Model-Level Dual Learning

Abstract

Many artificial intelligence tasks appear in dual forms like English$\leftrightarrow$French translation and speech$\leftrightarrow$text transformation. Existing dual learning schemes, which are proposed to solve a pair of such dual tasks, explore how to leverage such dualities from data level. In this work, we propose a new learning framework, model-level dual learning, which takes duality of tasks into consideration while designing the architectures for the primal/dual models, and ties the model parameters that playing similar roles in the two tasks. We study both symmetric and asymmetric model-level dual learning. Our algorithms achieve significant improvements on neural machine translation and sentiment analysis.

Cite

Text

Xia et al. "Model-Level Dual Learning." International Conference on Machine Learning, 2018.

Markdown

[Xia et al. "Model-Level Dual Learning." International Conference on Machine Learning, 2018.](https://mlanthology.org/icml/2018/xia2018icml-modellevel/)

BibTeX

@inproceedings{xia2018icml-modellevel,
  title     = {{Model-Level Dual Learning}},
  author    = {Xia, Yingce and Tan, Xu and Tian, Fei and Qin, Tao and Yu, Nenghai and Liu, Tie-Yan},
  booktitle = {International Conference on Machine Learning},
  year      = {2018},
  pages     = {5383-5392},
  volume    = {80},
  url       = {https://mlanthology.org/icml/2018/xia2018icml-modellevel/}
}