True Bilingual NMT
Abstract
Bilingual machine translation permits training a single model that translates monolingual sentences from one language to another. However, a model is not truly bilingual unless it can translate back and forth in both language directions it was trained on, along with translating code-switched sentences to either language. We propose a true bilingual model trained on WMT14 English-French (En-Fr) dataset. For better use of parallel data, we generated synthetic code-switched (CSW) data along with an alignment loss on the encoder to align representations across languages. Our model strongly outperforms bilingual baselines on CSW translation while maintaining quality for non-code switched data.
Cite
Text
Anwar et al. "True Bilingual NMT." ICLR 2022 Workshops: AfricaNLP, 2022.Markdown
[Anwar et al. "True Bilingual NMT." ICLR 2022 Workshops: AfricaNLP, 2022.](https://mlanthology.org/iclrw/2022/anwar2022iclrw-true/)BibTeX
@inproceedings{anwar2022iclrw-true,
title = {{True Bilingual NMT}},
author = {Anwar, Mohamed and Raheem, Lekan and Elrasheed, Maab and Johnson, Melvin and Kreutzer, Julia},
booktitle = {ICLR 2022 Workshops: AfricaNLP},
year = {2022},
url = {https://mlanthology.org/iclrw/2022/anwar2022iclrw-true/}
}