Cross-Lingual Retrieval for Iterative Self-Supervised Training

Abstract

Recent studies have demonstrated the cross-lingual alignment ability of multilingual pretrained language models. In this work, we found that the cross-lingual alignment can be further improved by training seq2seq models on sentence pairs mined using their own encoder outputs. We utilized these findings to develop a new approach --- cross-lingual retrieval for iterative self-supervised training (CRISS), where mining and training processes are applied iteratively, improving cross-lingual alignment and translation ability at the same time. Using this method, we achieved state-of-the-art unsupervised machine translation results on 9 language directions with an average improvement of 2.4 BLEU, and on the Tatoeba sentence retrieval task in the XTREME benchmark on 16 languages with an average improvement of 21.5% in absolute accuracy. Furthermore, CRISS also brings an additional 1.8 BLEU improvement on average compared to mBART, when finetuned on supervised machine translation downstream tasks.

Cite

Text

Tran et al. "Cross-Lingual Retrieval for Iterative Self-Supervised Training." Neural Information Processing Systems, 2020.

Markdown

[Tran et al. "Cross-Lingual Retrieval for Iterative Self-Supervised Training." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/tran2020neurips-crosslingual/)

BibTeX

@inproceedings{tran2020neurips-crosslingual,
  title     = {{Cross-Lingual Retrieval for Iterative Self-Supervised Training}},
  author    = {Tran, Chau and Tang, Yuqing and Li, Xian and Gu, Jiatao},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/tran2020neurips-crosslingual/}
}