Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation

Abstract

Meta-learning has been sufficiently validated to be beneficial for low-resource neural machine translation (NMT). However, we find that meta-trained NMT fails to improve the translation performance of the domain unseen at the meta-training stage. In this paper, we aim to alleviate this issue by proposing a novel meta-curriculum learning for domain adaptation in NMT. During meta-training, the NMT first learns the similar curricula from each domain to avoid falling into a bad local optimum early, and finally learns the curricula of individualities to improve the model robustness for learning domain-specific knowledge. Experimental results on 10 different low-resource domains show that meta-curriculum learning can improve the translation performance of both familiar and unfamiliar domains. All the codes and data are freely available at https://github.com/NLP2CT/Meta-Curriculum.

Cite

Text

Zhan et al. "Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I16.17683

Markdown

[Zhan et al. "Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/zhan2021aaai-meta/) doi:10.1609/AAAI.V35I16.17683

BibTeX

@inproceedings{zhan2021aaai-meta,
  title     = {{Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation}},
  author    = {Zhan, Runzhe and Liu, Xuebo and Wong, Derek F. and Chao, Lidia S.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {14310-14318},
  doi       = {10.1609/AAAI.V35I16.17683},
  url       = {https://mlanthology.org/aaai/2021/zhan2021aaai-meta/}
}