Regularizing Neural Machine Translation by Target-Bidirectional Agreement
Abstract
Although Neural Machine Translation (NMT) has achieved remarkable progress in the past several years, most NMT systems still suffer from a fundamental shortcoming as in other sequence generation tasks: errors made early in generation process are fed as inputs to the model and can be quickly amplified, harming subsequent sequence generation. To address this issue, we propose a novel model regularization method for NMT training, which aims to improve the agreement between translations generated by left-to-right (L2R) and right-to-left (R2L) NMT decoders. This goal is achieved by introducing two Kullback-Leibler divergence regularization terms into the NMT training objective to reduce the mismatch between output probabilities of L2R and R2L models. In addition, we also employ a joint training strategy to allow L2R and R2L models to improve each other in an interactive update process. Experimental results show that our proposed method significantly outperforms state-of-the-art baselines on Chinese-English and English-German translation tasks.
Cite
Text
Zhang et al. "Regularizing Neural Machine Translation by Target-Bidirectional Agreement." AAAI Conference on Artificial Intelligence, 2019. doi:10.1609/AAAI.V33I01.3301443Markdown
[Zhang et al. "Regularizing Neural Machine Translation by Target-Bidirectional Agreement." AAAI Conference on Artificial Intelligence, 2019.](https://mlanthology.org/aaai/2019/zhang2019aaai-regularizing/) doi:10.1609/AAAI.V33I01.3301443BibTeX
@inproceedings{zhang2019aaai-regularizing,
title = {{Regularizing Neural Machine Translation by Target-Bidirectional Agreement}},
author = {Zhang, Zhirui and Wu, Shuangzhi and Liu, Shujie and Li, Mu and Zhou, Ming and Xu, Tong},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2019},
pages = {443-450},
doi = {10.1609/AAAI.V33I01.3301443},
url = {https://mlanthology.org/aaai/2019/zhang2019aaai-regularizing/}
}