Addressing the Under-Translation Problem from the Entropy Perspective
Abstract
Neural Machine Translation (NMT) has drawn much attention due to its promising translation performance in recent years. However, the under-translation problem still remains a big challenge. In this paper, we focus on the under-translation problem and attempt to find out what kinds of source words are more likely to be ignored. Through analysis, we observe that a source word with a large translation entropy is more inclined to be dropped. To address this problem, we propose a coarse-to-fine framework. In coarse-grained phase, we introduce a simple strategy to reduce the entropy of highentropy words through constructing the pseudo target sentences. In fine-grained phase, we propose three methods, including pre-training method, multitask method and two-pass method, to encourage the neural model to correctly translate these high-entropy words. Experimental results on various translation tasks show that our method can significantly improve the translation quality and substantially reduce the under-translation cases of high-entropy words.
Cite
Text
Zhao et al. "Addressing the Under-Translation Problem from the Entropy Perspective." AAAI Conference on Artificial Intelligence, 2019. doi:10.1609/AAAI.V33I01.3301451Markdown
[Zhao et al. "Addressing the Under-Translation Problem from the Entropy Perspective." AAAI Conference on Artificial Intelligence, 2019.](https://mlanthology.org/aaai/2019/zhao2019aaai-addressing/) doi:10.1609/AAAI.V33I01.3301451BibTeX
@inproceedings{zhao2019aaai-addressing,
title = {{Addressing the Under-Translation Problem from the Entropy Perspective}},
author = {Zhao, Yang and Zhang, Jiajun and Zong, Chengqing and He, Zhongjun and Wu, Hua},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2019},
pages = {451-458},
doi = {10.1609/AAAI.V33I01.3301451},
url = {https://mlanthology.org/aaai/2019/zhao2019aaai-addressing/}
}