Cross-Lingual Natural Language Generation via Pre-Training
Abstract
In this work we focus on transferring supervision signals of natural language generation (NLG) tasks between multiple languages. We propose to pretrain the encoder and the decoder of a sequence-to-sequence model under both monolingual and cross-lingual settings. The pre-training objective encourages the model to represent different languages in the shared space, so that we can conduct zero-shot cross-lingual transfer. After the pre-training procedure, we use monolingual data to fine-tune the pre-trained model on downstream NLG tasks. Then the sequence-to-sequence model trained in a single language can be directly evaluated beyond that language (i.e., accepting multi-lingual input and producing multi-lingual output). Experimental results on question generation and abstractive summarization show that our model outperforms the machine-translation-based pipeline methods for zero-shot cross-lingual generation. Moreover, cross-lingual transfer improves NLG performance of low-resource languages by leveraging rich-resource language data. Our implementation and data are available at this https URL.
Cite
Text
Chi et al. "Cross-Lingual Natural Language Generation via Pre-Training." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I05.6256Markdown
[Chi et al. "Cross-Lingual Natural Language Generation via Pre-Training." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/chi2020aaai-cross/) doi:10.1609/AAAI.V34I05.6256BibTeX
@inproceedings{chi2020aaai-cross,
title = {{Cross-Lingual Natural Language Generation via Pre-Training}},
author = {Chi, Zewen and Dong, Li and Wei, Furu and Wang, Wenhui and Mao, Xian-Ling and Huang, Heyan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2020},
pages = {7570-7577},
doi = {10.1609/AAAI.V34I05.6256},
url = {https://mlanthology.org/aaai/2020/chi2020aaai-cross/}
}