ERNIE-GEN: An Enhanced Multi-Flow Pre-Training and Fine-Tuning Framework for Natural Language Generation
Abstract
Current pre-training works in natural language generation pay little attention to the problem of exposure bias on downstream tasks. To address this issue, we propose an enhanced multi-flow sequence to sequence pre-training and fine-tuning framework named ERNIE-GEN, which bridges the discrepancy between training and inference with an infilling generation mechanism and a noise-aware generation method. To make generation closer to human writing patterns, this framework introduces a span-by-span generation flow that trains the model to predict semantically-complete spans consecutively rather than predicting word by word. Unlike existing pre-training methods, ERNIE-GEN incorporates multi-granularity target sampling to construct pre-training data, which enhances the correlation between encoder and decoder. Experimental results demonstrate that ERNIE-GEN achieves state-of-the-art results with a much smaller amount of pre-training data and parameters on a range of language generation tasks, including abstractive summarization (Gigaword and CNN/DailyMail), question generation (SQuAD), dialogue generation (Persona-Chat) and generative question answering (CoQA). The source codes and pre-trained models have been released at https://github.com/PaddlePaddle/ERNIE/ernie-gen.
Cite
Text
Xiao et al. "ERNIE-GEN: An Enhanced Multi-Flow Pre-Training and Fine-Tuning Framework for Natural Language Generation." International Joint Conference on Artificial Intelligence, 2020. doi:10.24963/IJCAI.2020/553Markdown
[Xiao et al. "ERNIE-GEN: An Enhanced Multi-Flow Pre-Training and Fine-Tuning Framework for Natural Language Generation." International Joint Conference on Artificial Intelligence, 2020.](https://mlanthology.org/ijcai/2020/xiao2020ijcai-ernie/) doi:10.24963/IJCAI.2020/553BibTeX
@inproceedings{xiao2020ijcai-ernie,
title = {{ERNIE-GEN: An Enhanced Multi-Flow Pre-Training and Fine-Tuning Framework for Natural Language Generation}},
author = {Xiao, Dongling and Zhang, Han and Li, Yu-Kun and Sun, Yu and Tian, Hao and Wu, Hua and Wang, Haifeng},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2020},
pages = {3997-4003},
doi = {10.24963/IJCAI.2020/553},
url = {https://mlanthology.org/ijcai/2020/xiao2020ijcai-ernie/}
}