Reinforcing Coherence for Sequence to Sequence Model in Dialogue Generation

Abstract

Sequence to sequence (Seq2Seq) approach has gained great attention in the field of single-turn dialogue generation. However, one serious problem is that most existing Seq2Seq based models tend to generate common responses lacking specific meanings. Our analysis show that the underlying reason is that Seq2Seq is equivalent to optimizing Kullback–Leibler (KL) divergence, thus does not penalize the case whose generated probability is high while the true probability is low. However, the true probability is unknown, which poses challenges for tackling this problem. Inspired by the fact that the coherence (i.e. similarity) between post and response is consistent with human evaluation, we hypothesize that the true probability of a response is proportional to the coherence degree. The coherence scores are then used as the reward function in a reinforcement learning framework to penalize the case whose generated probability is high while the true probability is low. Three different types of coherence models, including an unlearned similarity function, a pretrained semantic matching function, and an end-to-end dual learning architecture, are proposed in this paper. Experimental results on both Chinese Weibo dataset and English Subtitle dataset show that the proposed models produce more specific and meaningful responses, yielding better performances against Seq2Seq models in terms of both metric-based and human evaluations.

Cite

Text

Zhang et al. "Reinforcing Coherence for Sequence to Sequence Model in Dialogue Generation." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/635

Markdown

[Zhang et al. "Reinforcing Coherence for Sequence to Sequence Model in Dialogue Generation." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/zhang2018ijcai-reinforcing/) doi:10.24963/IJCAI.2018/635

BibTeX

@inproceedings{zhang2018ijcai-reinforcing,
  title     = {{Reinforcing Coherence for Sequence to Sequence Model in Dialogue Generation}},
  author    = {Zhang, Hainan and Lan, Yanyan and Guo, Jiafeng and Xu, Jun and Cheng, Xueqi},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {4567-4573},
  doi       = {10.24963/IJCAI.2018/635},
  url       = {https://mlanthology.org/ijcai/2018/zhang2018ijcai-reinforcing/}
}