Relevance-Promoting Language Model for Short-Text Conversation
Abstract
Despite the effectiveness of sequence-to-sequence framework on the task of Short-Text Conversation (STC), the issue of under-exploitation of training data (i.e., the supervision signals from query text is ignored) still remains unresolved. Also, the adopted maximization-based decoding strategies, inclined to generating the generic responses or responses with repetition, are unsuited to the STC task. In this paper, we propose to formulate the STC task as a language modeling problem and tailor-make a training strategy to adapt a language model for response generation. To enhance generation performance, we design a relevance-promoting transformer language model, which performs additional supervised source attention after the self-attention to increase the importance of informative query tokens in calculating the token-level representation. The model further refines the query representation with relevance clues inferred from its multiple references during training. In testing, we adopt a randomization-over-maximization strategy to reduce the generation of generic responses. Experimental results on a large Chinese STC dataset demonstrate the superiority of the proposed model on relevance metrics and diversity metrics.1
Cite
Text
Li et al. "Relevance-Promoting Language Model for Short-Text Conversation." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I05.6340Markdown
[Li et al. "Relevance-Promoting Language Model for Short-Text Conversation." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/li2020aaai-relevance/) doi:10.1609/AAAI.V34I05.6340BibTeX
@inproceedings{li2020aaai-relevance,
title = {{Relevance-Promoting Language Model for Short-Text Conversation}},
author = {Li, Xin and Li, Piji and Bi, Wei and Liu, Xiaojiang and Lam, Wai},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2020},
pages = {8253-8260},
doi = {10.1609/AAAI.V34I05.6340},
url = {https://mlanthology.org/aaai/2020/li2020aaai-relevance/}
}