Reinforced History Backtracking for Conversational Question Answering
Abstract
To model the context history in multi-turn conversations has become a critical step towards a better understanding of the user query in question answering systems. To utilize the context history, most existing studies treat the whole context as input, which will inevitably face the following two challenges. First, modeling a long history can be costly as it requires more computation resources. Second, the long context history consists of a lot of irrelevant information that makes it difficult to model appropriate information relevant to the user query. To alleviate these problems, we propose a reinforcement learning based method to capture and backtrack the related conversation history to boost model performance in this paper. Our method seeks to automatically backtrack the history information with the implicit feedback from the model performance. We further consider both immediate and delayed rewards to guide the reinforced backtracking policy. Extensive experiments on a large conversational question answering dataset show that the proposed method can help to alleviate the problems arising from longer context history. Meanwhile, experiments show that the method yields better performance than other strong baselines, and the actions made by the method are insightful.
Cite
Text
Qiu et al. "Reinforced History Backtracking for Conversational Question Answering." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I15.17617Markdown
[Qiu et al. "Reinforced History Backtracking for Conversational Question Answering." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/qiu2021aaai-reinforced/) doi:10.1609/AAAI.V35I15.17617BibTeX
@inproceedings{qiu2021aaai-reinforced,
title = {{Reinforced History Backtracking for Conversational Question Answering}},
author = {Qiu, Minghui and Huang, Xinjing and Chen, Cen and Ji, Feng and Qu, Chen and Wei, Wei and Huang, Jun and Zhang, Yin},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2021},
pages = {13718-13726},
doi = {10.1609/AAAI.V35I15.17617},
url = {https://mlanthology.org/aaai/2021/qiu2021aaai-reinforced/}
}