Multi-Step Retriever-Reader Interaction for Scalable Open-Domain Question Answering
Abstract
This paper introduces a new framework for open-domain question answering in which the retriever and the reader \emph{iteratively interact} with each other. The framework is agnostic to the architecture of the machine reading model provided it has \emph{access} to the token-level hidden representations of the reader. The retriever uses fast nearest neighbor search that allows it to scale to corpora containing millions of paragraphs. A gated recurrent unit updates the query at each step conditioned on the \emph{state} of the reader and the \emph{reformulated} query is used to re-rank the paragraphs by the retriever. We conduct analysis and show that iterative interaction helps in retrieving informative paragraphs from the corpus. Finally, we show that our multi-step-reasoning framework brings consistent improvement when applied to two widely used reader architectures (\drqa and \bidaf) on various large open-domain datasets ---\tqau, \quasart, \searchqa, and \squado\footnote{Code and pretrained models are available at \url{https://github.com/rajarshd/Multi-Step-Reasoning}}.
Cite
Text
Das et al. "Multi-Step Retriever-Reader Interaction for Scalable Open-Domain Question Answering." International Conference on Learning Representations, 2019.Markdown
[Das et al. "Multi-Step Retriever-Reader Interaction for Scalable Open-Domain Question Answering." International Conference on Learning Representations, 2019.](https://mlanthology.org/iclr/2019/das2019iclr-multistep/)BibTeX
@inproceedings{das2019iclr-multistep,
title = {{Multi-Step Retriever-Reader Interaction for Scalable Open-Domain Question Answering}},
author = {Das, Rajarshi and Dhuliawala, Shehzaad and Zaheer, Manzil and McCallum, Andrew},
booktitle = {International Conference on Learning Representations},
year = {2019},
url = {https://mlanthology.org/iclr/2019/das2019iclr-multistep/}
}