Human Adversarial QA: Did the Model Understand the Paragraph?
Abstract
Recently, adversarial attacks have become an important means of gauging the robustness of natural language models as training and testing set methodology has proved inadequate. In this paper we explore an evaluation based on human-in-the-loop adversarial example generation. These adversarial examples aid us in finding the loopholes in the models and give insights into their working. In the published work on adversarial question-answering, perturbations are made on the questions without changing the background context on which the question is based. In the current work, we examine the complementary idea of perturbing the background context while keeping the question constant. We analyze the state-of-the-art language model BERT for the task of question-answering on SQuAD dataset using novel adversarial examples crafted by humans exposing the weaknesses of the model. We present the typology of the successful attacks here as a baseline for stress-testing QA systems.
Cite
Text
Rahurkar et al. "Human Adversarial QA: Did the Model Understand the Paragraph?." NeurIPS 2020 Workshops: HAMLETS, 2020.Markdown
[Rahurkar et al. "Human Adversarial QA: Did the Model Understand the Paragraph?." NeurIPS 2020 Workshops: HAMLETS, 2020.](https://mlanthology.org/neuripsw/2020/rahurkar2020neuripsw-human/)BibTeX
@inproceedings{rahurkar2020neuripsw-human,
title = {{Human Adversarial QA: Did the Model Understand the Paragraph?}},
author = {Rahurkar, Prachi Shriram and Olson, Matthew Lyle and Tadepalli, Prasad},
booktitle = {NeurIPS 2020 Workshops: HAMLETS},
year = {2020},
url = {https://mlanthology.org/neuripsw/2020/rahurkar2020neuripsw-human/}
}