Exploring Human-like Attention Supervision in Visual Question Answering
Abstract
Attention mechanisms have been widely applied in the Visual Question Answering (VQA) task, as they help to focus on the area-of-interest of both visual and textual information. To answer the questions correctly, the model needs to selectively target different areas of an image, which suggests that an attention-based model may benefit from an explicit attention supervision. In this work, we aim to address the problem of adding attention supervision to VQA models. Since there is a lack of human attention data, we first propose a Human Attention Network (HAN) to generate human-like attention maps, training on a recently released dataset called Human ATtention Dataset (VQA-HAT). Then, we apply the pre-trained HAN on the VQA v2.0 dataset to automatically produce the human-like attention maps for all image-question pairs. The generated human-like attention map dataset for the VQA v2.0 dataset is named as Human-Like ATtention (HLAT) dataset. Finally, we apply human-like attention supervision to an attention-based VQA model. The experiments show that adding human-like supervision yields a more accurate attention together with a better performance, showing a promising future for human-like attention supervision in VQA.
Cite
Text
Qiao et al. "Exploring Human-like Attention Supervision in Visual Question Answering." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.12272Markdown
[Qiao et al. "Exploring Human-like Attention Supervision in Visual Question Answering." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/qiao2018aaai-exploring/) doi:10.1609/AAAI.V32I1.12272BibTeX
@inproceedings{qiao2018aaai-exploring,
title = {{Exploring Human-like Attention Supervision in Visual Question Answering}},
author = {Qiao, Tingting and Dong, Jianfeng and Xu, Duanqing},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2018},
pages = {7300-7307},
doi = {10.1609/AAAI.V32I1.12272},
url = {https://mlanthology.org/aaai/2018/qiao2018aaai-exploring/}
}