Deep Reinforced Attention Learning for Quality-Aware Visual Recognition
Abstract
In this paper, we build upon the weakly-supervised generation mechanism of intermediate attention maps in any convolutional neural networks and disclose the effectiveness of attention modules more straightforwardly to fully exploit their potential. Given an existing neural network equipped with arbitrary attention modules, we introduce a meta critic network to evaluate the quality of attention maps in the main network. Due to the discreteness of our designed reward, the proposed learning method is arranged in a reinforcement learning setting, where the attention actors and recurrent critics are alternately optimized to provide instant critique and revision for the temporary attention representation, hence coined as Deep REinforced Attention Learning (DREAL). It could be applied universally to network architectures with different types of attention modules and promotes their expressive ability by maximizing the relative gain of the final recognition performance arising from each individual attention module, as demonstrated by extensive experiments on both category and instance recognition benchmarks.
Cite
Text
Li and Chen. "Deep Reinforced Attention Learning for Quality-Aware Visual Recognition." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58517-4_29Markdown
[Li and Chen. "Deep Reinforced Attention Learning for Quality-Aware Visual Recognition." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/li2020eccv-deep/) doi:10.1007/978-3-030-58517-4_29BibTeX
@inproceedings{li2020eccv-deep,
title = {{Deep Reinforced Attention Learning for Quality-Aware Visual Recognition}},
author = {Li, Duo and Chen, Qifeng},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58517-4_29},
url = {https://mlanthology.org/eccv/2020/li2020eccv-deep/}
}