ResMatch: Residual Attention Learning for Feature Matching
Abstract
Attention-based graph neural networks have made great progress in feature matching. However, the literature lacks a comprehensive understanding of how the attention mechanism operates for feature matching. In this paper, we rethink cross- and self-attention from the viewpoint of traditional feature matching and filtering. To facilitate the learning of matching and filtering, we incorporate the similarity of descriptors into cross-attention and relative positions into self-attention. In this way, the attention can concentrate on learning residual matching and filtering functions with reference to the basic functions of measuring visual and spatial correlation. Moreover, we leverage descriptor similarity and relative positions to extract inter- and intra-neighbors. Then sparse attention for each point can be performed only within its neighborhoods to acquire higher computation efficiency. Extensive experiments, including feature matching, pose estimation and visual localization, confirm the superiority of the proposed method. Our codes are available at https://github.com/ACuOoOoO/ResMatch.
Cite
Text
Deng et al. "ResMatch: Residual Attention Learning for Feature Matching." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I2.27915Markdown
[Deng et al. "ResMatch: Residual Attention Learning for Feature Matching." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/deng2024aaai-resmatch/) doi:10.1609/AAAI.V38I2.27915BibTeX
@inproceedings{deng2024aaai-resmatch,
title = {{ResMatch: Residual Attention Learning for Feature Matching}},
author = {Deng, Yuxin and Zhang, Kaining and Zhang, Shihua and Li, Yansheng and Ma, Jiayi},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2024},
pages = {1501-1509},
doi = {10.1609/AAAI.V38I2.27915},
url = {https://mlanthology.org/aaai/2024/deng2024aaai-resmatch/}
}