Learning to Make Keypoints Sub-Pixel Accurate
Abstract
This work addresses the challenge of sub-pixel accuracy in detecting 2D local features, a cornerstone problem in computer vision. Despite the advancements brought by neural network-based methods like SuperPoint and ALIKED, these modern approaches lag behind classical ones such as SIFT in keypoint localization accuracy due to their lack of sub-pixel precision. We propose a novel network that enhances any detector with sub-pixel precision by learning an offset vector for detected features, thereby eliminating the need for designing specialized sub-pixel accurate detectors. This optimization directly minimizes test-time evaluation metrics like relative pose error. Through extensive testing with both nearest neighbors matching and the recent LightGlue matcher across various real-world datasets, our method consistently outperforms existing methods in accuracy. Moreover, it adds only around 7 ms to the time of a particular detector. The code is available at https://github.com/KimSinjeong/keypt2subpx.
Cite
Text
Kim et al. "Learning to Make Keypoints Sub-Pixel Accurate." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-73229-4_24Markdown
[Kim et al. "Learning to Make Keypoints Sub-Pixel Accurate." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/kim2024eccv-learning-a/) doi:10.1007/978-3-031-73229-4_24BibTeX
@inproceedings{kim2024eccv-learning-a,
title = {{Learning to Make Keypoints Sub-Pixel Accurate}},
author = {Kim, Shinjeong and Pollefeys, Marc and Barath, Daniel},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2024},
doi = {10.1007/978-3-031-73229-4_24},
url = {https://mlanthology.org/eccv/2024/kim2024eccv-learning-a/}
}