SuperPoint: Self-Supervised Interest Point Detection and Description
Abstract
This paper presents a self-supervised framework for training interest point detectors and descriptors suitable for a large number of multiple-view geometry problems in computer vision. As opposed to patch-based neural networks, our fully-convolutional model operates on full-sized images and jointly computes pixel-level interest point locations and associated descriptors in one forward pass. We introduce Homographic Adaptation, a multi-scale, multi-homography approach for boosting interest point detection repeatability and performing cross-domain adaptation (e.g., synthetic-to-real). Our model, when trained on the MS-COCO generic image dataset using Homographic Adaptation, is able to repeatedly detect a much richer set of interest points than the initial pre-adapted deep model and any other traditional corner detector. The final system gives rise to state-of-the-art homography estimation results on HPatches when compared to LIFT, SIFT and ORB.
Cite
Text
DeTone et al. "SuperPoint: Self-Supervised Interest Point Detection and Description." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018. doi:10.1109/CVPRW.2018.00060Markdown
[DeTone et al. "SuperPoint: Self-Supervised Interest Point Detection and Description." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/detone2018cvprw-superpoint/) doi:10.1109/CVPRW.2018.00060BibTeX
@inproceedings{detone2018cvprw-superpoint,
title = {{SuperPoint: Self-Supervised Interest Point Detection and Description}},
author = {DeTone, Daniel and Malisiewicz, Tomasz and Rabinovich, Andrew},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2018},
pages = {224-236},
doi = {10.1109/CVPRW.2018.00060},
url = {https://mlanthology.org/cvprw/2018/detone2018cvprw-superpoint/}
}