Deep Visual Geo-Localization Benchmark

Abstract

In this paper, we propose a new open-source benchmarking framework for Visual Geo-localization (VG) that allows to build, train, and test a wide range of commonly used architectures, with the flexibility to change individual components of a geo-localization pipeline. The purpose of this framework is twofold: i) gaining insights into how different components and design choices in a VG pipeline impact the final results, both in terms of performance (recall@N metric) and system requirements (such as execution time and memory consumption); ii) establish a systematic evaluation protocol for comparing different methods. Using the proposed framework, we perform a large suite of experiments which provide criteria for choosing backbone, aggregation and negative mining depending on the use-case and requirements. We also assess the impact of engineering techniques like pre/post-processing, data augmentation and image resizing, showing that better performance can be obtained through somewhat simple procedures: for example, downscaling the images' resolution to 80% can lead to similar results with a 36% savings in extraction time and dataset storage requirement. Code and trained models are available at https://deep-vg-bench.herokuapp.com/.

Cite

Text

Berton et al. "Deep Visual Geo-Localization Benchmark." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00532

Markdown

[Berton et al. "Deep Visual Geo-Localization Benchmark." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/berton2022cvpr-deep/) doi:10.1109/CVPR52688.2022.00532

BibTeX

@inproceedings{berton2022cvpr-deep,
  title     = {{Deep Visual Geo-Localization Benchmark}},
  author    = {Berton, Gabriele and Mereu, Riccardo and Trivigno, Gabriele and Masone, Carlo and Csurka, Gabriela and Sattler, Torsten and Caputo, Barbara},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {5396-5407},
  doi       = {10.1109/CVPR52688.2022.00532},
  url       = {https://mlanthology.org/cvpr/2022/berton2022cvpr-deep/}
}