Efficient Representation of Local Geometry for Large Scale Object Retrieval

Abstract

State of the art methods for image and object retrieval exploit both appearance (via visual words) and local geometry (spatial extent, relative pose). In large scale problems, memory becomes a limiting factor - local geometry is stored for each feature detected in each image and requires storage larger than the inverted file and term frequency and inverted document frequency weights together. We propose a novel method for learning discretized local geometry representation based on minimization of average reprojection error in the space of ellipses. The representation requires only 24 bits per feature without drop in performance. Additionally, we show that if the gravity vector assumption is used consistently from the feature description to spatial verification, it improves retrieval performance and decreases the memory footprint. The proposed method outperforms state of the art retrieval algorithms in a standard image retrieval benchmark.

Cite

Text

Perdoch et al. "Efficient Representation of Local Geometry for Large Scale Object Retrieval." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2009. doi:10.1109/CVPR.2009.5206529

Markdown

[Perdoch et al. "Efficient Representation of Local Geometry for Large Scale Object Retrieval." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2009.](https://mlanthology.org/cvpr/2009/perdoch2009cvpr-efficient/) doi:10.1109/CVPR.2009.5206529

BibTeX

@inproceedings{perdoch2009cvpr-efficient,
  title     = {{Efficient Representation of Local Geometry for Large Scale Object Retrieval}},
  author    = {Perdoch, Michal and Chum, Ondrej and Matas, Jiri},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2009},
  pages     = {9-16},
  doi       = {10.1109/CVPR.2009.5206529},
  url       = {https://mlanthology.org/cvpr/2009/perdoch2009cvpr-efficient/}
}