Contextual Weighting for Vocabulary Tree Based Image Retrieval

Abstract

In this paper we address the problem of image retrieval from millions of database images. We improve the vocabulary tree based approach by introducing contextual weighting of local features in both descriptor and spatial domains. Specifically, we propose to incorporate efficient statistics of neighbor descriptors both on the vocabulary tree and in the image spatial domain into the retrieval. These contextual cues substantially enhance the discriminative power of individual local features with very small computational overhead. We have conducted extensive experiments on benchmark datasets, i.e., the UKbench, Holidays, and our new Mobile dataset, which show that our method reaches state-of-the-art performance with much less computation. Furthermore, the proposed method demonstrates excellent scalability in terms of both retrieval accuracy and efficiency on large-scale experiments using 1.26 million images from the ImageNet database as distractors.

Cite

Text

Wang et al. "Contextual Weighting for Vocabulary Tree Based Image Retrieval." IEEE/CVF International Conference on Computer Vision, 2011. doi:10.1109/ICCV.2011.6126244

Markdown

[Wang et al. "Contextual Weighting for Vocabulary Tree Based Image Retrieval." IEEE/CVF International Conference on Computer Vision, 2011.](https://mlanthology.org/iccv/2011/wang2011iccv-contextual/) doi:10.1109/ICCV.2011.6126244

BibTeX

@inproceedings{wang2011iccv-contextual,
  title     = {{Contextual Weighting for Vocabulary Tree Based Image Retrieval}},
  author    = {Wang, Xiaoyu and Yang, Ming and Cour, Timothée and Zhu, Shenghuo and Yu, Kai and Han, Tony X.},
  booktitle = {IEEE/CVF International Conference on Computer Vision},
  year      = {2011},
  pages     = {209-216},
  doi       = {10.1109/ICCV.2011.6126244},
  url       = {https://mlanthology.org/iccv/2011/wang2011iccv-contextual/}
}