Randomized Visual Phrases for Object Search

Abstract

Accurate matching of local features plays an essential role in visual object search. Instead of matching individual features separately, using the spatial context, e.g., bundling a group of co-located features into a visual phrase, has shown to enable more discriminative matching. Despite previous work, it remains a challenging problem to extract appropriate spatial context for matching. We propose a randomized approach to deriving visual phrase, in the form of spatial random partition. By averaging the matching scores over multiple randomized visual phrases, our approach offers three benefits: 1) the aggregation of the matching scores over a collection of visual phrases of varying sizes and shapes provides robust local matching; 2) object localization is achieved by simple thresholding on the voting map, which is more efficient than subimage search; 3) our algorithm lends itself to easy parallelization and also allows a flexible trade-off between accuracy and speed by adjusting the number of partition times. Both theoretical studies and experimental comparisons with the state-of-the-art methods validate the advantages of our approach.

Cite

Text

Jiang et al. "Randomized Visual Phrases for Object Search." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2012. doi:10.1109/CVPR.2012.6248042

Markdown

[Jiang et al. "Randomized Visual Phrases for Object Search." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2012.](https://mlanthology.org/cvpr/2012/jiang2012cvpr-randomized/) doi:10.1109/CVPR.2012.6248042

BibTeX

@inproceedings{jiang2012cvpr-randomized,
  title     = {{Randomized Visual Phrases for Object Search}},
  author    = {Jiang, Yuning and Meng, Jingjing and Yuan, Junsong},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2012},
  pages     = {3100-3107},
  doi       = {10.1109/CVPR.2012.6248042},
  url       = {https://mlanthology.org/cvpr/2012/jiang2012cvpr-randomized/}
}