Fine-Grained Visual Comparisons with Local Learning

Abstract

Given two images, we want to predict which exhibits a particular visual attribute more than the other---even when the two images are quite similar. Existing relative attribute methods rely on global ranking functions; yet rarely will the visual cues relevant to a comparison be constant for all data, nor will humans' perception of the attribute necessarily permit a global ordering. To address these issues, we propose a local learning approach for fine-grained visual comparisons. Given a novel pair of images, we learn a local ranking model on the fly, using only analogous training comparisons. We show how to identify these analogous pairs using learned metrics. With results on three challenging datasets -- including a large newly curated dataset for fine-grained comparisons -- our method outperforms state-of-the-art methods for relative attribute prediction.

Cite

Text

Yu and Grauman. "Fine-Grained Visual Comparisons with Local Learning." Conference on Computer Vision and Pattern Recognition, 2014. doi:10.1109/CVPR.2014.32

Markdown

[Yu and Grauman. "Fine-Grained Visual Comparisons with Local Learning." Conference on Computer Vision and Pattern Recognition, 2014.](https://mlanthology.org/cvpr/2014/yu2014cvpr-finegrained/) doi:10.1109/CVPR.2014.32

BibTeX

@inproceedings{yu2014cvpr-finegrained,
  title     = {{Fine-Grained Visual Comparisons with Local Learning}},
  author    = {Yu, Aron and Grauman, Kristen},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2014},
  doi       = {10.1109/CVPR.2014.32},
  url       = {https://mlanthology.org/cvpr/2014/yu2014cvpr-finegrained/}
}