Relative Attributes for Enhanced Human-Machine Communication

Abstract

We propose to model relative attributes that capture the relationships between images and objects in terms of human-nameable visual properties. For example, the models can capture that animal A is 'furrier' than animal B, or image X is 'brighter' than image B. Given training data stating how object/scene categories relate according to different attributes, we learn a ranking function per attribute. The learned ranking functions predict the relative strength of each property in novel images. We show how these relative attribute predictions enable a variety of novel applications, including zero-shot learning from relative comparisons, automatic image description, image search with interactive feedback, and active learning of discriminative classifiers. We overview results demonstrating these applications with images of faces and natural scenes. Overall, we find that relative attributes enhance the precision of communication between humans and computer vision algorithms, providing the richer language needed to fluidly "teach" a system about visual concepts.

Cite

Text

Parikh et al. "Relative Attributes for Enhanced Human-Machine Communication." AAAI Conference on Artificial Intelligence, 2012. doi:10.1609/AAAI.V26I1.8443

Markdown

[Parikh et al. "Relative Attributes for Enhanced Human-Machine Communication." AAAI Conference on Artificial Intelligence, 2012.](https://mlanthology.org/aaai/2012/parikh2012aaai-relative/) doi:10.1609/AAAI.V26I1.8443

BibTeX

@inproceedings{parikh2012aaai-relative,
  title     = {{Relative Attributes for Enhanced Human-Machine Communication}},
  author    = {Parikh, Devi and Kovashka, Adriana and Parkash, Amar and Grauman, Kristen},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2012},
  pages     = {2153-2159},
  doi       = {10.1609/AAAI.V26I1.8443},
  url       = {https://mlanthology.org/aaai/2012/parikh2012aaai-relative/}
}