Love Thy Neighbors: Image Annotation by Exploiting Image Metadata
Abstract
Some images that are difficult to recognize on their own may become more clear in the context of a neighborhood of related images with similar social-network metadata. We build on this intuition to improve multilabel image annotation. Our model uses image metadata nonparametrically to generate neighborhoods of related images using Jaccard similarities, then uses a deep neural network to blend visual information from the image and its neighbors. Prior work typically models image metadata parametrically; in contrast, our nonparametric treatment allows our model to perform well even when the vocabulary of metadata changes between training and testing. We perform comprehensive experiments on the NUS-WIDE dataset, where we show that our model outperforms state-of-the-art methods for multilabel image annotation even when our model is forced to generalize to new types of metadata.
Cite
Text
Johnson et al. "Love Thy Neighbors: Image Annotation by Exploiting Image Metadata." International Conference on Computer Vision, 2015. doi:10.1109/ICCV.2015.525Markdown
[Johnson et al. "Love Thy Neighbors: Image Annotation by Exploiting Image Metadata." International Conference on Computer Vision, 2015.](https://mlanthology.org/iccv/2015/johnson2015iccv-love/) doi:10.1109/ICCV.2015.525BibTeX
@inproceedings{johnson2015iccv-love,
title = {{Love Thy Neighbors: Image Annotation by Exploiting Image Metadata}},
author = {Johnson, Justin and Ballan, Lamberto and Fei-Fei, Li},
booktitle = {International Conference on Computer Vision},
year = {2015},
doi = {10.1109/ICCV.2015.525},
url = {https://mlanthology.org/iccv/2015/johnson2015iccv-love/}
}