Location Sensitive Image Retrieval and Tagging
Abstract
People from different parts of the globe describe objects and concepts in distinct manners. Visual appearance can thus vary across different geographic locations, which makes location a relevant contextual information when analysing visual data. In this work, we address the task of image retrieval related to a given tag conditioned on a certain location on Earth. We present LocSens, a model that learns to rank triplets of images, tags and coordinates by plausibility, and two training strategies to balance the location influence in the final ranking. LocSens learns to fuse textual and location information of multimodal queries to retrieve related images at different levels of location granularity, and successfully utilizes location information to improve image tagging. Code and models will be available to ensure reproducibility.
Cite
Text
Gomez et al. "Location Sensitive Image Retrieval and Tagging." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58517-4_38Markdown
[Gomez et al. "Location Sensitive Image Retrieval and Tagging." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/gomez2020eccv-location/) doi:10.1007/978-3-030-58517-4_38BibTeX
@inproceedings{gomez2020eccv-location,
title = {{Location Sensitive Image Retrieval and Tagging}},
author = {Gomez, Raul and Gibert, Jaume and Gomez, Lluis and Karatzas, Dimosthenis},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58517-4_38},
url = {https://mlanthology.org/eccv/2020/gomez2020eccv-location/}
}