Multi-Modal Reasoning Graph for Scene-Text Based Fine-Grained Image Classification and Retrieval

Abstract

Scene text instances found in natural images carry explicit semantic information that can provide important cues to solve a wide array of computer vision problems. In this paper, we focus on leveraging multi-modal content in the form of visual and textual cues to tackle the task of fine-grained image classification and retrieval. First, we obtain the text instances from images by employing a text reading system. Then, we combine textual features with salient image regions to exploit the complementary information carried by the two sources. Specifically, we employ a Graph Convolutional Network to perform multi-modal reasoning and obtain relationship-enhanced features by learning a common semantic space between salient objects and text found in an image. By obtaining an enhanced set of visual and textual features, the proposed model greatly outperforms previous state-of-the-art in two different tasks, fine-grained classification and image retrieval in the Con-Text and Drink Bottle datasets.

Cite

Text

Mafla et al. "Multi-Modal Reasoning Graph for Scene-Text Based Fine-Grained Image Classification and Retrieval." Winter Conference on Applications of Computer Vision, 2021.

Markdown

[Mafla et al. "Multi-Modal Reasoning Graph for Scene-Text Based Fine-Grained Image Classification and Retrieval." Winter Conference on Applications of Computer Vision, 2021.](https://mlanthology.org/wacv/2021/mafla2021wacv-multimodal/)

BibTeX

@inproceedings{mafla2021wacv-multimodal,
  title     = {{Multi-Modal Reasoning Graph for Scene-Text Based Fine-Grained Image Classification and Retrieval}},
  author    = {Mafla, Andres and Dey, Sounak and Biten, Ali Furkan and Gomez, Lluis and Karatzas, Dimosthenis},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2021},
  pages     = {4023-4033},
  url       = {https://mlanthology.org/wacv/2021/mafla2021wacv-multimodal/}
}