Fine-Grained Image Classification by Visual-Semantic Embedding

Abstract

This paper investigates a challenging problem,which is known as fine-grained image classification(FGIC). Different from conventional computer visionproblems, FGIC suffers from the large intraclassdiversities and subtle inter-class differences.Existing FGIC approaches are limited to exploreonly the visual information embedded in the images.In this paper, we present a novel approachwhich can use handy prior knowledge from eitherstructured knowledge bases or unstructured text tofacilitate FGIC. Specifically, we propose a visual-semanticembedding model which explores semanticembedding from knowledge bases and text, andfurther trains a novel end-to-end CNN frameworkto linearly map image features to a rich semanticembedding space. Experimental results on a challenginglarge-scale UCSD Bird-200-2011 datasetverify that our approach outperforms several state-of-the-art methods with significant advances.

Cite

Text

Xu et al. "Fine-Grained Image Classification by Visual-Semantic Embedding." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/145

Markdown

[Xu et al. "Fine-Grained Image Classification by Visual-Semantic Embedding." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/xu2018ijcai-fine/) doi:10.24963/IJCAI.2018/145

BibTeX

@inproceedings{xu2018ijcai-fine,
  title     = {{Fine-Grained Image Classification by Visual-Semantic Embedding}},
  author    = {Xu, Huapeng and Qi, Guilin and Li, Jingjing and Wang, Meng and Xu, Kang and Gao, Huan},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {1043-1049},
  doi       = {10.24963/IJCAI.2018/145},
  url       = {https://mlanthology.org/ijcai/2018/xu2018ijcai-fine/}
}