Fine-Grained Image Classification by Visual-Semantic Embedding
Abstract
This paper investigates a challenging problem,which is known as fine-grained image classification(FGIC). Different from conventional computer visionproblems, FGIC suffers from the large intraclassdiversities and subtle inter-class differences.Existing FGIC approaches are limited to exploreonly the visual information embedded in the images.In this paper, we present a novel approachwhich can use handy prior knowledge from eitherstructured knowledge bases or unstructured text tofacilitate FGIC. Specifically, we propose a visual-semanticembedding model which explores semanticembedding from knowledge bases and text, andfurther trains a novel end-to-end CNN frameworkto linearly map image features to a rich semanticembedding space. Experimental results on a challenginglarge-scale UCSD Bird-200-2011 datasetverify that our approach outperforms several state-of-the-art methods with significant advances.
Cite
Text
Xu et al. "Fine-Grained Image Classification by Visual-Semantic Embedding." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/145Markdown
[Xu et al. "Fine-Grained Image Classification by Visual-Semantic Embedding." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/xu2018ijcai-fine/) doi:10.24963/IJCAI.2018/145BibTeX
@inproceedings{xu2018ijcai-fine,
title = {{Fine-Grained Image Classification by Visual-Semantic Embedding}},
author = {Xu, Huapeng and Qi, Guilin and Li, Jingjing and Wang, Meng and Xu, Kang and Gao, Huan},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {1043-1049},
doi = {10.24963/IJCAI.2018/145},
url = {https://mlanthology.org/ijcai/2018/xu2018ijcai-fine/}
}