Fine-Grained Object Classification via Self-Supervised Pose Alignment
Abstract
Semantic patterns of fine-grained objects are determined by subtle appearance difference of local parts, which thus inspires a number of part-based methods. However, due to uncontrollable object poses in images, distinctive details carried by local regions can be spatially distributed or even self-occluded, leading to a large variation on object representation. For discounting pose variations, this paper proposes to learn a novel graph based object representation to reveal a global configuration of local parts for self-supervised pose alignment across classes, which is employed as an auxiliary feature regularization on a deep representation learning network. Moreover, a coarse-to-fine supervision together with the proposed pose-insensitive constraint on shallow-to-deep sub-networks encourages discriminative features in a curriculum learning manner. We evaluate our method on three popular fine-grained object classification benchmarks, consistently achieving the state-of-the-art performance. Source codes are available at https://github.com/yangxh11/P2P-Net.
Cite
Text
Yang et al. "Fine-Grained Object Classification via Self-Supervised Pose Alignment." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00725Markdown
[Yang et al. "Fine-Grained Object Classification via Self-Supervised Pose Alignment." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/yang2022cvpr-finegrained/) doi:10.1109/CVPR52688.2022.00725BibTeX
@inproceedings{yang2022cvpr-finegrained,
title = {{Fine-Grained Object Classification via Self-Supervised Pose Alignment}},
author = {Yang, Xuhui and Wang, Yaowei and Chen, Ke and Xu, Yong and Tian, Yonghong},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2022},
pages = {7399-7408},
doi = {10.1109/CVPR52688.2022.00725},
url = {https://mlanthology.org/cvpr/2022/yang2022cvpr-finegrained/}
}