Long-Tail Recognition via Compositional Knowledge Transfer

Abstract

In this work, we introduce a novel strategy for long-tail recognition that addresses the tail classes' few-shot problem via training-free knowledge transfer. Our objective is to transfer knowledge acquired from information-rich common classes to semantically similar, and yet data-hungry, rare classes in order to obtain stronger tail class representations. We leverage the fact that class prototypes and learned cosine classifiers provide two different, complementary representations of class cluster centres in feature space, and use an attention mechanism to select and recompose learned classifiers features from common classes to obtain higher quality rare class representations. Our knowledge transfer process is training free, reducing overfitting risks, and can afford continual extension of classifiers to new classes. Experiments show that our approach can achieve significant performance boosts on rare classes while maintaining robust common class performance, outperforming directly comparable state-of-the-art models.

Cite

Text

Parisot et al. "Long-Tail Recognition via Compositional Knowledge Transfer." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00681

Markdown

[Parisot et al. "Long-Tail Recognition via Compositional Knowledge Transfer." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/parisot2022cvpr-longtail/) doi:10.1109/CVPR52688.2022.00681

BibTeX

@inproceedings{parisot2022cvpr-longtail,
  title     = {{Long-Tail Recognition via Compositional Knowledge Transfer}},
  author    = {Parisot, Sarah and Esperança, Pedro M. and McDonagh, Steven and Madarasz, Tamas J. and Yang, Yongxin and Li, Zhenguo},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {6939-6948},
  doi       = {10.1109/CVPR52688.2022.00681},
  url       = {https://mlanthology.org/cvpr/2022/parisot2022cvpr-longtail/}
}