Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning

Abstract

Few-shot class-incremental learning is to recognize the new classes given few samples and not forget the old classes. It is a challenging task since representation optimization and prototype reorganization can only be achieved under little supervision. To address this problem, we propose a novel incremental prototype learning scheme. Our scheme consists of a random episode selection strategy that adapts the feature representation to various generated incremental episodes to enhance the corresponding extensibility, and a self-promoted prototype refinement mechanism which strengthens the expression ability of the new class by explicitly considering the dependencies among different classes. Particularly, a dynamic relation projection module is proposed to calculate the relation matrix in a shared embedding space and leverage it as the factor for bootstrapping the update of prototypes. Extensive experiments on three benchmark datasets demonstrate the above-par incremental performance, outperforming state-of-the-art methods by a margin of 13%, 17% and 11%, respectively.

Cite

Text

Zhu et al. "Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00673

Markdown

[Zhu et al. "Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/zhu2021cvpr-selfpromoted/) doi:10.1109/CVPR46437.2021.00673

BibTeX

@inproceedings{zhu2021cvpr-selfpromoted,
  title     = {{Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning}},
  author    = {Zhu, Kai and Cao, Yang and Zhai, Wei and Cheng, Jie and Zha, Zheng-Jun},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {6801-6810},
  doi       = {10.1109/CVPR46437.2021.00673},
  url       = {https://mlanthology.org/cvpr/2021/zhu2021cvpr-selfpromoted/}
}