A Theoretical Analysis of the Number of Shots in Few-Shot Learning

Abstract

Few-shot classification is the task of predicting the category of an example from a set of few labeled examples. The number of labeled examples per category is called the number of shots (or shot number). Recent works tackle this task through meta-learning, where a meta-learner extracts information from observed tasks during meta-training to quickly adapt to new tasks during meta-testing. In this formulation, the number of shots exploited during meta-training has an impact on the recognition performance at meta-test time. Generally, the shot number used in meta-training should match the one used in meta-testing to obtain the best performance. We introduce a theoretical analysis of the impact of the shot number on Prototypical Networks, a state-of-the-art few-shot classification method. From our analysis, we propose a simple method that is robust to the choice of shot number used during meta-training, which is a crucial hyperparameter. The performance of our model trained for an arbitrary meta-training shot number shows great performance for different values of meta-testing shot numbers. We experimentally demonstrate our approach on different few-shot classification benchmarks.

Cite

Text

Cao et al. "A Theoretical Analysis of the Number of Shots in Few-Shot Learning." International Conference on Learning Representations, 2020.

Markdown

[Cao et al. "A Theoretical Analysis of the Number of Shots in Few-Shot Learning." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/cao2020iclr-theoretical/)

BibTeX

@inproceedings{cao2020iclr-theoretical,
  title     = {{A Theoretical Analysis of the Number of Shots in Few-Shot Learning}},
  author    = {Cao, Tianshi and Law, Marc and Fidler, Sanja},
  booktitle = {International Conference on Learning Representations},
  year      = {2020},
  url       = {https://mlanthology.org/iclr/2020/cao2020iclr-theoretical/}
}