Few-Shot Learning with Online Self-Distillation
Abstract
Few-shot learning has been a long-standing problem in learning to learn. This problem typically involves training a model on an extremely small amount of data and testing the model on the out-of-distribution data. The focus of recent few-shot learning research has been on the development of good representation models that can quickly adapt to test tasks. To that end, we come up with a model that learns representation through online self-distillation. Our model combines supervised training with knowledge distillation via a continuously updated teacher. We also identify that data augmentation plays an important role in producing robust features. Our final model is trained with CutMix augmentation and online self-distillation. On the commonly used benchmark miniImageNet, our model achieves 67.07% and 83.03% under the 5-way 1-shot setting and the 5-way 5-shot setting, respectively. It outperforms counterparts of its kind by 2.25% and 0.89%.
Cite
Text
Liu and Wang. "Few-Shot Learning with Online Self-Distillation." IEEE/CVF International Conference on Computer Vision Workshops, 2021. doi:10.1109/ICCVW54120.2021.00124Markdown
[Liu and Wang. "Few-Shot Learning with Online Self-Distillation." IEEE/CVF International Conference on Computer Vision Workshops, 2021.](https://mlanthology.org/iccvw/2021/liu2021iccvw-fewshot/) doi:10.1109/ICCVW54120.2021.00124BibTeX
@inproceedings{liu2021iccvw-fewshot,
title = {{Few-Shot Learning with Online Self-Distillation}},
author = {Liu, Sihan and Wang, Yue},
booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
year = {2021},
pages = {1067-1070},
doi = {10.1109/ICCVW54120.2021.00124},
url = {https://mlanthology.org/iccvw/2021/liu2021iccvw-fewshot/}
}