Distilling Virtual Examples for Long-Tailed Recognition

Abstract

We tackle the long-tailed visual recognition problem from the knowledge distillation perspective by proposing a Distill the Virtual Examples (DiVE) method. Specifically, by treating the predictions of a teacher model as virtual exam- ples, we prove that distilling from these virtual examples is equivalent to label distribution learning under certain con- straints. We show that when the virtual example distribu- tion becomes flatter than the original input distribution, the under-represented tail classes will receive significant im- provements, which is crucial in long-tailed recognition. The proposed DiVE method can explicitly tune the virtual exam- ple distribution to become flat. Extensive experiments on three benchmark datasets, including the large-scale iNat- uralist ones, justify that the proposed DiVE method can significantly outperform state-of-the-art methods. Further- more, additional analyses and experiments verify the virtual example interpretation, and demonstrate the effectiveness of tailored designs in DiVE for long-tailed problems.

Cite

Text

He et al. "Distilling Virtual Examples for Long-Tailed Recognition." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.00030

Markdown

[He et al. "Distilling Virtual Examples for Long-Tailed Recognition." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/he2021iccv-distilling/) doi:10.1109/ICCV48922.2021.00030

BibTeX

@inproceedings{he2021iccv-distilling,
  title     = {{Distilling Virtual Examples for Long-Tailed Recognition}},
  author    = {He, Yin-Yin and Wu, Jianxin and Wei, Xiu-Shen},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {235-244},
  doi       = {10.1109/ICCV48922.2021.00030},
  url       = {https://mlanthology.org/iccv/2021/he2021iccv-distilling/}
}