DiffuLT: Diffusion for Long-Tail Recognition Without External Knowledge
Abstract
This paper introduces a novel pipeline for long-tail (LT) recognition that diverges from conventional strategies. Instead, it leverages the long-tailed dataset itself to generate a balanced proxy dataset without utilizing external data or model. We deploy a diffusion model trained from scratch on only the long-tailed dataset to create this proxy and verify the effectiveness of the data produced. Our analysis identifies approximately-in-distribution (AID) samples, which slightly deviate from the real data distribution and incorporate a blend of class information, as the crucial samples for enhancing the generative model's performance in long-tail classification. We promote the generation of AID samples during the training of a generative model by utilizing a feature extractor to guide the process and filter out detrimental samples during generation. Our approach, termed Diffusion model for Long-Tail recognition (DiffuLT), represents a pioneer application of generative models in long-tail recognition. DiffuLT achieves state-of-the-art results on CIFAR10-LT, CIFAR100-LT, and ImageNet-LT, surpassing leading competitors by significant margins. Comprehensive ablations enhance the interpretability of our pipeline. Notably, the entire generative process is conducted without relying on external data or pre-trained model weights, which leads to its generalizability to real-world long-tailed scenarios.
Cite
Text
Shao et al. "DiffuLT: Diffusion for Long-Tail Recognition Without External Knowledge." Neural Information Processing Systems, 2024. doi:10.52202/079017-3909Markdown
[Shao et al. "DiffuLT: Diffusion for Long-Tail Recognition Without External Knowledge." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/shao2024neurips-diffult/) doi:10.52202/079017-3909BibTeX
@inproceedings{shao2024neurips-diffult,
title = {{DiffuLT: Diffusion for Long-Tail Recognition Without External Knowledge}},
author = {Shao, Jie and Zhu, Ke and Zhang, Hanxiao and Wu, Jianxin},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-3909},
url = {https://mlanthology.org/neurips/2024/shao2024neurips-diffult/}
}