Label Noise: Ignorance Is Bliss

Abstract

We establish a new theoretical framework for learning under multi-class, instance-dependent label noise. At the heart of our framework is the concept of \emph{relative signal strength} (RSS), which is a point-wise measure of noisiness. We use relative signal strength to establish matching upper and lower bounds for excess risk. Our theoretical findings reveal a surprising result: the extremely simple \emph{Noise Ignorant Empirical Risk Minimization (NI-ERM)} principle, which conducts empirical risk minimization as if no label noise exists, is minimax optimal. Finally, we translate these theoretical insights into practice: by using NI-ERM to fit a linear classifier on top of a frozen foundation model, we achieve state-of-the-art performance on the CIFAR-N data challenge.

Cite

Text

Zhu et al. "Label Noise: Ignorance Is Bliss." NeurIPS 2024 Workshops: M3L, 2024.

Markdown

[Zhu et al. "Label Noise: Ignorance Is Bliss." NeurIPS 2024 Workshops: M3L, 2024.](https://mlanthology.org/neuripsw/2024/zhu2024neuripsw-label/)

BibTeX

@inproceedings{zhu2024neuripsw-label,
  title     = {{Label Noise: Ignorance Is Bliss}},
  author    = {Zhu, Yilun and Zhang, Jianxin and Gangrade, Aditya and Scott, Clayton},
  booktitle = {NeurIPS 2024 Workshops: M3L},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/zhu2024neuripsw-label/}
}