Malign Overfitting: Interpolation and Invariance Are Fundamentally at Odds

Abstract

Learned classifiers should often possess certain invariance properties meant to encourage fairness, robustness, or out-of-distribution generalization. However, multiple recent works empirically demonstrate that common invariance-inducing regularizers are ineffective in the over-parameterized regime, in which classifiers perfectly fit (i.e. interpolate) the training data. This suggests that the phenomenon of ``benign overfitting", in which models generalize well despite interpolating, might not favorably extend to settings in which robustness or fairness are desirable. In this work we provide a theoretical justification for these observations. We prove that - even in the simplest of settings - any interpolating classifier (with nonzero margin) will not satisfy these invariance properties. We then propose and analyze an algorithm that - in the same setting - successfully learns a non-interpolating classifier that is provably invariant. We validate our theoretical observations regarding the conflict between interpolation and invariance on simulated data and the Waterbirds dataset.

Cite

Text

Wald et al. "Malign Overfitting: Interpolation and Invariance Are Fundamentally at Odds." NeurIPS 2022 Workshops: DistShift, 2022.

Markdown

[Wald et al. "Malign Overfitting: Interpolation and Invariance Are Fundamentally at Odds." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/wald2022neuripsw-malign/)

BibTeX

@inproceedings{wald2022neuripsw-malign,
  title     = {{Malign Overfitting: Interpolation and Invariance Are Fundamentally at Odds}},
  author    = {Wald, Yoav and Yona, Gal and Shalit, Uri and Carmon, Yair},
  booktitle = {NeurIPS 2022 Workshops: DistShift},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/wald2022neuripsw-malign/}
}