Maximizing Entropy on Adversarial Examples Can Improve Generalization

Abstract

Supervised classification methods that directly optimize maximize the likelihood of the training data often overfit. This overfitting is typically mitigated through regularizing the loss function (e.g., label smoothing, weight decay) or by minimizing the same loss on new examples (e.g., data augmentation, adversarial training). In this work, we propose a complementary regularization strategy: training the model to be unconfident on examples that are generated so they have unclear labels. We call our approach Maximum Predictive Entropy (MPE). These automatically generated examples are cheap to compute, so our method is only 30% slower than standard data augmentation. Adding MPE to existing regularization techniques, such as label smoothing, increases test accuracy by 1-3%, with larger gains in the small data regime.

Cite

Text

Setlur et al. "Maximizing Entropy on Adversarial Examples Can Improve Generalization." ICLR 2022 Workshops: PAIR2Struct, 2022.

Markdown

[Setlur et al. "Maximizing Entropy on Adversarial Examples Can Improve Generalization." ICLR 2022 Workshops: PAIR2Struct, 2022.](https://mlanthology.org/iclrw/2022/setlur2022iclrw-maximizing/)

BibTeX

@inproceedings{setlur2022iclrw-maximizing,
  title     = {{Maximizing Entropy on Adversarial Examples Can Improve Generalization}},
  author    = {Setlur, Amrith and Eysenbach, Benjamin and Smith, Virginia and Levine, Sergey},
  booktitle = {ICLR 2022 Workshops: PAIR2Struct},
  year      = {2022},
  url       = {https://mlanthology.org/iclrw/2022/setlur2022iclrw-maximizing/}
}