Adversarial Training for Free!

Abstract

Adversarial training, in which a network is trained on adversarial examples, is one of the few defenses against adversarial attacks that withstands strong attacks. Unfortunately, the high cost of generating strong adversarial examples makes standard adversarial training impractical on large-scale problems like ImageNet. We present an algorithm that eliminates the overhead cost of generating adversarial examples by recycling the gradient information computed when updating model parameters. Our "free" adversarial training algorithm achieves comparable robustness to PGD adversarial training on the CIFAR-10 and CIFAR-100 datasets at negligible additional cost compared to natural training, and can be 7 to 30 times faster than other strong adversarial training methods. Using a single workstation with 4 P100 GPUs and 2 days of runtime, we can train a robust model for the large-scale ImageNet classification task that maintains 40% accuracy against PGD attacks.

Cite

Text

Shafahi et al. "Adversarial Training for Free!." Neural Information Processing Systems, 2019.

Markdown

[Shafahi et al. "Adversarial Training for Free!." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/shafahi2019neurips-adversarial/)

BibTeX

@inproceedings{shafahi2019neurips-adversarial,
  title     = {{Adversarial Training for Free!}},
  author    = {Shafahi, Ali and Najibi, Mahyar and Ghiasi, Mohammad Amin and Xu, Zheng and Dickerson, John and Studer, Christoph and Davis, Larry S. and Taylor, Gavin and Goldstein, Tom},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {3358-3369},
  url       = {https://mlanthology.org/neurips/2019/shafahi2019neurips-adversarial/}
}