Gradient Descent: Robustness to Adversarial Corruption

Abstract

Optimization using gradient descent (GD) is a ubiquitous practice in various machine learning problems including training large neural networks. Noise-free GD and stochastic GD--corrupted by random noise--have been extensively studied in the literature, but less attention has been paid to an adversarial setting, that is subject to adversarial corruptions in the gradient values. In this work, we analyze the performance of GD under a proposed general adversarial framework. For the class of functions satisfying the Polyak-Łojasiewicz condition, we derive finite time bounds on a minimax optimization error. Based on this bound, we provide a guideline on the choice of learning rate sequence with theoretical guarantees on the robustness of GD against adversarial corruption.

Cite

Text

Chang et al. "Gradient Descent: Robustness to Adversarial Corruption." NeurIPS 2022 Workshops: OPT, 2022.

Markdown

[Chang et al. "Gradient Descent: Robustness to Adversarial Corruption." NeurIPS 2022 Workshops: OPT, 2022.](https://mlanthology.org/neuripsw/2022/chang2022neuripsw-gradient/)

BibTeX

@inproceedings{chang2022neuripsw-gradient,
  title     = {{Gradient Descent: Robustness to Adversarial Corruption}},
  author    = {Chang, Fu-Chieh and Nabiei, Farhang and Wu, Pei-Yuan and Cioba, Alexandru and Vakili, Sattar and Bernacchia, Alberto},
  booktitle = {NeurIPS 2022 Workshops: OPT},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/chang2022neuripsw-gradient/}
}