Anderson Acceleration of Proximal Gradient Methods

Abstract

Anderson acceleration is a well-established and simple technique for speeding up fixed-point computations with countless applications. This work introduces novel methods for adapting Anderson acceleration to proximal gradient algorithms. Under some technical conditions, we extend existing local convergence results of Anderson acceleration for smooth fixed-point mappings to the proposed non-smooth setting. We also prove analytically that it is in general, impossible to guarantee global convergence of native Anderson acceleration. We therefore propose a simple scheme for stabilization that combines the global worst-case guarantees of proximal gradient methods with the local adaptation and practical speed-up of Anderson acceleration. Finally, we provide the first applications of Anderson acceleration to non-Euclidean geometry.

Cite

Text

Mai and Johansson. "Anderson Acceleration of Proximal Gradient Methods." International Conference on Machine Learning, 2020.

Markdown

[Mai and Johansson. "Anderson Acceleration of Proximal Gradient Methods." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/mai2020icml-anderson/)

BibTeX

@inproceedings{mai2020icml-anderson,
  title     = {{Anderson Acceleration of Proximal Gradient Methods}},
  author    = {Mai, Vien and Johansson, Mikael},
  booktitle = {International Conference on Machine Learning},
  year      = {2020},
  pages     = {6620-6629},
  volume    = {119},
  url       = {https://mlanthology.org/icml/2020/mai2020icml-anderson/}
}