Accelerated Gradient Descent: A Guaranteed Bound for a Heuristic Restart Strategy

Abstract

The $\mathcal{O}(1/k^2)$ convergence rate in function value of accelerated gradient descent is optimal, but there are many modifications that have been used to speed up convergence in practice. Among these modifications are restarts, that is, starting the algorithm anew with the current iteration being considered as the initial point. We focus on the adaptive restart techniques introduced by O'Donoghue and Candes, specifically their gradient restart strategy. While the gradient restart strategy is a heuristic in general, we prove that applying gradient restarts preserves and in fact improves the $\mathcal{O}(1/k^2)$ bound, hence establishing function value convergence, for one-dimensional functions. Applications of our results to separable and nearly separable functions are presented.

Cite

Text

Moursi et al. "Accelerated Gradient Descent: A Guaranteed Bound for a Heuristic Restart Strategy." NeurIPS 2023 Workshops: OPT, 2023.

Markdown

[Moursi et al. "Accelerated Gradient Descent: A Guaranteed Bound for a Heuristic Restart Strategy." NeurIPS 2023 Workshops: OPT, 2023.](https://mlanthology.org/neuripsw/2023/moursi2023neuripsw-accelerated/)

BibTeX

@inproceedings{moursi2023neuripsw-accelerated,
  title     = {{Accelerated Gradient Descent: A Guaranteed Bound for a Heuristic Restart Strategy}},
  author    = {Moursi, Walaa and Vavasis, Stephen A. and Pavlovic, Viktor},
  booktitle = {NeurIPS 2023 Workshops: OPT},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/moursi2023neuripsw-accelerated/}
}