Riemannian Accelerated Gradient Methods via Extrapolation

Abstract

In this paper, we propose a convergence acceleration scheme for general Riemannian optimization problems by extrapolating iterates on manifolds. We show that when the iterates are generated from the Riemannian gradient descent method, the scheme achieves the optimal convergence rate asymptotically and is computationally more favorable than the recently proposed Riemannian Nesterov accelerated gradient methods. A salient feature of our analysis is the convergence guarantees with respect to the use of general retraction and vector transport. Empirically, we verify the practical benefits of the proposed acceleration strategy, including robustness to the choice of different averaging schemes on manifolds.

Cite

Text

Han et al. "Riemannian Accelerated Gradient Methods via Extrapolation." Artificial Intelligence and Statistics, 2023.

Markdown

[Han et al. "Riemannian Accelerated Gradient Methods via Extrapolation." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/han2023aistats-riemannian/)

BibTeX

@inproceedings{han2023aistats-riemannian,
  title     = {{Riemannian Accelerated Gradient Methods via Extrapolation}},
  author    = {Han, Andi and Mishra, Bamdev and Jawanpuria, Pratik and Gao, Junbin},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2023},
  pages     = {1554-1585},
  volume    = {206},
  url       = {https://mlanthology.org/aistats/2023/han2023aistats-riemannian/}
}