Riemannian Accelerated Gradient Methods via Extrapolation
Abstract
In this paper, we propose a convergence acceleration scheme for general Riemannian optimization problems by extrapolating iterates on manifolds. We show that when the iterates are generated from the Riemannian gradient descent method, the scheme achieves the optimal convergence rate asymptotically and is computationally more favorable than the recently proposed Riemannian Nesterov accelerated gradient methods. A salient feature of our analysis is the convergence guarantees with respect to the use of general retraction and vector transport. Empirically, we verify the practical benefits of the proposed acceleration strategy, including robustness to the choice of different averaging schemes on manifolds.
Cite
Text
Han et al. "Riemannian Accelerated Gradient Methods via Extrapolation." Artificial Intelligence and Statistics, 2023.Markdown
[Han et al. "Riemannian Accelerated Gradient Methods via Extrapolation." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/han2023aistats-riemannian/)BibTeX
@inproceedings{han2023aistats-riemannian,
title = {{Riemannian Accelerated Gradient Methods via Extrapolation}},
author = {Han, Andi and Mishra, Bamdev and Jawanpuria, Pratik and Gao, Junbin},
booktitle = {Artificial Intelligence and Statistics},
year = {2023},
pages = {1554-1585},
volume = {206},
url = {https://mlanthology.org/aistats/2023/han2023aistats-riemannian/}
}