Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives
Abstract
We analyze the convergence rate of various momentum-based optimization algorithms from a dynamical systems point of view. Our analysis exploits fundamental topological properties, such as the continuous dependence of iterates on their initial conditions, to provide a simple characterization of convergence rates. In many cases, closed-form expressions are obtained that relate algorithm parameters to the convergence rate. The analysis encompasses discrete time and continuous time, as well as time-invariant and time-variant formulations, and is not limited to a convex or Euclidean setting. In addition, the article rigorously establishes why symplectic discretization schemes are important for momentum-based optimization algorithms, and provides a characterization of algorithms that exhibit accelerated convergence.
Cite
Text
Muehlebach and Jordan. "Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives." Journal of Machine Learning Research, 2021.Markdown
[Muehlebach and Jordan. "Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/muehlebach2021jmlr-optimization/)BibTeX
@article{muehlebach2021jmlr-optimization,
title = {{Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives}},
author = {Muehlebach, Michael and Jordan, Michael I.},
journal = {Journal of Machine Learning Research},
year = {2021},
pages = {1-50},
volume = {22},
url = {https://mlanthology.org/jmlr/2021/muehlebach2021jmlr-optimization/}
}