Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives

Abstract

We analyze the convergence rate of various momentum-based optimization algorithms from a dynamical systems point of view. Our analysis exploits fundamental topological properties, such as the continuous dependence of iterates on their initial conditions, to provide a simple characterization of convergence rates. In many cases, closed-form expressions are obtained that relate algorithm parameters to the convergence rate. The analysis encompasses discrete time and continuous time, as well as time-invariant and time-variant formulations, and is not limited to a convex or Euclidean setting. In addition, the article rigorously establishes why symplectic discretization schemes are important for momentum-based optimization algorithms, and provides a characterization of algorithms that exhibit accelerated convergence.

Cite

Text

Muehlebach and Jordan. "Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives." Journal of Machine Learning Research, 2021.

Markdown

[Muehlebach and Jordan. "Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/muehlebach2021jmlr-optimization/)

BibTeX

@article{muehlebach2021jmlr-optimization,
  title     = {{Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives}},
  author    = {Muehlebach, Michael and Jordan, Michael I.},
  journal   = {Journal of Machine Learning Research},
  year      = {2021},
  pages     = {1-50},
  volume    = {22},
  url       = {https://mlanthology.org/jmlr/2021/muehlebach2021jmlr-optimization/}
}