Acceleration Through Spectral Density Estimation
Abstract
We develop a framework for the average-case analysis of random quadratic problems and derive algorithms that are optimal under this analysis. This yields a new class of methods that achieve acceleration given a model of the Hessian’s eigenvalue distribution. We develop explicit algorithms for the uniform, Marchenko-Pastur, and exponential distributions. These methods have a simple momentum-like update, in which each update only makes use on the current gradient and previous two iterates. Furthermore, the momentum and step-size parameters can be estimated without knowledge of the Hessian’s smallest singular value, in contrast with classical accelerated methods like Nesterov acceleration and Polyak momentum. Through empirical benchmarks on quadratic and logistic regression problems, we identify regimes in which the the proposed methods improve over classical (worst-case) accelerated methods.
Cite
Text
Pedregosa and Scieur. "Acceleration Through Spectral Density Estimation." International Conference on Machine Learning, 2020.Markdown
[Pedregosa and Scieur. "Acceleration Through Spectral Density Estimation." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/pedregosa2020icml-acceleration/)BibTeX
@inproceedings{pedregosa2020icml-acceleration,
title = {{Acceleration Through Spectral Density Estimation}},
author = {Pedregosa, Fabian and Scieur, Damien},
booktitle = {International Conference on Machine Learning},
year = {2020},
pages = {7553-7562},
volume = {119},
url = {https://mlanthology.org/icml/2020/pedregosa2020icml-acceleration/}
}