Scalable Second-Order Optimization Algorithms for Minimizing Low-Rank Functions

Abstract

We present a random-subspace variant of cubic regularization algorithm that chooses the size of the subspace adaptively, based on the rank of the projected second derivative matrix. Iteratively, our variant only requires access to (small-dimensional) projections of first- and second-order problem derivatives and calculates a reduced step inexpensively. The ensuing method maintains the optimal global rate of convergence of (full-dimensional) cubic regularization, while showing improved scalability both theoretically and numerically, particularly when applied to low-rank functions. When applied to the latter, our algorithm naturally adapts the subspace size to the true rank of the function, without knowing it a priori.

Cite

Text

Tansley and Cartis. "Scalable Second-Order Optimization Algorithms for Minimizing Low-Rank Functions." NeurIPS 2024 Workshops: OPT, 2024.

Markdown

[Tansley and Cartis. "Scalable Second-Order Optimization Algorithms for Minimizing Low-Rank Functions." NeurIPS 2024 Workshops: OPT, 2024.](https://mlanthology.org/neuripsw/2024/tansley2024neuripsw-scalable/)

BibTeX

@inproceedings{tansley2024neuripsw-scalable,
  title     = {{Scalable Second-Order Optimization Algorithms for Minimizing Low-Rank Functions}},
  author    = {Tansley, Edward and Cartis, Coralia},
  booktitle = {NeurIPS 2024 Workshops: OPT},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/tansley2024neuripsw-scalable/}
}