Online Control for Meta-Optimization

Abstract

Choosing the optimal hyperparameters, including learning rate and momentum, for specific optimization instances is a significant yet non-convex challenge. This makes conventional iterative techniques such as hypergradient descent \cite{baydin2017online} insufficient in obtaining global optimality guarantees.We consider the more general task of meta-optimization -- online learning of the best optimization algorithm given problem instances, and introduce a novel approach based on control theory. We show how meta-optimization can be formulated as an optimal control problem, departing from existing literature that use stability-based methods to study optimization. Our approach leverages convex relaxation techniques in the recently-proposed nonstochastic control framework to overcome the challenge of nonconvexity, and obtains regret guarantees vs. the best offline solution. This guarantees that in meta-optimization, we can learn a method that attains convergence comparable to that of the best optimization method in hindsight from a class of methods.

Cite

Text

Chen and Hazan. "Online Control for Meta-Optimization." Neural Information Processing Systems, 2023.

Markdown

[Chen and Hazan. "Online Control for Meta-Optimization." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/chen2023neurips-online/)

BibTeX

@inproceedings{chen2023neurips-online,
  title     = {{Online Control for Meta-Optimization}},
  author    = {Chen, Xinyi and Hazan, Elad},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/chen2023neurips-online/}
}