Level Set Teleportation: The Good, the Bad, and the Ugly
Abstract
We study level set teleportation, a sub-routine which seeks to accelerate gradient methods by maximizing the gradient over the set of parameters with the same objective value. Since the descent lemma implies that gradient descent (GD) decreases the objective proportional to the gradient-norm, level-set teleportation maximizes guaranteed one-step progress. We prove level-set teleportation neither improves nor worsens the convergence of GD for strongly convex functions, while for convex functions teleportation can arbitrarily increase the distance to the global minima. To solve teleportation problems, we develop a projected-gradient-type method requiring only Hessian-vector products; we use our method to show that initializing GD with teleportation slightly under-performs standard initializations for both convex and non-convex optimization problems. As a result, we report a mixed picture: teleportation can be efficiently evaluated, but appears to offer marginal gains.
Cite
Text
Mishkin et al. "Level Set Teleportation: The Good, the Bad, and the Ugly." NeurIPS 2023 Workshops: OPT, 2023.Markdown
[Mishkin et al. "Level Set Teleportation: The Good, the Bad, and the Ugly." NeurIPS 2023 Workshops: OPT, 2023.](https://mlanthology.org/neuripsw/2023/mishkin2023neuripsw-level/)BibTeX
@inproceedings{mishkin2023neuripsw-level,
title = {{Level Set Teleportation: The Good, the Bad, and the Ugly}},
author = {Mishkin, Aaron and Bietti, Alberto and Gower, Robert M.},
booktitle = {NeurIPS 2023 Workshops: OPT},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/mishkin2023neuripsw-level/}
}