Continually Adapting Optimizers Improve Meta-Generalization
Abstract
Meta-learned optimizers increasingly outperform analytical handcrafted optimizers such as SGD and Adam. On some tasks, however, they fail to generalize strongly, underperforming handcrafted methods. Then one can fall back on handcrafted methods through a guard, to combine the efficiency benefits of learned optimizers and the guarantees of analytical methods. At some point in the iterative optimization process, however, such guards may make the learned optimizer incompatible with the remaining optimization, and thus useless for further progress. Our novel method Meta Guard keeps adapting the learned optimizer to the target optimization problem. It experimentally outperforms other baselines, adapting to new tasks during training.
Cite
Text
Wang et al. "Continually Adapting Optimizers Improve Meta-Generalization." NeurIPS 2023 Workshops: DistShift, 2023.Markdown
[Wang et al. "Continually Adapting Optimizers Improve Meta-Generalization." NeurIPS 2023 Workshops: DistShift, 2023.](https://mlanthology.org/neuripsw/2023/wang2023neuripsw-continually/)BibTeX
@inproceedings{wang2023neuripsw-continually,
title = {{Continually Adapting Optimizers Improve Meta-Generalization}},
author = {Wang, Wenyi and Kirsch, Louis and Faccio, Francesco and Zhuge, Mingchen and Schmidhuber, Jürgen},
booktitle = {NeurIPS 2023 Workshops: DistShift},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/wang2023neuripsw-continually/}
}