Neural Optimizer Search with Reinforcement Learning
Abstract
We present an approach to automate the process of discovering optimization methods, with a focus on deep learning architectures. We train a Recurrent Neural Network controller to generate a string in a specific domain language that describes a mathematical update equation based on a list of primitive functions, such as the gradient, running average of the gradient, etc. The controller is trained with Reinforcement Learning to maximize the performance of a model after a few epochs. On CIFAR-10, our method discovers several update rules that are better than many commonly used optimizers, such as Adam, RMSProp, or SGD with and without Momentum on a ConvNet model. These optimizers can also be transferred to perform well on different neural network architectures, including Google’s neural machine translation system.
Cite
Text
Bello et al. "Neural Optimizer Search with Reinforcement Learning." International Conference on Machine Learning, 2017.Markdown
[Bello et al. "Neural Optimizer Search with Reinforcement Learning." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/bello2017icml-neural/)BibTeX
@inproceedings{bello2017icml-neural,
title = {{Neural Optimizer Search with Reinforcement Learning}},
author = {Bello, Irwan and Zoph, Barret and Vasudevan, Vijay and Le, Quoc V.},
booktitle = {International Conference on Machine Learning},
year = {2017},
pages = {459-468},
volume = {70},
url = {https://mlanthology.org/icml/2017/bello2017icml-neural/}
}