Towards Faster Stochastic Gradient Search

Abstract

Stochastic gradient descent is a general algorithm which includes LMS, on-line backpropagation, and adaptive k-means clustering as special cases. The standard choices of the learning rate 1] (both adaptive and fixed func(cid:173) tions of time) often perform quite poorly. In contrast, our recently pro(cid:173) posed class of "search then converge" learning rate schedules (Darken and Moody, 1990) display the theoretically optimal asymptotic convergence rate and a superior ability to escape from poor local minima. However, the user is responsible for setting a key parameter. We propose here a new method(cid:173) ology for creating the first completely automatic adaptive learning rates which achieve the optimal rate of convergence.

Cite

Text

Darken and Moody. "Towards Faster Stochastic Gradient Search." Neural Information Processing Systems, 1991.

Markdown

[Darken and Moody. "Towards Faster Stochastic Gradient Search." Neural Information Processing Systems, 1991.](https://mlanthology.org/neurips/1991/darken1991neurips-faster/)

BibTeX

@inproceedings{darken1991neurips-faster,
  title     = {{Towards Faster Stochastic Gradient Search}},
  author    = {Darken, Christian and Moody, John},
  booktitle = {Neural Information Processing Systems},
  year      = {1991},
  pages     = {1009-1016},
  url       = {https://mlanthology.org/neurips/1991/darken1991neurips-faster/}
}