A Unified Framework of Online Learning Algorithms for Training Recurrent Neural Networks

Abstract

We present a framework for compactly summarizing many recent results in efficient and/or biologically plausible online training of recurrent neural networks (RNN). The framework organizes algorithms according to several criteria: (a) past vs. future facing, (b) tensor structure, (c) stochastic vs. deterministic, and (d) closed form vs. numerical. These axes reveal latent conceptual connections among several recent advances in online learning. Furthermore, we provide novel mathematical intuitions for their degree of success. Testing these algorithms on two parametric task families shows that performances cluster according to our criteria. Although a similar clustering is also observed for pairwise gradient alignment, alignment with exact methods does not explain ultimate performance. This suggests the need for better comparison metrics.

Cite

Text

Marschall et al. "A Unified Framework of Online Learning Algorithms for Training Recurrent Neural Networks." Journal of Machine Learning Research, 2020.

Markdown

[Marschall et al. "A Unified Framework of Online Learning Algorithms for Training Recurrent Neural Networks." Journal of Machine Learning Research, 2020.](https://mlanthology.org/jmlr/2020/marschall2020jmlr-unified/)

BibTeX

@article{marschall2020jmlr-unified,
  title     = {{A Unified Framework of Online Learning Algorithms for Training Recurrent Neural Networks}},
  author    = {Marschall, Owen and Cho, Kyunghyun and Savin, Cristina},
  journal   = {Journal of Machine Learning Research},
  year      = {2020},
  pages     = {1-34},
  volume    = {21},
  url       = {https://mlanthology.org/jmlr/2020/marschall2020jmlr-unified/}
}