Pre-Training of Recurrent Neural Networks via Linear Autoencoders

Abstract

We propose a pre-training technique for recurrent neural networks based on linear autoencoder networks for sequences, i.e. linear dynamical systems modelling the target sequences. We start by giving a closed form solution for the definition of the optimal weights of a linear autoencoder given a training set of sequences. This solution, however, is computationally very demanding, so we suggest a procedure to get an approximate solution for a given number of hidden units. The weights obtained for the linear autoencoder are then used as initial weights for the input-to-hidden connections of a recurrent neural network, which is then trained on the desired task. Using four well known datasets of sequences of polyphonic music, we show that the proposed pre-training approach is highly effective, since it allows to largely improve the state of the art results on all the considered datasets.

Cite

Text

Pasa and Sperduti. "Pre-Training of Recurrent Neural Networks via Linear Autoencoders." Neural Information Processing Systems, 2014.

Markdown

[Pasa and Sperduti. "Pre-Training of Recurrent Neural Networks via Linear Autoencoders." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/pasa2014neurips-pretraining/)

BibTeX

@inproceedings{pasa2014neurips-pretraining,
  title     = {{Pre-Training of Recurrent Neural Networks via Linear Autoencoders}},
  author    = {Pasa, Luca and Sperduti, Alessandro},
  booktitle = {Neural Information Processing Systems},
  year      = {2014},
  pages     = {3572-3580},
  url       = {https://mlanthology.org/neurips/2014/pasa2014neurips-pretraining/}
}