Pre-Training of Recurrent Neural Networks via Linear Autoencoders
Abstract
We propose a pre-training technique for recurrent neural networks based on linear autoencoder networks for sequences, i.e. linear dynamical systems modelling the target sequences. We start by giving a closed form solution for the definition of the optimal weights of a linear autoencoder given a training set of sequences. This solution, however, is computationally very demanding, so we suggest a procedure to get an approximate solution for a given number of hidden units. The weights obtained for the linear autoencoder are then used as initial weights for the input-to-hidden connections of a recurrent neural network, which is then trained on the desired task. Using four well known datasets of sequences of polyphonic music, we show that the proposed pre-training approach is highly effective, since it allows to largely improve the state of the art results on all the considered datasets.
Cite
Text
Pasa and Sperduti. "Pre-Training of Recurrent Neural Networks via Linear Autoencoders." Neural Information Processing Systems, 2014.Markdown
[Pasa and Sperduti. "Pre-Training of Recurrent Neural Networks via Linear Autoencoders." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/pasa2014neurips-pretraining/)BibTeX
@inproceedings{pasa2014neurips-pretraining,
title = {{Pre-Training of Recurrent Neural Networks via Linear Autoencoders}},
author = {Pasa, Luca and Sperduti, Alessandro},
booktitle = {Neural Information Processing Systems},
year = {2014},
pages = {3572-3580},
url = {https://mlanthology.org/neurips/2014/pasa2014neurips-pretraining/}
}