Sequential Neural Models with Stochastic Layers

Abstract

How can we efficiently propagate uncertainty in a latent state representation with recurrent neural networks? This paper introduces stochastic recurrent neural networks which glue a deterministic recurrent neural network and a state space model together to form a stochastic and sequential neural generative model. The clear separation of deterministic and stochastic layers allows a structured variational inference network to track the factorization of the model’s posterior distribution. By retaining both the nonlinear recursive structure of a recurrent neural network and averaging over the uncertainty in a latent path, like a state space model, we improve the state of the art results on the Blizzard and TIMIT speech modeling data sets by a large margin, while achieving comparable performances to competing methods on polyphonic music modeling.

Cite

Text

Fraccaro et al. "Sequential Neural Models with Stochastic Layers." Neural Information Processing Systems, 2016.

Markdown

[Fraccaro et al. "Sequential Neural Models with Stochastic Layers." Neural Information Processing Systems, 2016.](https://mlanthology.org/neurips/2016/fraccaro2016neurips-sequential/)

BibTeX

@inproceedings{fraccaro2016neurips-sequential,
  title     = {{Sequential Neural Models with Stochastic Layers}},
  author    = {Fraccaro, Marco and Sønderby, Søren Kaae and Paquet, Ulrich and Winther, Ole},
  booktitle = {Neural Information Processing Systems},
  year      = {2016},
  pages     = {2199-2207},
  url       = {https://mlanthology.org/neurips/2016/fraccaro2016neurips-sequential/}
}