Deep Temporal Sigmoid Belief Networks for Sequence Modeling

Abstract

Deep dynamic generative models are developed to learn sequential dependencies in time-series data. The multi-layered model is designed by constructing a hierarchy of temporal sigmoid belief networks (TSBNs), defined as a sequential stack of sigmoid belief networks (SBNs). Each SBN has a contextual hidden state, inherited from the previous SBNs in the sequence, and is used to regulate its hidden bias. Scalable learning and inference algorithms are derived by introducing a recognition model that yields fast sampling from the variational posterior. This recognition model is trained jointly with the generative model, by maximizing its variational lower bound on the log-likelihood. Experimental results on bouncing balls, polyphonic music, motion capture, and text streams show that the proposed approach achieves state-of-the-art predictive performance, and has the capacity to synthesize various sequences.

Cite

Text

Gan et al. "Deep Temporal Sigmoid Belief Networks for Sequence Modeling." Neural Information Processing Systems, 2015.

Markdown

[Gan et al. "Deep Temporal Sigmoid Belief Networks for Sequence Modeling." Neural Information Processing Systems, 2015.](https://mlanthology.org/neurips/2015/gan2015neurips-deep/)

BibTeX

@inproceedings{gan2015neurips-deep,
  title     = {{Deep Temporal Sigmoid Belief Networks for Sequence Modeling}},
  author    = {Gan, Zhe and Li, Chunyuan and Henao, Ricardo and Carlson, David E and Carin, Lawrence},
  booktitle = {Neural Information Processing Systems},
  year      = {2015},
  pages     = {2467-2475},
  url       = {https://mlanthology.org/neurips/2015/gan2015neurips-deep/}
}