Latent Matters: Learning Deep State-Space Models
Abstract
Deep state-space models (DSSMs) enable temporal predictions by learning the underlying dynamics of observed sequence data. They are often trained by maximising the evidence lower bound. However, as we show, this does not ensure the model actually learns the underlying dynamics. We therefore propose a constrained optimisation framework as a general approach for training DSSMs. Building upon this, we introduce the extended Kalman VAE (EKVAE), which combines amortised variational inference with classic Bayesian filtering/smoothing to model dynamics more accurately than RNN-based DSSMs. Our results show that the constrained optimisation framework significantly improves system identification and prediction accuracy on the example of established state-of-the-art DSSMs. The EKVAE outperforms previous models w.r.t. prediction accuracy, achieves remarkable results in identifying dynamical systems, and can furthermore successfully learn state-space representations where static and dynamic features are disentangled.
Cite
Text
Klushyn et al. "Latent Matters: Learning Deep State-Space Models." Neural Information Processing Systems, 2021.Markdown
[Klushyn et al. "Latent Matters: Learning Deep State-Space Models." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/klushyn2021neurips-latent/)BibTeX
@inproceedings{klushyn2021neurips-latent,
title = {{Latent Matters: Learning Deep State-Space Models}},
author = {Klushyn, Alexej and Kurle, Richard and Soelch, Maximilian and Cseke, Botond and van der Smagt, Patrick},
booktitle = {Neural Information Processing Systems},
year = {2021},
url = {https://mlanthology.org/neurips/2021/klushyn2021neurips-latent/}
}