Overcoming Mean-Field Approximations in Recurrent Gaussian Process Models
Abstract
We identify a new variational inference scheme for dynamical systems whose transition function is modelled by a Gaussian process. Inference in this setting has either employed computationally intensive MCMC methods, or relied on factorisations of the variational posterior. As we demonstrate in our experiments, the factorisation between latent system states and transition function can lead to a miscalibrated posterior and to learning unnecessarily large noise terms. We eliminate this factorisation by explicitly modelling the dependence between state trajectories and the low-rank representation of our Gaussian process posterior. Samples of the latent states can then be tractably generated by conditioning on this representation. The method we obtain gives better predictive performance and more calibrated estimates of the transition function, yet maintains the same time and space complexities as mean-field methods.
Cite
Text
Ialongo et al. "Overcoming Mean-Field Approximations in Recurrent Gaussian Process Models." International Conference on Machine Learning, 2019.Markdown
[Ialongo et al. "Overcoming Mean-Field Approximations in Recurrent Gaussian Process Models." International Conference on Machine Learning, 2019.](https://mlanthology.org/icml/2019/ialongo2019icml-overcoming/)BibTeX
@inproceedings{ialongo2019icml-overcoming,
title = {{Overcoming Mean-Field Approximations in Recurrent Gaussian Process Models}},
author = {Ialongo, Alessandro Davide and Van Der Wilk, Mark and Hensman, James and Rasmussen, Carl Edward},
booktitle = {International Conference on Machine Learning},
year = {2019},
pages = {2931-2940},
volume = {97},
url = {https://mlanthology.org/icml/2019/ialongo2019icml-overcoming/}
}