Bayesian Transformed Gaussian Processes

Abstract

The Bayesian transformed Gaussian (BTG) model, proposed by Kedem and Oliviera in 1997, was developed as a Bayesian approach to trans-Kriging in the spatial statistics community. In this paper, we revisit BTG in the context of modern Gaussian process literature by framing it as a fully Bayesian counterpart to the Warped Gaussian process that marginalizes out a joint prior over input warping and kernel hyperparameters. As with any other fully Bayesian approach, this treatment introduces prohibitively expensive computational overhead; unsurprisingly, the BTG posterior predictive distribution, itself estimated through high-dimensional integration, must be inverted in order to perform model prediction. To address these challenges, we introduce principled numerical techniques for computing with BTG efficiently using a combination of doubly sparse quadrature rules, tight quantile bounds, and rank-one matrix algebra to enable both fast model prediction and model selection. These efficient methods allow us to compute with higher-dimensional datasets and apply BTG with layered transformations that greatly improve its expressibility. We demonstrate that BTG achieves superior empirical performance over MLE-based models in the low-data regime ---situations in which MLE tends to overfit.

Cite

Text

Zhu et al. "Bayesian Transformed Gaussian Processes." Transactions on Machine Learning Research, 2023.

Markdown

[Zhu et al. "Bayesian Transformed Gaussian Processes." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/zhu2023tmlr-bayesian/)

BibTeX

@article{zhu2023tmlr-bayesian,
  title     = {{Bayesian Transformed Gaussian Processes}},
  author    = {Zhu, Xinran and Huang, Leo and Lee, Eric Hans and Ibrahim, Cameron Alexander and Bindel, David},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/zhu2023tmlr-bayesian/}
}