Stochastic Gradient MCMC for Gaussian Process Inference on Massive Geostatistical Data

Abstract

Gaussian processes (GPs) are the workhorses of spatial data analyses, but are difficult to scale to large spatial datasets. The Vecchia approximation induces sparsity in the dependence structure and is one of several methods proposed to scale GP inference. We develop a stochastic gradient Markov chain Monte Carlo framework for efficient computation in GPs for spatial data. At each step, the algorithm subsamples a minibatch of locations and subsequently updates process parameters through stochastic gradient Riemannian Langevin dynamics (SGRLD) on a Vecchia-approximated GP likelihood. We are able to conduct full Bayesian analysis for GPs with up to 100,000 locations using our spatial SGRLD, and demonstrate its efficacy through numerical studies and an application using ocean temperature data.

Cite

Text

Abba et al. "Stochastic Gradient MCMC for Gaussian Process Inference on Massive Geostatistical Data." NeurIPS 2024 Workshops: BDU, 2024.

Markdown

[Abba et al. "Stochastic Gradient MCMC for Gaussian Process Inference on Massive Geostatistical Data." NeurIPS 2024 Workshops: BDU, 2024.](https://mlanthology.org/neuripsw/2024/abba2024neuripsw-stochastic/)

BibTeX

@inproceedings{abba2024neuripsw-stochastic,
  title     = {{Stochastic Gradient MCMC for Gaussian Process Inference on Massive Geostatistical Data}},
  author    = {Abba, Mohamed A. and Reich, Brian J. and Majumder, Reetam and Feng, Brandon},
  booktitle = {NeurIPS 2024 Workshops: BDU},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/abba2024neuripsw-stochastic/}
}