On the Theoretical Guarantees for Parameter Estimation of Gaussian Random Field Models: A Sparse Precision Matrix Approach
Abstract
Iterative methods for fitting a Gaussian Random Field (GRF) model via maximum likelihood (ML) estimation requires solving a nonconvex optimization problem. The problem is aggravated for anisotropic GRFs where the number of covariance function parameters increases with the dimension. Even evaluation of the likelihood function requires $O(n^3)$ floating point operations, where $n$ denotes the number of data locations. In this paper, we propose a new two-stage procedure to estimate the parameters of second-order stationary GRFs. First, a convex likelihood problem regularized with a weighted $\ell_1$-norm, utilizing the available distance information between observation locations, is solved to fit a sparse precision (inverse covariance) matrix to the observed data. Second, the parameters of the covariance function are estimated by solving a least squares problem. Theoretical error bounds for the solutions of stage I and II problems are provided, and their tightness are investigated.
Cite
Text
Tajbakhsh et al. "On the Theoretical Guarantees for Parameter Estimation of Gaussian Random Field Models: A Sparse Precision Matrix Approach." Journal of Machine Learning Research, 2020.Markdown
[Tajbakhsh et al. "On the Theoretical Guarantees for Parameter Estimation of Gaussian Random Field Models: A Sparse Precision Matrix Approach." Journal of Machine Learning Research, 2020.](https://mlanthology.org/jmlr/2020/tajbakhsh2020jmlr-theoretical/)BibTeX
@article{tajbakhsh2020jmlr-theoretical,
title = {{On the Theoretical Guarantees for Parameter Estimation of Gaussian Random Field Models: A Sparse Precision Matrix Approach}},
author = {Tajbakhsh, Sam Davanloo and Aybat, Necdet Serhat and Del Castillo, Enrique},
journal = {Journal of Machine Learning Research},
year = {2020},
pages = {1-41},
volume = {21},
url = {https://mlanthology.org/jmlr/2020/tajbakhsh2020jmlr-theoretical/}
}