Convex Optimization Techniques for Fitting Sparse Gaussian Graphical Models

Abstract

We consider the problem of fitting a large-scale covariance matrix to multivariate Gaussian data in such a way that the inverse is sparse, thus providing model selection. Beginning with a dense empirical covariance matrix, we solve a maximum likelihood problem with an l1-norm penalty term added to encourage sparsity in the inverse. For models with tens of nodes, the resulting problem can be solved using standard interior-point algorithms for convex optimization, but these methods scale poorly with problem size. We present two new algorithms aimed at solving problems with a thousand nodes. The first, based on Nesterov's first-order algorithm, yields a rigorous complexity estimate for the problem, with a much better dependence on problem size than interior-point methods. Our second algorithm uses block coordinate descent, updating row/columns of the covariance matrix sequentially. Experiments with genomic data show that our method is able to uncover biologically interpretable connections among genes.

Cite

Text

Banerjee et al. "Convex Optimization Techniques for Fitting Sparse Gaussian Graphical Models." International Conference on Machine Learning, 2006. doi:10.1145/1143844.1143856

Markdown

[Banerjee et al. "Convex Optimization Techniques for Fitting Sparse Gaussian Graphical Models." International Conference on Machine Learning, 2006.](https://mlanthology.org/icml/2006/banerjee2006icml-convex/) doi:10.1145/1143844.1143856

BibTeX

@inproceedings{banerjee2006icml-convex,
  title     = {{Convex Optimization Techniques for Fitting Sparse Gaussian Graphical Models}},
  author    = {Banerjee, Onureena and El Ghaoui, Laurent and d'Aspremont, Alexandre and Natsoulis, Georges},
  booktitle = {International Conference on Machine Learning},
  year      = {2006},
  pages     = {89-96},
  doi       = {10.1145/1143844.1143856},
  url       = {https://mlanthology.org/icml/2006/banerjee2006icml-convex/}
}