Group Sparse Priors for Covariance Estimation

Abstract

Recently it has become popular to learn sparse Gaussian graphical models (GGMs) by imposing l1 or group l1, 2 penalties on the elements of the precision matrix. This penalized likelihood approach results in a tractable convex optimization problem. In this paper, we reinterpret these results as performing MAP estimation under a novel prior which we call the group l1 and l1, 2 positive-definite matrix distributions. This enables us to build a hierarchical model in which the l1 regularization terms vary depending on which group the entries are assigned to, which in turn allows us to learn block structured sparse GGMs with unknown group assignments. Exact inference in this hierarchical model is intractable, due to the need to compute the normalization constant of these matrix distributions. However, we derive upper bounds on the partition functions, which lets us use fast variational inference (optimizing a lower bound on the joint posterior). We show that on two real world data sets (motion capture and financial data), our method which infers the block structure outperforms a method that uses a fixed block structure, which in turn outperforms baseline methods that ignore block structure.

Cite

Text

Marlin et al. "Group Sparse Priors for Covariance Estimation." Conference on Uncertainty in Artificial Intelligence, 2009.

Markdown

[Marlin et al. "Group Sparse Priors for Covariance Estimation." Conference on Uncertainty in Artificial Intelligence, 2009.](https://mlanthology.org/uai/2009/marlin2009uai-group/)

BibTeX

@inproceedings{marlin2009uai-group,
  title     = {{Group Sparse Priors for Covariance Estimation}},
  author    = {Marlin, Benjamin M. and Schmidt, Mark and Murphy, Kevin P.},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2009},
  pages     = {383-392},
  url       = {https://mlanthology.org/uai/2009/marlin2009uai-group/}
}