PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption

Abstract

We propose and analyze a new vantage point for the learning of mixtures of Gaussians: namely, the PAC-style model of learning probability distributions introduced by Kearns et al. [13]. Here the task is to construct a hypothesis mixture of Gaussians that is statistically indistinguishable from the actual mixture generating the data; specifically, the KL divergence should be at most In this scenario, we give a poly(n/e) time algorithm that learns the class of mixtures of any constant number of axis-aligned Gaussians in R. Our algorithm makes no assumptions about the separation between the means of the Gaussians, nor does it have any dependence on the minimum mixing weight. This is in contrast to learning results known in the clustering model, where such assumptions are unavoidable. Our algorithm relies on the method of moments, and a subalgorithm developed in [9] for a discrete mixture-learning problem.

Cite

Text

Feldman et al. "PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption." Annual Conference on Computational Learning Theory, 2006. doi:10.1007/11776420_5

Markdown

[Feldman et al. "PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption." Annual Conference on Computational Learning Theory, 2006.](https://mlanthology.org/colt/2006/feldman2006colt-pac/) doi:10.1007/11776420_5

BibTeX

@inproceedings{feldman2006colt-pac,
  title     = {{PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption}},
  author    = {Feldman, Jon and Servedio, Rocco A. and O'Donnell, Ryan},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2006},
  pages     = {20-34},
  doi       = {10.1007/11776420_5},
  url       = {https://mlanthology.org/colt/2006/feldman2006colt-pac/}
}