Finding Metric Structure in Information Theoretic Clustering

Abstract

We study the problem of clustering discrete probability distributions with respect to the Kullback-Leibler (KL) divergence. This problem arises naturally in many applications. Our goal is to pick k distributions as "representatives" such that the average or maximum KL- divergence between an input distribution and the closest representative distribution is minimized. Unfortunately, no polynomial-time algorithms with worst-case performance guarantees are known for either of these problems. The analogous problems for $\ell_1$, $\ell_2$ and $\ell_2$ 2 (i.e., k-center, k-median and k-means) have been extensively studied and efficient algorithms with good approximation guarantees are known. However, these algorithms rely crucially on the (geo-)metric properties of these metrics and do not apply to KL-divergence. In this paper, our contribution is to find a "relaxed" metricstructure for KL-divergence. In doing so, we provide the first polynomial-time algorithm for clustering using KL-divergences with provable guarantees for general inputs.

Cite

Text

Chaudhuri and McGregor. "Finding Metric Structure in Information Theoretic Clustering." Annual Conference on Computational Learning Theory, 2008.

Markdown

[Chaudhuri and McGregor. "Finding Metric Structure in Information Theoretic Clustering." Annual Conference on Computational Learning Theory, 2008.](https://mlanthology.org/colt/2008/chaudhuri2008colt-finding/)

BibTeX

@inproceedings{chaudhuri2008colt-finding,
  title     = {{Finding Metric Structure in Information Theoretic Clustering}},
  author    = {Chaudhuri, Kamalika and McGregor, Andrew},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2008},
  pages     = {391-402},
  url       = {https://mlanthology.org/colt/2008/chaudhuri2008colt-finding/}
}