Finding Metric Structure in Information Theoretic Clustering
Abstract
We study the problem of clustering discrete probability distributions with respect to the Kullback-Leibler (KL) divergence. This problem arises naturally in many applications. Our goal is to pick k distributions as "representatives" such that the average or maximum KL- divergence between an input distribution and the closest representative distribution is minimized. Unfortunately, no polynomial-time algorithms with worst-case performance guarantees are known for either of these problems. The analogous problems for $\ell_1$, $\ell_2$ and $\ell_2$ 2 (i.e., k-center, k-median and k-means) have been extensively studied and efficient algorithms with good approximation guarantees are known. However, these algorithms rely crucially on the (geo-)metric properties of these metrics and do not apply to KL-divergence. In this paper, our contribution is to find a "relaxed" metricstructure for KL-divergence. In doing so, we provide the first polynomial-time algorithm for clustering using KL-divergences with provable guarantees for general inputs.
Cite
Text
Chaudhuri and McGregor. "Finding Metric Structure in Information Theoretic Clustering." Annual Conference on Computational Learning Theory, 2008.Markdown
[Chaudhuri and McGregor. "Finding Metric Structure in Information Theoretic Clustering." Annual Conference on Computational Learning Theory, 2008.](https://mlanthology.org/colt/2008/chaudhuri2008colt-finding/)BibTeX
@inproceedings{chaudhuri2008colt-finding,
title = {{Finding Metric Structure in Information Theoretic Clustering}},
author = {Chaudhuri, Kamalika and McGregor, Andrew},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2008},
pages = {391-402},
url = {https://mlanthology.org/colt/2008/chaudhuri2008colt-finding/}
}