Estimation of Information Theoretic Measures for Continuous Random Variables

Abstract

We analyze the estimation of information theoretic measures of continuous random variables such as: differential entropy, mutual information or Kullback-Leibler divergence. The objective of this paper is two-fold. First, we prove that the information theoretic measure estimates using the k-nearest-neighbor density estimation with fixed k converge almost surely, even though the k-nearest-neighbor density estimation with fixed k does not converge to its true measure. Second, we show that the information theoretic measure estimates do not converge for k growing linearly with the number of samples. Nevertheless, these nonconvergent estimates can be used for solving the two-sample problem and assessing if two random variables are independent. We show that the two-sample and independence tests based on these nonconvergent estimates compare favorably with the maximum mean discrepancy test and the Hilbert Schmidt independence criterion, respectively.

Cite

Text

Pérez-Cruz. "Estimation of Information Theoretic Measures for Continuous Random Variables." Neural Information Processing Systems, 2008.

Markdown

[Pérez-Cruz. "Estimation of Information Theoretic Measures for Continuous Random Variables." Neural Information Processing Systems, 2008.](https://mlanthology.org/neurips/2008/perezcruz2008neurips-estimation/)

BibTeX

@inproceedings{perezcruz2008neurips-estimation,
  title     = {{Estimation of Information Theoretic Measures for Continuous Random Variables}},
  author    = {Pérez-Cruz, Fernando},
  booktitle = {Neural Information Processing Systems},
  year      = {2008},
  pages     = {1257-1264},
  url       = {https://mlanthology.org/neurips/2008/perezcruz2008neurips-estimation/}
}