Estimation of Information Theoretic Measures for Continuous Random Variables
Abstract
We analyze the estimation of information theoretic measures of continuous random variables such as: differential entropy, mutual information or Kullback-Leibler divergence. The objective of this paper is two-fold. First, we prove that the information theoretic measure estimates using the k-nearest-neighbor density estimation with fixed k converge almost surely, even though the k-nearest-neighbor density estimation with fixed k does not converge to its true measure. Second, we show that the information theoretic measure estimates do not converge for k growing linearly with the number of samples. Nevertheless, these nonconvergent estimates can be used for solving the two-sample problem and assessing if two random variables are independent. We show that the two-sample and independence tests based on these nonconvergent estimates compare favorably with the maximum mean discrepancy test and the Hilbert Schmidt independence criterion, respectively.
Cite
Text
Pérez-Cruz. "Estimation of Information Theoretic Measures for Continuous Random Variables." Neural Information Processing Systems, 2008.Markdown
[Pérez-Cruz. "Estimation of Information Theoretic Measures for Continuous Random Variables." Neural Information Processing Systems, 2008.](https://mlanthology.org/neurips/2008/perezcruz2008neurips-estimation/)BibTeX
@inproceedings{perezcruz2008neurips-estimation,
title = {{Estimation of Information Theoretic Measures for Continuous Random Variables}},
author = {Pérez-Cruz, Fernando},
booktitle = {Neural Information Processing Systems},
year = {2008},
pages = {1257-1264},
url = {https://mlanthology.org/neurips/2008/perezcruz2008neurips-estimation/}
}