Estimating Mutual Information for Discrete-Continuous Mixtures
Abstract
Estimation of mutual information from observed samples is a basic primitive in machine learning, useful in several learning tasks including correlation mining, information bottleneck, Chow-Liu tree, and conditional independence testing in (causal) graphical models. While mutual information is a quantity well-defined for general probability spaces, estimators have been developed only in the special case of discrete or continuous pairs of random variables. Most of these estimators operate using the 3H -principle, i.e., by calculating the three (differential) entropies of X, Y and the pair (X,Y). However, in general mixture spaces, such individual entropies are not well defined, even though mutual information is. In this paper, we develop a novel estimator for estimating mutual information in discrete-continuous mixtures. We prove the consistency of this estimator theoretically as well as demonstrate its excellent empirical performance. This problem is relevant in a wide-array of applications, where some variables are discrete, some continuous, and others are a mixture between continuous and discrete components.
Cite
Text
Gao et al. "Estimating Mutual Information for Discrete-Continuous Mixtures." Neural Information Processing Systems, 2017.Markdown
[Gao et al. "Estimating Mutual Information for Discrete-Continuous Mixtures." Neural Information Processing Systems, 2017.](https://mlanthology.org/neurips/2017/gao2017neurips-estimating/)BibTeX
@inproceedings{gao2017neurips-estimating,
title = {{Estimating Mutual Information for Discrete-Continuous Mixtures}},
author = {Gao, Weihao and Kannan, Sreeram and Oh, Sewoong and Viswanath, Pramod},
booktitle = {Neural Information Processing Systems},
year = {2017},
pages = {5986-5997},
url = {https://mlanthology.org/neurips/2017/gao2017neurips-estimating/}
}