Improved Mutual Information Estimation
Abstract
We propose to estimate the KL divergence using a relaxed likelihood ratio estimation in a Reproducing Kernel Hilbert space. We show that the dual of our ratio estimator for KL in the particular case of Mutual Information estimation corresponds to a lower bound on the MI that is related to the so called Donsker Varadhan lower bound. In this dual form, MI is estimated via learning a witness function discriminating between the joint density and the product of marginal, as well as an auxiliary scalar variable that enforces a normalization constraint on the likelihood ratio. By extending the function space to neural networks, we propose an efficient neural MI estimator, and validate its performance on synthetic examples, showing advantage over the existing baselines. We demonstrate its strength in large-scale self-supervised representation learning through MI maximization.
Cite
Text
Mroueh et al. "Improved Mutual Information Estimation." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I10.17089Markdown
[Mroueh et al. "Improved Mutual Information Estimation." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/mroueh2021aaai-improved/) doi:10.1609/AAAI.V35I10.17089BibTeX
@inproceedings{mroueh2021aaai-improved,
title = {{Improved Mutual Information Estimation}},
author = {Mroueh, Youssef and Melnyk, Igor and Dognin, Pierre L. and Ross, Jarret and Sercu, Tom},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2021},
pages = {9009-9017},
doi = {10.1609/AAAI.V35I10.17089},
url = {https://mlanthology.org/aaai/2021/mroueh2021aaai-improved/}
}