A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence
Abstract
We introduce a novel kernel-based information-theoretic framework for two-sample testing, leveraging the representation Jensen-Shannon divergence (RJSD). RJSD captures higher-order information from covariance operators in reproducing Kernel Hilbert spaces and avoids Gaussianity assumptions, providing a robust and flexible measure of divergence between distributions. We develop RJSD-based variants of Maximum Mean Discrepancy (MMD) approaches, demonstrating superior discriminative power in extensive experiments on synthetic and real-world datasets. Our results position RJSD as a powerful alternative to MMD, with the potential to significantly impact kernel-based learning and distribution comparison. By establishing RJSD as a benchmark for two-sample testing, this work lays the foundation for future research in kernel-based divergence estimation and its broad range of applications in machine learning.
Cite
Text
Hoyos and Giraldo. "A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence." NeurIPS 2024 Workshops: LXAI, 2024.Markdown
[Hoyos and Giraldo. "A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence." NeurIPS 2024 Workshops: LXAI, 2024.](https://mlanthology.org/neuripsw/2024/hoyos2024neuripsw-kernel/)BibTeX
@inproceedings{hoyos2024neuripsw-kernel,
title = {{A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence}},
author = {Hoyos, Jhoan Keider and Giraldo, Luis Gonzalo Sanchez},
booktitle = {NeurIPS 2024 Workshops: LXAI},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/hoyos2024neuripsw-kernel/}
}