A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence

Abstract

We introduce a novel kernel-based information-theoretic framework for two-sample testing, leveraging the representation Jensen-Shannon divergence (RJSD). RJSD captures higher-order information from covariance operators in reproducing Kernel Hilbert spaces and avoids Gaussianity assumptions, providing a robust and flexible measure of divergence between distributions. We develop RJSD-based variants of Maximum Mean Discrepancy (MMD) approaches, demonstrating superior discriminative power in extensive experiments on synthetic and real-world datasets. Our results position RJSD as a powerful alternative to MMD, with the potential to significantly impact kernel-based learning and distribution comparison. By establishing RJSD as a benchmark for two-sample testing, this work lays the foundation for future research in kernel-based divergence estimation and its broad range of applications in machine learning.

Cite

Text

Hoyos and Giraldo. "A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence." NeurIPS 2024 Workshops: LXAI, 2024.

Markdown

[Hoyos and Giraldo. "A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence." NeurIPS 2024 Workshops: LXAI, 2024.](https://mlanthology.org/neuripsw/2024/hoyos2024neuripsw-kernel/)

BibTeX

@inproceedings{hoyos2024neuripsw-kernel,
  title     = {{A Kernel Two-Sample Test with the Representation Jensen-Shannon Divergence}},
  author    = {Hoyos, Jhoan Keider and Giraldo, Luis Gonzalo Sanchez},
  booktitle = {NeurIPS 2024 Workshops: LXAI},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/hoyos2024neuripsw-kernel/}
}