Measuring Statistical Dependence with Hilbert-Schmidt Norms

Abstract

We propose an independence criterion based on the eigenspectrum of covariance operators in reproducing kernel Hilbert spaces (RKHSs), consisting of an empirical estimate of the Hilbert-Schmidt norm of the cross-covariance operator (we term this a Hilbert-Schmidt Independence Criterion, or HSIC). This approach has several advantages, compared with previous kernel-based independence criteria. First, the empirical estimate is simpler than any other kernel dependence test, and requires no user-defined regularisation. Second, there is a clearly defined population quantity which the empirical estimate approaches in the large sample limit, with exponential convergence guaranteed between the two: this ensures that independence tests based on HSIC do not suffer from slow learning rates. Finally, we show in the context of independent component analysis (ICA) that the performance of HSIC is competitive with that of previously published kernel-based criteria, and of other recently published ICA methods.

Cite

Text

Gretton et al. "Measuring Statistical Dependence with Hilbert-Schmidt Norms." International Conference on Algorithmic Learning Theory, 2005. doi:10.1007/11564089_7

Markdown

[Gretton et al. "Measuring Statistical Dependence with Hilbert-Schmidt Norms." International Conference on Algorithmic Learning Theory, 2005.](https://mlanthology.org/alt/2005/gretton2005alt-measuring/) doi:10.1007/11564089_7

BibTeX

@inproceedings{gretton2005alt-measuring,
  title     = {{Measuring Statistical Dependence with Hilbert-Schmidt Norms}},
  author    = {Gretton, Arthur and Bousquet, Olivier and Smola, Alexander J. and Schölkopf, Bernhard},
  booktitle = {International Conference on Algorithmic Learning Theory},
  year      = {2005},
  pages     = {63-77},
  doi       = {10.1007/11564089_7},
  url       = {https://mlanthology.org/alt/2005/gretton2005alt-measuring/}
}