Self-Discrepancy Conditional Independence Test

Abstract

Tests of conditional independence (CI) of random variables play an important role in machine learning and causal inference. Of particular interest are kernel-based CI tests which allow us to test for independence among random variables with complex distribution functions. The efficacy of a CI test is measured in terms of its power and its calibratedness. We show that the Kernel CI Permutation Test (KCIPT) suffers from a loss of calibratedness as its power is increased by increasing the number of bootstraps. To address this limitation, we propose a novel CI test, called Self-Discrepancy Conditional Independence Test (SDCIT). SDCIT uses a test statistic that is a modified unbiased estimate of maximum mean discrepancy (MMD), the largest difference in the means of features of the given sample and its permuted counterpart in the kernel-induced Hilbert space. We present results of experiments that demonstrate SDCIT is, relative to the other methods: (i) competitive in terms of its power and calibratedness, outperforming other methods when the number of conditioning variables is large; (ii) more robust with respect to the choice of the kernel function; and (iii) competitive in run time.

Cite

Text

Lee and Honavar. "Self-Discrepancy Conditional Independence Test." Conference on Uncertainty in Artificial Intelligence, 2017.

Markdown

[Lee and Honavar. "Self-Discrepancy Conditional Independence Test." Conference on Uncertainty in Artificial Intelligence, 2017.](https://mlanthology.org/uai/2017/lee2017uai-self/)

BibTeX

@inproceedings{lee2017uai-self,
  title     = {{Self-Discrepancy Conditional Independence Test}},
  author    = {Lee, Sanghack and Honavar, Vasant G.},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2017},
  url       = {https://mlanthology.org/uai/2017/lee2017uai-self/}
}