Robust Independence Testing for Constraint-Based Learning of Causal Structure

Abstract

Constraint-based (CB) learning is a formalism for learning a causal network with a database D by performing a series of conditional-independence tests to infer structural information. This paper considers a new test of independence that combines ideas from Bayesian learning, Bayesian network inference, and classical hypothesis testing to produce a more reliable and robust test. The new test can be calculated in the same asymptotic time and space required for the standard tests such as the chi-squared test, but it allows the specification of a prior distribution over parameters and can be used when the database is incomplete. We prove that the test is correct, and we demonstrate empirically that, when used with a CB causal discovery algorithm with noninformative priors, it recovers structural features more reliably and it produces networks with smaller KL-Divergence, especially as the number of nodes increases or the number of records decreases. Another benefit is the dramatic reduction in the probability that a CB algorithm will stall during the search, providing a remedy for an annoying problem plaguing CB learning when the database is small.

Cite

Text

Dash and Druzdzel. "Robust Independence Testing for Constraint-Based Learning of Causal Structure." Conference on Uncertainty in Artificial Intelligence, 2003.

Markdown

[Dash and Druzdzel. "Robust Independence Testing for Constraint-Based Learning of Causal Structure." Conference on Uncertainty in Artificial Intelligence, 2003.](https://mlanthology.org/uai/2003/dash2003uai-robust/)

BibTeX

@inproceedings{dash2003uai-robust,
  title     = {{Robust Independence Testing for Constraint-Based Learning of Causal Structure}},
  author    = {Dash, Denver and Druzdzel, Marek J.},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2003},
  pages     = {167-174},
  url       = {https://mlanthology.org/uai/2003/dash2003uai-robust/}
}