A Scoring Function for Learning Bayesian Networks Based on Mutual Information and Conditional Independence Tests
Abstract
We propose a new scoring function for learning Bayesian networks from data using score+search algorithms. This is based on the concept of mutual information and exploits some well-known properties of this measure in a novel way. Essentially, a statistical independence test based on the chi-square distribution, associated with the mutual information measure, together with a property of additive decomposition of this measure, are combined in order to measure the degree of interaction between each variable and its parent variables in the network. The result is a non-Bayesian scoring function called MIT (mutual information tests) which belongs to the family of scores based on information theory. The MIT score also represents a penalization of the Kullback-Leibler divergence between the joint probability distributions associated with a candidate network and with the available data set. Detailed results of a complete experimental evaluation of the proposed scoring function and its comparison with the well-known K2, BDeu and BIC/MDL scores are also presented.
Cite
Text
de Campos. "A Scoring Function for Learning Bayesian Networks Based on Mutual Information and Conditional Independence Tests." Journal of Machine Learning Research, 2006.Markdown
[de Campos. "A Scoring Function for Learning Bayesian Networks Based on Mutual Information and Conditional Independence Tests." Journal of Machine Learning Research, 2006.](https://mlanthology.org/jmlr/2006/decampos2006jmlr-scoring/)BibTeX
@article{decampos2006jmlr-scoring,
title = {{A Scoring Function for Learning Bayesian Networks Based on Mutual Information and Conditional Independence Tests}},
author = {de Campos, Luis M.},
journal = {Journal of Machine Learning Research},
year = {2006},
pages = {2149-2187},
volume = {7},
url = {https://mlanthology.org/jmlr/2006/decampos2006jmlr-scoring/}
}