Learning Bayesian Network Classifiers by Maximizing Conditional Likelihood
Abstract
Bayesian networks are a powerful probabilistic representation, and their usefor classification has received considerable attention. However, they tend toperform poorly when learned in the standard way. This is attributable to amismatch between the objective function used (likelihood or a functionthereof) and the goal of classification (maximizing accuracy or conditionallikelihood). Unfortunately, the computational cost of optimizing structureand parameters for conditional likelihood is prohibitive. In this paper weshow that a simple approximation - choosing structures by maximizingconditional likelihood while setting parameters by maximum likelihood - yieldsgood results. On a large suite of benchmark datasets, this approach producesbetter class probability estimates than naive Bayes, TAN, andgeneratively-trained Bayesian networks.
Cite
Text
Grossman and Domingos. "Learning Bayesian Network Classifiers by Maximizing Conditional Likelihood." International Conference on Machine Learning, 2004. doi:10.1145/1015330.1015339Markdown
[Grossman and Domingos. "Learning Bayesian Network Classifiers by Maximizing Conditional Likelihood." International Conference on Machine Learning, 2004.](https://mlanthology.org/icml/2004/grossman2004icml-learning/) doi:10.1145/1015330.1015339BibTeX
@inproceedings{grossman2004icml-learning,
title = {{Learning Bayesian Network Classifiers by Maximizing Conditional Likelihood}},
author = {Grossman, Daniel and Domingos, Pedro M.},
booktitle = {International Conference on Machine Learning},
year = {2004},
doi = {10.1145/1015330.1015339},
url = {https://mlanthology.org/icml/2004/grossman2004icml-learning/}
}