Large-Sample Learning of Bayesian Networks Is NP-Hard
Abstract
In this paper, we provide new complexity results for algorithms that learn discretevariable Bayesian networks from data. Our results apply whenever the learning algorithm uses a scoring criterion that favors the simplest model able to represent the generative distribution exactly. Our results therefore hold whenever the learning algorithm uses a consistent scoring criterion and is applied to a sufficiently large dataset. We show that identifying high-scoring structures is NP-hard, even when we are given an independence oracle, an inference oracle, and/or an information oracle. Our negative results also apply when learning discrete-variable Bayesian networks in which each node has at most k parents, for all k ≥ 3.
Cite
Text
Chickering et al. "Large-Sample Learning of Bayesian Networks Is NP-Hard." Conference on Uncertainty in Artificial Intelligence, 2003.Markdown
[Chickering et al. "Large-Sample Learning of Bayesian Networks Is NP-Hard." Conference on Uncertainty in Artificial Intelligence, 2003.](https://mlanthology.org/uai/2003/chickering2003uai-large/)BibTeX
@inproceedings{chickering2003uai-large,
title = {{Large-Sample Learning of Bayesian Networks Is NP-Hard}},
author = {Chickering, David Maxwell and Meek, Christopher and Heckerman, David},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {2003},
pages = {124-133},
url = {https://mlanthology.org/uai/2003/chickering2003uai-large/}
}