Estimating Continuous Distributions in Bayesian Classifiers
Abstract
When modeling a probability distribution with a Bayesian network, we are faced with the problem of how to handle continuous variables. Most previous work has either solved the problem by discretizing, or assumed that the data are generated by a single Gaussian. In this paper we abandon the normality assumption and instead use statistical methods for nonparametric density estimation. For a naive Bayesian classifier, we present experimental results on a variety of natural and artificial domains, comparing two methods of density estimation: assuming normality and modeling each conditional distribution with a single Gaussian; and using nonparametric kernel density estimation. We observe large reductions in error on several natural and artificial data sets, which suggests that kernel estimation is a useful tool for learning Bayesian models.
Cite
Text
John and Langley. "Estimating Continuous Distributions in Bayesian Classifiers." Conference on Uncertainty in Artificial Intelligence, 1995.Markdown
[John and Langley. "Estimating Continuous Distributions in Bayesian Classifiers." Conference on Uncertainty in Artificial Intelligence, 1995.](https://mlanthology.org/uai/1995/john1995uai-estimating/)BibTeX
@inproceedings{john1995uai-estimating,
title = {{Estimating Continuous Distributions in Bayesian Classifiers}},
author = {John, George H. and Langley, Pat},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {1995},
pages = {338-345},
url = {https://mlanthology.org/uai/1995/john1995uai-estimating/}
}