Fast Learning from Sparse Data
Abstract
We describe two techniques that significantly improve the running time of several standard machine-learning algorithms when data is sparse. The first technique is an algorithm that efficiently extracts one-way and two-way counts-either real or expected-from discrete data. Extracting such counts is a fundamental step in learning algorithms for constructing a variety of models including decision trees, decision graphs, Bayesian networks, and naive-Bayes clustering models. The second technique is an algorithm that efficiently performs the E-step of the EM algorithm (i.e., inference) when applied to a naive-Bayes clustering model. Using real-world data sets, we demonstrate a dramatic decrease in running time for algorithms that incorporate these techniques.
Cite
Text
Chickering and Heckerman. "Fast Learning from Sparse Data." Conference on Uncertainty in Artificial Intelligence, 1999.Markdown
[Chickering and Heckerman. "Fast Learning from Sparse Data." Conference on Uncertainty in Artificial Intelligence, 1999.](https://mlanthology.org/uai/1999/chickering1999uai-fast/)BibTeX
@inproceedings{chickering1999uai-fast,
title = {{Fast Learning from Sparse Data}},
author = {Chickering, David Maxwell and Heckerman, David},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {1999},
pages = {109-115},
url = {https://mlanthology.org/uai/1999/chickering1999uai-fast/}
}