Decision Trees: More Theoretical Justification for Practical Algorithms
Abstract
We study impurity-based decision tree algorithms such as CART, C4.5, etc. , so as to better understand their theoretical underpinnings. We consider such algorithms on special forms of functions and distributions. We deal with the uniform distribution and functions that can be described as a boolean linear threshold functions or a read-once DNF. We show that for boolean linear threshold functions and read-once DNF, maximal purity gain and maximal influence are logically equivalent. This leads us to the exact identification of these classes of functions by impurity-based algorithms given sufficiently many noise-free examples. We show that the decision tree resulting from these algorithms has minimal size and height amongst all decision trees representing the function. Based on the statistical query learning model, we introduce the noise-tolerant version of practical decision tree algorithms. We show that if the input examples have small classification noise and are uniformly distributed, then all our results for practical noise-free impurity-based algorithms also hold for their noise-tolerant version.
Cite
Text
Fiat and Pechyony. "Decision Trees: More Theoretical Justification for Practical Algorithms." International Conference on Algorithmic Learning Theory, 2004. doi:10.1007/978-3-540-30215-5_13Markdown
[Fiat and Pechyony. "Decision Trees: More Theoretical Justification for Practical Algorithms." International Conference on Algorithmic Learning Theory, 2004.](https://mlanthology.org/alt/2004/fiat2004alt-decision/) doi:10.1007/978-3-540-30215-5_13BibTeX
@inproceedings{fiat2004alt-decision,
title = {{Decision Trees: More Theoretical Justification for Practical Algorithms}},
author = {Fiat, Amos and Pechyony, Dmitry},
booktitle = {International Conference on Algorithmic Learning Theory},
year = {2004},
pages = {156-170},
doi = {10.1007/978-3-540-30215-5_13},
url = {https://mlanthology.org/alt/2004/fiat2004alt-decision/}
}