In Defense of C4.5: Notes in Learning One-Level Decision Trees
Abstract
We discuss the implications of Holte's recently-published article, which demonstrated that on the most commonly used data very simple classification rules are almost as accurate as decision trees produced by Quinlan's C4.5. We consider, in particular, what is the significance of Holte's results for the future of top-down induction of decision trees. To an extent, Holte questioned the sense of further research on multilevel decision tree learning. We go in detail through all the parts of Holte's study. We try to put the results into perspective. We argue that the (in absolute terms) small difference in accuracy between 1R and C4.5 that was witnessed by Holte is still significant. We claim that C4.5 possesses additional accuracy-related advantages over 1R. In addition we discuss the representativeness of the databases used by Holte. We compare empirically the optimal accuracies of multilevel and one-level decision trees and observe some significant differences. We point out several deficiencies of limited-complexity classifiers.
Cite
Text
Elomaa. "In Defense of C4.5: Notes in Learning One-Level Decision Trees." International Conference on Machine Learning, 1994. doi:10.1016/B978-1-55860-335-6.50016-7Markdown
[Elomaa. "In Defense of C4.5: Notes in Learning One-Level Decision Trees." International Conference on Machine Learning, 1994.](https://mlanthology.org/icml/1994/elomaa1994icml-defense/) doi:10.1016/B978-1-55860-335-6.50016-7BibTeX
@inproceedings{elomaa1994icml-defense,
title = {{In Defense of C4.5: Notes in Learning One-Level Decision Trees}},
author = {Elomaa, Tapio},
booktitle = {International Conference on Machine Learning},
year = {1994},
pages = {62-69},
doi = {10.1016/B978-1-55860-335-6.50016-7},
url = {https://mlanthology.org/icml/1994/elomaa1994icml-defense/}
}