Boosting Using Branching Programs
Abstract
It is known that decision tree learning can be viewed as a form of boosting. Given a weak learning hypothesis one can show that the training error of a decision tree declines as |T|−β where |T| is the size of the decision tree and β is a constant determined by the weak learning hypothesis. Here we consider the case of decision DAGs—decision trees in which a given node can be shared by different branches of the tree, also called branching programs (BP). Node sharing allows a branching program to be exponentially more compact than the corresponding decision tree. We show that under the same weak learning assumption used for decision tree learning there exists a greedy BP-growth algorithm whose training error is guaranteed to decline as 2−β |T| , where |T| is the size of the branching program and β is a constant determined by the weak learning hypothesis. Therefore, from the perspective of boosting theory, branching programs are exponentially more efficient than decision trees.
Cite
Text
Mansour and McAllester. "Boosting Using Branching Programs." Annual Conference on Computational Learning Theory, 2000. doi:10.1006/jcss.2001.1796Markdown
[Mansour and McAllester. "Boosting Using Branching Programs." Annual Conference on Computational Learning Theory, 2000.](https://mlanthology.org/colt/2000/mansour2000colt-boosting/) doi:10.1006/jcss.2001.1796BibTeX
@inproceedings{mansour2000colt-boosting,
title = {{Boosting Using Branching Programs}},
author = {Mansour, Yishay and McAllester, David A.},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2000},
pages = {220-224},
doi = {10.1006/jcss.2001.1796},
url = {https://mlanthology.org/colt/2000/mansour2000colt-boosting/}
}