Boosting Using Branching Programs

Abstract

It is known that decision tree learning can be viewed as a form of boosting. Given a weak learning hypothesis one can show that the training error of a decision tree declines as |T|−β where |T| is the size of the decision tree and β is a constant determined by the weak learning hypothesis. Here we consider the case of decision DAGs—decision trees in which a given node can be shared by different branches of the tree, also called branching programs (BP). Node sharing allows a branching program to be exponentially more compact than the corresponding decision tree. We show that under the same weak learning assumption used for decision tree learning there exists a greedy BP-growth algorithm whose training error is guaranteed to decline as 2−β |T| , where |T| is the size of the branching program and β is a constant determined by the weak learning hypothesis. Therefore, from the perspective of boosting theory, branching programs are exponentially more efficient than decision trees.

Cite

Text

Mansour and McAllester. "Boosting Using Branching Programs." Annual Conference on Computational Learning Theory, 2000. doi:10.1006/jcss.2001.1796

Markdown

[Mansour and McAllester. "Boosting Using Branching Programs." Annual Conference on Computational Learning Theory, 2000.](https://mlanthology.org/colt/2000/mansour2000colt-boosting/) doi:10.1006/jcss.2001.1796

BibTeX

@inproceedings{mansour2000colt-boosting,
  title     = {{Boosting Using Branching Programs}},
  author    = {Mansour, Yishay and McAllester, David A.},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2000},
  pages     = {220-224},
  doi       = {10.1006/jcss.2001.1796},
  url       = {https://mlanthology.org/colt/2000/mansour2000colt-boosting/}
}