Bayesian and Empirical Bayesian Forests

Abstract

We derive ensembles of decision trees through a nonparametric Bayesian model, allowing us to view such ensembles as samples from a posterior distribution. This insight motivates a class of Bayesian Forest (BF) algorithms that provide small gains in performance and large gains in interpretability. Based on the BF framework, we are able to show that high-level tree hierarchy is stable in large samples. This motivates an empirical Bayesian Forest (EBF) algorithm for building approximate BFs on massive distributed datasets and we show that EBFs outperform sub-sampling based alternatives by a large margin.

Cite

Text

Matthew et al. "Bayesian and Empirical Bayesian Forests." International Conference on Machine Learning, 2015.

Markdown

[Matthew et al. "Bayesian and Empirical Bayesian Forests." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/matthew2015icml-bayesian/)

BibTeX

@inproceedings{matthew2015icml-bayesian,
  title     = {{Bayesian and Empirical Bayesian Forests}},
  author    = {Matthew, Taddy and Chen, Chun-Sheng and Yu, Jun and Wyle, Mitch},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {967-976},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/matthew2015icml-bayesian/}
}