Regression Conformal Prediction with Random Forests

Abstract

Regression conformal prediction produces prediction intervals that are valid, i.e., the probability of excluding the correct target value is bounded by a predefined confidence level. The most important criterion when comparing conformal regressors is efficiency; the prediction intervals should be as tight (informative) as possible. In this study, the use of random forests as the underlying model for regression conformal prediction is investigated and compared to existing state-of-the-art techniques, which are based on neural networks and k-nearest neighbors. In addition to their robust predictive performance, random forests allow for determining the size of the prediction intervals by using out-of-bag estimates instead of requiring a separate calibration set. An extensive empirical investigation, using 33 publicly available data sets, was undertaken to compare the use of random forests to existing state-of-the-art conformal predictors. The results show that the suggested approach, on almost all confidence levels and using both standard and normalized nonconformity functions, produced significantly more efficient conformal predictors than the existing alternatives.

Cite

Text

Johansson et al. "Regression Conformal Prediction with Random Forests." Machine Learning, 2014. doi:10.1007/S10994-014-5453-0

Markdown

[Johansson et al. "Regression Conformal Prediction with Random Forests." Machine Learning, 2014.](https://mlanthology.org/mlj/2014/johansson2014mlj-regression/) doi:10.1007/S10994-014-5453-0

BibTeX

@article{johansson2014mlj-regression,
  title     = {{Regression Conformal Prediction with Random Forests}},
  author    = {Johansson, Ulf and Boström, Henrik and Löfström, Tuve and Linusson, Henrik},
  journal   = {Machine Learning},
  year      = {2014},
  pages     = {155-176},
  doi       = {10.1007/S10994-014-5453-0},
  volume    = {97},
  url       = {https://mlanthology.org/mlj/2014/johansson2014mlj-regression/}
}