Practical Confidence and Prediction Intervals

Abstract

We propose a new method to compute prediction intervals. Espe(cid:173) cially for small data sets the width of a prediction interval does not only depend on the variance of the target distribution, but also on the accuracy of our estimator of the mean of the target, i.e., on the width of the confidence interval. The confidence interval follows from the variation in an ensemble of neural networks, each of them trained and stopped on bootstrap replicates of the original data set. A second improvement is the use of the residuals on validation pat(cid:173) terns instead of on training patterns for estimation of the variance of the target distribution. As illustrated on a synthetic example, our method is better than existing methods with regard to extrap(cid:173) olation and interpolation in data regimes with a limited amount of data, and yields prediction intervals which actual confidence levels are closer to the desired confidence levels.

Cite

Text

Heskes. "Practical Confidence and Prediction Intervals." Neural Information Processing Systems, 1996.

Markdown

[Heskes. "Practical Confidence and Prediction Intervals." Neural Information Processing Systems, 1996.](https://mlanthology.org/neurips/1996/heskes1996neurips-practical/)

BibTeX

@inproceedings{heskes1996neurips-practical,
  title     = {{Practical Confidence and Prediction Intervals}},
  author    = {Heskes, Tom},
  booktitle = {Neural Information Processing Systems},
  year      = {1996},
  pages     = {176-182},
  url       = {https://mlanthology.org/neurips/1996/heskes1996neurips-practical/}
}