Evaluating Bayesian Models with Posterior Dispersion Indices

Abstract

Probabilistic modeling is cyclical: we specify a model, infer its posterior, and evaluate its performance. Evaluation drives the cycle, as we revise our model based on how it performs. This requires a metric. Traditionally, predictive accuracy prevails. Yet, predictive accuracy does not tell the whole story. We propose to evaluate a model through posterior dispersion. The idea is to analyze how each datapoint fares in relation to posterior uncertainty around the hidden structure. This highlights datapoints the model struggles to explain and provides complimentary insight to datapoints with low predictive accuracy. We present a family of posterior dispersion indices (PDI) that capture this idea. We show how a PDI identifies patterns of model mismatch in three real data examples: voting preferences, supermarket shopping, and population genetics.

Cite

Text

Kucukelbir et al. "Evaluating Bayesian Models with Posterior Dispersion Indices." International Conference on Machine Learning, 2017.

Markdown

[Kucukelbir et al. "Evaluating Bayesian Models with Posterior Dispersion Indices." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/kucukelbir2017icml-evaluating/)

BibTeX

@inproceedings{kucukelbir2017icml-evaluating,
  title     = {{Evaluating Bayesian Models with Posterior Dispersion Indices}},
  author    = {Kucukelbir, Alp and Wang, Yixin and Blei, David M.},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {1925-1934},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/kucukelbir2017icml-evaluating/}
}