Robust Probabilistic Modeling with Bayesian Data Reweighting

Abstract

Probabilistic models analyze data by relying on a set of assumptions. Data that exhibit deviations from these assumptions can undermine inference and prediction quality. Robust models offer protection against mismatch between a model’s assumptions and reality. We propose a way to systematically detect and mitigate mismatch of a large class of probabilistic models. The idea is to raise the likelihood of each observation to a weight and then to infer both the latent variables and the weights from data. Inferring the weights allows a model to identify observations that match its assumptions and down-weight others. This enables robust inference and improves predictive accuracy. We study four different forms of mismatch with reality, ranging from missing latent groups to structure misspecification. A Poisson factorization analysis of the Movielens 1M dataset shows the benefits of this approach in a practical scenario.

Cite

Text

Wang et al. "Robust Probabilistic Modeling with Bayesian Data Reweighting." International Conference on Machine Learning, 2017.

Markdown

[Wang et al. "Robust Probabilistic Modeling with Bayesian Data Reweighting." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/wang2017icml-robust-a/)

BibTeX

@inproceedings{wang2017icml-robust-a,
  title     = {{Robust Probabilistic Modeling with Bayesian Data Reweighting}},
  author    = {Wang, Yixin and Kucukelbir, Alp and Blei, David M.},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {3646-3655},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/wang2017icml-robust-a/}
}