Bayesian Quantification with Black-Box Estimators

Abstract

Understanding how different classes are distributed in an unlabeled data set is important for the calibration of probabilistic classifiers and uncertainty quantification. Methods like adjusted classify and count, black-box shift estimators, and invariant ratio estimators use an auxiliary and potentially biased black-box classifier trained on a different data set to estimate the class distribution on the current data set and yield asymptotic guarantees under weak assumptions. We demonstrate that these algorithms are closely related to the inference in a particular probabilistic graphical model approximating the assumed ground-truth generative process, and we propose a Bayesian estimator. Then, we discuss an efficient Markov chain Monte Carlo sampling scheme for the introduced model and show an asymptotic consistency guarantee in the large-data limit. We compare the introduced model against the established point estimators in a variety of scenarios, and show it is competitive, and in some cases superior, with the non-Bayesian alternatives.

Cite

Text

Ziegler and Czyż. "Bayesian Quantification with Black-Box Estimators." Transactions on Machine Learning Research, 2024.

Markdown

[Ziegler and Czyż. "Bayesian Quantification with Black-Box Estimators." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/ziegler2024tmlr-bayesian/)

BibTeX

@article{ziegler2024tmlr-bayesian,
  title     = {{Bayesian Quantification with Black-Box Estimators}},
  author    = {Ziegler, Albert and Czyż, Paweł},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/ziegler2024tmlr-bayesian/}
}