Estimating Fréchet Bounds for Validating Programmatic Weak Supervision
Abstract
We develop methods for estimating Fréchet bounds on (possibly high-dimensional) distribution classes in which some variables are continuous-valued. We establish the statistical correctness of the computed bounds under uncertainty in the marginal constraints and demonstrate the usefulness of our algorithms by evaluating the performance of machine learning (ML) models trained with programmatic weak supervision (PWS). PWS is a framework for principled learning from weak supervision inputs (e.g., crowdsourced labels, knowledge bases, pre-trained models on related tasks, etc.), and it has achieved remarkable success in many areas of science and engineering. Unfortunately, it is generally difficult to validate the performance of ML models trained with PWS due to the absence of labeled data. Our algorithms address this issue by estimating sharp lower and upper bounds for performance metrics such as accuracy/recall/precision drawing connections to tools from computational optimal transport.
Cite
Text
Polo et al. "Estimating Fréchet Bounds for Validating Programmatic Weak Supervision." NeurIPS 2023 Workshops: OTML, 2023.Markdown
[Polo et al. "Estimating Fréchet Bounds for Validating Programmatic Weak Supervision." NeurIPS 2023 Workshops: OTML, 2023.](https://mlanthology.org/neuripsw/2023/polo2023neuripsw-estimating/)BibTeX
@inproceedings{polo2023neuripsw-estimating,
title = {{Estimating Fréchet Bounds for Validating Programmatic Weak Supervision}},
author = {Polo, Felipe Maia and Yurochkin, Mikhail and Banerjee, Moulinath and Maity, Subha and Sun, Yuekai},
booktitle = {NeurIPS 2023 Workshops: OTML},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/polo2023neuripsw-estimating/}
}