Learning Model Uncertainty as Variance-Minimizing Instance Weights
Abstract
Predictive uncertainty--a model’s self-awareness regarding its accuracy on an input--is key for both building robust models via training interventions and for test-time applications such as selective classification. We propose a novel instance-conditional reweighting approach that captures predictive uncertainty using an auxiliary network, and unifies these train- and test-time applications. The auxiliary network is trained using a meta-objective in a bilevel optimization framework. A key contribution of our proposal is the meta-objective of minimizing dropout variance, an approximation of Bayesian predictive uncertainty, We show in controlled experiments that we effectively capture diverse specific notions of uncertainty through this meta-objective, while previous approaches only capture certain aspects. These results translate to significant gains in real-world settings–selective classification, label noise, domain adaptation, calibration–and across datasets–Imagenet, Cifar100, diabetic retinopathy, Camelyon, WILDs, Imagenet-C,-A,-R, Clothing-1.6M, etc. For Diabetic Retinopathy, we see upto 3.4\%/3.3\% accuracy & AUC gains over SOTA in selective classification. We also improve upon large-scale pretrained models such as PLEX.
Cite
Text
Jain et al. "Learning Model Uncertainty as Variance-Minimizing Instance Weights." International Conference on Learning Representations, 2024.Markdown
[Jain et al. "Learning Model Uncertainty as Variance-Minimizing Instance Weights." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/jain2024iclr-learning/)BibTeX
@inproceedings{jain2024iclr-learning,
title = {{Learning Model Uncertainty as Variance-Minimizing Instance Weights}},
author = {Jain, Nishant and Shanmugam, Karthikeyan and Shenoy, Pradeep},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/jain2024iclr-learning/}
}