Quantifying Uncertainty in Foundation Models via Ensembles
Abstract
As large-scale foundation models begin to have increasing impact in real-world applications, to guarantee reliability and trustworthiness it is important for these models to "know what they don't know": to be capable of quantifying uncertainty about their own outputs. In this work, we propose disagreement of model ensembles as an effective and compute-efficient method to quantify uncertainty. We also conduct a systematic study of uncertainty quantification spanning multiple tasks - a synthetic string task, and natural language arithmetic and question-answering tasks - over a progression of increasingly out of distribution inputs. We find that considering ensemble disagreement results in improved uncertainty prediction over only considering a single model's likelihood. We hope that our investigation and results encourage more research in the area of uncertainty quantification in foundation models and the use of model ensembles.
Cite
Text
Sun et al. "Quantifying Uncertainty in Foundation Models via Ensembles." NeurIPS 2022 Workshops: RobustSeq, 2022.Markdown
[Sun et al. "Quantifying Uncertainty in Foundation Models via Ensembles." NeurIPS 2022 Workshops: RobustSeq, 2022.](https://mlanthology.org/neuripsw/2022/sun2022neuripsw-quantifying/)BibTeX
@inproceedings{sun2022neuripsw-quantifying,
title = {{Quantifying Uncertainty in Foundation Models via Ensembles}},
author = {Sun, Meiqi and Yan, Wilson and Abbeel, Pieter and Mordatch, Igor},
booktitle = {NeurIPS 2022 Workshops: RobustSeq},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/sun2022neuripsw-quantifying/}
}