Finetuning Language Models to Emit Linguistic Expressions of Uncertainty

Abstract

Large language models (LLMs) are increasingly employed in information-seeking and decision-making tasks. Despite their broad utility, LLMs tend to generate information that conflict with real-world facts, and their persuasive style can make these inaccuracies appear confident and convincing. As a result, end-users struggle to consistently align the confidence expressed by LLMs with the accuracy of their predictions, often leading to either blind trust in all outputs or a complete disregard for their reliability. In this work, we explore supervised fine-tuning on uncertainty-augmented predictions as a method to develop models that produce linguistic expressions of uncertainty. Specifically, we measure the calibration of pre-trained models and fine-tune language models to generate calibrated linguistic expressions of uncertainty. Through experiments on various question-answering datasets, we demonstrate that LLMs are well-calibrated in assessing their predictions, and supervised fine-tuning based on the model’s own confidence leads to well-calibrated expressions of uncertainty, particularly for single-claim answers.

Cite

Text

Chaudhry et al. "Finetuning Language Models to Emit Linguistic Expressions of Uncertainty." ICLR 2025 Workshops: QUESTION, 2025.

Markdown

[Chaudhry et al. "Finetuning Language Models to Emit Linguistic Expressions of Uncertainty." ICLR 2025 Workshops: QUESTION, 2025.](https://mlanthology.org/iclrw/2025/chaudhry2025iclrw-finetuning/)

BibTeX

@inproceedings{chaudhry2025iclrw-finetuning,
  title     = {{Finetuning Language Models to Emit Linguistic Expressions of Uncertainty}},
  author    = {Chaudhry, Arslan and Thiagarajan, Sridhar and Gorur, Dilan},
  booktitle = {ICLR 2025 Workshops: QUESTION},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/chaudhry2025iclrw-finetuning/}
}