Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles

Abstract

Fine-tuning large language models can improve task specific performance, although a general understanding of what the fine-tuned model has learned, forgotten and how to trust its predictions is still missing. We derive principled uncertainty quantification for fine-tuned LLMs with posterior approximations using computationally efficient low-rank adaptation ensembles. We analyze three common multiple-choice datasets using low-rank adaptation ensembles based on Mistral-7b, and draw quantitative and qualitative conclusions on their perceived complexity and balance between retained prior knowledge and domain specific adaptation during and after fine-tuning. We identify unexpected retention of acquired knowledge during fine-tuning in the overfitting regime.

Cite

Text

Balabanov and Linander. "Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles." ICLR 2025 Workshops: QUESTION, 2025.

Markdown

[Balabanov and Linander. "Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles." ICLR 2025 Workshops: QUESTION, 2025.](https://mlanthology.org/iclrw/2025/balabanov2025iclrw-uncertainty/)

BibTeX

@inproceedings{balabanov2025iclrw-uncertainty,
  title     = {{Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles}},
  author    = {Balabanov, Oleksandr and Linander, Hampus},
  booktitle = {ICLR 2025 Workshops: QUESTION},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/balabanov2025iclrw-uncertainty/}
}