Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles
Abstract
Fine-tuning large language models can improve task specific performance, although a general understanding of what the fine-tuned model has learned, forgotten and how to trust its predictions is still missing. We derive principled uncertainty quantification for fine-tuned LLMs with posterior approximations using computationally efficient low-rank adaptation ensembles. We analyze three common multiple-choice datasets using low-rank adaptation ensembles based on Mistral-7b, and draw quantitative and qualitative conclusions on their perceived complexity and balance between retained prior knowledge and domain specific adaptation during and after fine-tuning. We identify unexpected retention of acquired knowledge during fine-tuning in the overfitting regime.
Cite
Text
Balabanov and Linander. "Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles." ICLR 2025 Workshops: QUESTION, 2025.Markdown
[Balabanov and Linander. "Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles." ICLR 2025 Workshops: QUESTION, 2025.](https://mlanthology.org/iclrw/2025/balabanov2025iclrw-uncertainty/)BibTeX
@inproceedings{balabanov2025iclrw-uncertainty,
title = {{Uncertainty Quantification in Fine-Tuned LLMs Using LoRA Ensembles}},
author = {Balabanov, Oleksandr and Linander, Hampus},
booktitle = {ICLR 2025 Workshops: QUESTION},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/balabanov2025iclrw-uncertainty/}
}