Fine-Tuning Pretrained Models with NVIB for Improved Generalisation
Abstract
Fine-tuned pretrained attention-based models often struggle with generalisation, leading to poor performance on tasks like out-of-domain transfer, distribution shifts, and few-shot learning. This limitation is prevalent across modalities such as speech, text, graphs, and vision. Nonparametric Variational Information Bottleneck (NVIB) is an attention-based information-theoretic regulariser applicable to pretrained models that has been shown to improve generalisation. However, prior work has applied NVIB only to the text modality and without fine-tuning. We investigate whether NVIB’s ability to remove information from pretrained embeddings helps the model avoid spurious correlations with noisy and superficial features during fine-tuning. We are the first to integrate NVIB regularisation during fine-tuning across multiple diverse models and modalities. This required modifications to the architecture which enhance adaptability and stability during fine-tuning and simplify the evaluation. We found improved out-of-distribution generalisation in: speech quality assessment and language identification, text with induced attention sparsity, graph-based link prediction, and few-shot image classification.
Cite
Text
Fehr et al. "Fine-Tuning Pretrained Models with NVIB for Improved Generalisation." ICLR 2025 Workshops: SCSL, 2025.Markdown
[Fehr et al. "Fine-Tuning Pretrained Models with NVIB for Improved Generalisation." ICLR 2025 Workshops: SCSL, 2025.](https://mlanthology.org/iclrw/2025/fehr2025iclrw-finetuning/)BibTeX
@inproceedings{fehr2025iclrw-finetuning,
title = {{Fine-Tuning Pretrained Models with NVIB for Improved Generalisation}},
author = {Fehr, Fabio James and Baia, Alina Elena and Chang, Xiaoguang and Coman, Andrei Catalin and El Hajal, Karl and El Zein, Dina and Kumar, Shashi and Gomez, Juan Pablo Zuluaga and Cavallaro, Andrea and Teney, Damien and Henderson, James},
booktitle = {ICLR 2025 Workshops: SCSL},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/fehr2025iclrw-finetuning/}
}