Model Diffusion for Certifiable Few-Shot Transfer Learning

Abstract

In contemporary deep learning, a prevalent and effective workflow for solving low-data problems is adapting powerful pre-trained foundation models (FMs) to new tasks via parameter-efficient fine-tuning (PEFT). However, while empirically effective, the resulting solutions lack generalisation guarantees to certify their accuracy - which may be required for ethical or legal reasons prior to deployment in high-importance applications. In this paper we develop a novel transfer learning approach that is designed to facilitate non-vacuous learning theoretic generalisation guarantees for downstream tasks, even in the low-shot regime. Specifically, we first use upstream tasks to train a distribution over PEFT parameters. We then learn the downstream task by a sample-and-evaluate procedure -- sampling plausible PEFTs from the trained diffusion model and selecting the one with the highest likelihood on the downstream data. Crucially, this confines our model hypothesis to a finite set of PEFT samples. In contrast to the typical continuous hypothesis spaces of neural network weights, this facilitates tighter risk certificates. We instantiate our bound and show non-trivial generalization guarantees compared to existing learning approaches which lead to vacuous bounds in the low-shot regime.

Cite

Text

Rezk et al. "Model Diffusion for Certifiable Few-Shot Transfer Learning." ICLR 2025 Workshops: WSL, 2025. doi:10.48550/arxiv.2502.06970

Markdown

[Rezk et al. "Model Diffusion for Certifiable Few-Shot Transfer Learning." ICLR 2025 Workshops: WSL, 2025.](https://mlanthology.org/iclrw/2025/rezk2025iclrw-model/) doi:10.48550/arxiv.2502.06970

BibTeX

@inproceedings{rezk2025iclrw-model,
  title     = {{Model Diffusion for Certifiable Few-Shot Transfer Learning}},
  author    = {Rezk, Fady and Lee, Royson and Gouk, Henry and Hospedales, Timothy and Kim, Minyoung},
  booktitle = {ICLR 2025 Workshops: WSL},
  year      = {2025},
  doi       = {10.48550/arxiv.2502.06970},
  url       = {https://mlanthology.org/iclrw/2025/rezk2025iclrw-model/}
}