Data-Efficient Supervised Fine-Tuning of Language Models Using Optimal Design

Abstract

Supervised fine-tuning (SFT) is a standard approach to adapting large language models (LLMs) to new domains. In this work, we improve the statistical efficiency of SFT by selecting an informative subset of training examples. Specifically, for a fixed budget of training examples, which determines the computational cost of fine-tuning, we determine the most informative ones. The key idea in our method is to select examples that maximize the Hessian of the log-likelihood of the LLM. We approximate it efficiently by linearizing the LLM at the last layer using multinomial logistic regression models. Our approach is computationally efficient, analyzable, and performs well empirically. We demonstrate this on several problems, and back our claims with both quantitative results and an LLM evaluation.

Cite

Text

Deb et al. "Data-Efficient Supervised Fine-Tuning of Language Models Using Optimal Design." ICLR 2025 Workshops: Data_Problems, 2025.

Markdown

[Deb et al. "Data-Efficient Supervised Fine-Tuning of Language Models Using Optimal Design." ICLR 2025 Workshops: Data_Problems, 2025.](https://mlanthology.org/iclrw/2025/deb2025iclrw-dataefficient/)

BibTeX

@inproceedings{deb2025iclrw-dataefficient,
  title     = {{Data-Efficient Supervised Fine-Tuning of Language Models Using Optimal Design}},
  author    = {Deb, Rohan and Thekumparampil, Kiran Koshy and Kalantari, Kousha and Hiranandani, Gaurush and Sabach, Shoham and Kveton, Branislav},
  booktitle = {ICLR 2025 Workshops: Data_Problems},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/deb2025iclrw-dataefficient/}
}