Overtrained Language Models Are Harder to Fine-Tune

Abstract

Large language models are pre-trained with an ever-increasing token budget operating under the largely unexamined premise that better pre-training performance translates to better downstream performance. In this work, we show that this widely-held assumption is in fact false! Pre-training on extremely large number of tokens eventually makes the model harder to fine-tune leading to worse downstream performance. For instance, after instruction tuning or multimodal fine tuning, OLMo-1B models pre-trained on 3T tokens under perform their 2.3T token counterpart by over $2\%$ on standard LLM benchmarks. Controlled experiments and theoretical analysis show that the phenomenon of catastrophic overtraining is both fundamental and universal. Our results suggest that as token budgets continue to scale, models will experience increasingly severe fine-tuning degradation across a wider range of tasks. This calls for a critical reassessment of pre-training design that takes into account the entire model lifecycle.

Cite

Text

Springer et al. "Overtrained Language Models Are Harder to Fine-Tune." ICLR 2025 Workshops: ICBINB, 2025.

Markdown

[Springer et al. "Overtrained Language Models Are Harder to Fine-Tune." ICLR 2025 Workshops: ICBINB, 2025.](https://mlanthology.org/iclrw/2025/springer2025iclrw-overtrained/)

BibTeX

@inproceedings{springer2025iclrw-overtrained,
  title     = {{Overtrained Language Models Are Harder to Fine-Tune}},
  author    = {Springer, Jacob Mitchell and Goyal, Sachin and Wen, Kaiyue and Kumar, Tanishq and Yue, Xiang and Malladi, Sadhika and Neubig, Graham and Raghunathan, Aditi},
  booktitle = {ICLR 2025 Workshops: ICBINB},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/springer2025iclrw-overtrained/}
}