Overtrained Language Models Are Harder to Fine-Tune

Abstract

Large language models are pre-trained on ever-growing token budgets under the assumption that better pre-training performance translates to improved downstream models. In this work, we challenge this assumption and show that extended pre-training can make models harder to fine-tune, leading to degraded final performance. We term this phenomenon catastrophic overtraining. For example, the instruction-tuned OLMo-1B model pre-trained on 3T tokens leads to over 2% worse performance on multiple standard LLM benchmarks than its 2.3T token counterpart. Through controlled experiments and theoretical analysis, we show that catastrophic overtraining arises from a systematic increase in the broad sensitivity of pre-trained parameters to modifications, including but not limited to fine-tuning. Our findings call for a critical reassessment of pre-training design that considers the downstream adaptability of the model.

Cite

Text

Springer et al. "Overtrained Language Models Are Harder to Fine-Tune." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Springer et al. "Overtrained Language Models Are Harder to Fine-Tune." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/springer2025icml-overtrained/)

BibTeX

@inproceedings{springer2025icml-overtrained,
  title     = {{Overtrained Language Models Are Harder to Fine-Tune}},
  author    = {Springer, Jacob Mitchell and Goyal, Sachin and Wen, Kaiyue and Kumar, Tanishq and Yue, Xiang and Malladi, Sadhika and Neubig, Graham and Raghunathan, Aditi},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {56719-56789},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/springer2025icml-overtrained/}
}