The Surprising Effectiveness of Test-Time Training for Few-Shot Learning
Abstract
Language models (LMs) have shown impressive performance on tasks within their training distribution, but often struggle with structurally novel tasks even when given a small number of in-context task examples. We investigate the effectiveness of test-time training (TTT)—temporarily updating model parameters during inference using a loss derived from input data—as a mechanism for improving LMs’ reasoning and few-shot learning capabilities. On the Abstraction and Reasoning Corpus (ARC), performing TTT with in-context examples yields up to $6\times$ higher accuracy compared to fine-tuned baselines—reaching $53.0%$ on the public validation set with an 8B-parameter LM and $61.9%$ when ensembled with program-synthesis methods, matching average human performance. On BIG-Bench Hard (BBH), TTT on in-context examples surpasses standard few-shot prompting in the $10$-shot setting by $7.3$ percentage points ($50.5%$ to $57.8%$). Our findings highlight the limitations of in-context learning for novel tasks and demonstrate the potential of test-time training to enhance language model adaptability.
Cite
Text
Akyürek et al. "The Surprising Effectiveness of Test-Time Training for Few-Shot Learning." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Akyürek et al. "The Surprising Effectiveness of Test-Time Training for Few-Shot Learning." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/akyurek2025icml-surprising/)BibTeX
@inproceedings{akyurek2025icml-surprising,
title = {{The Surprising Effectiveness of Test-Time Training for Few-Shot Learning}},
author = {Akyürek, Ekin and Damani, Mehul and Zweiger, Adam and Qiu, Linlu and Guo, Han and Pari, Jyothish and Kim, Yoon and Andreas, Jacob},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {942-963},
volume = {267},
url = {https://mlanthology.org/icml/2025/akyurek2025icml-surprising/}
}