Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners

Abstract

Recent advancements have demonstrated that the performance of large language models (LLMs) can be significantly enhanced by scaling computational resources at test time. A common strategy involves generating multiple Chain-of-Thought (CoT) trajectories and aggregating their outputs through various selection mechanisms. This raises a fundamental question: can models with lower complexity leverage their superior generation throughput to outperform similarly sized Transformers for a fixed computational budget? To address this question and overcome the lack of strong subquadratic reasoners, we distill pure and hybrid Mamba models from pretrained Transformers. Trained on only 8 billion tokens, our distilled models show strong performance and scaling on mathematical reasoning datasets while being much faster at inference for large batches and long sequences. Despite the zero-shot performance hit due to distillation, both pure and hybrid Mamba models can scale their coverage and accuracy performance past their Transformer teachers under fixed time budgets, opening a new direction for scaling inference compute.

Cite

Text

Paliotta et al. "Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners." ICLR 2025 Workshops: LLM_Reason_and_Plan, 2025.

Markdown

[Paliotta et al. "Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners." ICLR 2025 Workshops: LLM_Reason_and_Plan, 2025.](https://mlanthology.org/iclrw/2025/paliotta2025iclrw-thinking/)

BibTeX

@inproceedings{paliotta2025iclrw-thinking,
  title     = {{Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners}},
  author    = {Paliotta, Daniele and Wang, Junxiong and Pagliardini, Matteo and Li, Kevin and Bick, Aviv and Gu, Albert and Fleuret, François and Dao, Tri},
  booktitle = {ICLR 2025 Workshops: LLM_Reason_and_Plan},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/paliotta2025iclrw-thinking/}
}