Steering Large Reasoning Models Towards Concise Reasoning via Flow Matching

Abstract

Large Reasoning Models (LRMs) excel at complex reasoning tasks, but their efficiency is often hampered by overly verbose outputs. Prior steering methods attempt to address this issue by applying a single, global vector to hidden representations—an approach grounded in the restrictive \textit{linear representation hypothesis}. In this work, we introduce FlowSteer, a nonlinear steering method that goes beyond uniform linear shifts by learning a complete \textit{transformation between the distributions} associated with verbose and concise reasoning. This transformation is learned via \textit{Flow Matching} as a velocity field, enabling precise, input-dependent control over the model's reasoning process. By aligning steered representations with the distribution of concise-reasoning activations, FlowSteer yields more compact reasoning than the linear shifts. Across diverse reasoning benchmarks, FlowSteer demonstrates strong task performance and token efficiency compared to leading inference-time baselines. Our work demonstrates that modeling the full distributional transport with generative techniques offers a more effective and principled foundation for controlling LRMs.

Cite

Text

Li et al. "Steering Large Reasoning Models Towards Concise Reasoning via Flow Matching." Transactions on Machine Learning Research, 2026.

Markdown

[Li et al. "Steering Large Reasoning Models Towards Concise Reasoning via Flow Matching." Transactions on Machine Learning Research, 2026.](https://mlanthology.org/tmlr/2026/li2026tmlr-steering/)

BibTeX

@article{li2026tmlr-steering,
  title     = {{Steering Large Reasoning Models Towards Concise Reasoning via Flow Matching}},
  author    = {Li, Yawei and Bergner, Benjamin and Zhao, Yinghan and Patil, Vihang Prakash and Chen, Bei and Wang, Cheng},
  journal   = {Transactions on Machine Learning Research},
  year      = {2026},
  url       = {https://mlanthology.org/tmlr/2026/li2026tmlr-steering/}
}