When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs

Abstract

Reasoning-enhanced large language models (RLLMs), whether explicitly trained for reasoning or prompted via chain-of-thought (CoT), have achieved state-of-the-art performance on many complex reasoning tasks. However, we uncover a surprising and previously overlooked phenomenon: explicit CoT reasoning can significantly degrade instruction-following accuracy. Evaluating 20+ models on two benchmarks: IFEval (with simple, rule-verifiable constraints) and ComplexBench (with complex, compositional constraints), we consistently observe performance drops when CoT prompting is applied. Through large-scale case studies and an attention-based analysis, we identify common patterns where reasoning either helps (e.g., with formatting or lexical precision) or hurts (e.g., by neglecting simple constraints or introducing unnecessary content). We propose a metric, constraint attention, to quantify model focus during generation and show that CoT reasoning often diverts attention away from instruction-relevant tokens. To mitigate these effects, we introduce and evaluate four strategies: in-context learning, self-reflection, self-selective reasoning, and classifier-selective reasoning. Our results demonstrate that selective reasoning strategies, particularly classifier-selective reasoning, can substantially recover lost performance. To our knowledge, this is the first work to systematically expose reasoning-induced failures in instruction-following and offer practical mitigation strategies.

Cite

Text

Li et al. "When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs." Advances in Neural Information Processing Systems, 2025.

Markdown

[Li et al. "When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/li2025neurips-thinking/)

BibTeX

@inproceedings{li2025neurips-thinking,
  title     = {{When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs}},
  author    = {Li, Xiaomin and Yu, Zhou and Zhang, Zhiwei and Chen, Xupeng and Zhang, Ziji and Zhuang, Yingying and Sadagopan, Narayanan and Beniwal, Anurag},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/li2025neurips-thinking/}
}