Rethinking the Value of Training-Free Structured Pruning of LLMs
Abstract
This paper investigates the effectiveness of training-free structured pruning techniques for Large Language Models (LLMs), with a particular focus on depth and width pruning strategies. Through an extensive empirical evaluation across a diverse range of tasks, datasets and modalities, we reveal critical limitations in current pruning methods. While some tasks exhibit minimal performance degradation, others face significant deterioration, even at low pruning rates, contradicting prior findings that often rely on selective benchmarks. Our analysis also finds that depth pruning, despite its simplicity, usually outperforms the more granular width pruning approaches in maintaining downstream task performance. Our findings highlight that existing evaluations of pruned LLMs often overstate their effectiveness due to incomplete or limited evaluation tasks, necessitating a critical reassessment of the true value of pruning and emphasizing the need to explore more robust pruning algorithms.
Cite
Text
Lele et al. "Rethinking the Value of Training-Free Structured Pruning of LLMs." Transactions on Machine Learning Research, 2025.Markdown
[Lele et al. "Rethinking the Value of Training-Free Structured Pruning of LLMs." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/lele2025tmlr-rethinking/)BibTeX
@article{lele2025tmlr-rethinking,
title = {{Rethinking the Value of Training-Free Structured Pruning of LLMs}},
author = {Lele, Nahush and Chavan, Arnav and Thakur, Aryamaan and Gupta, Deepak},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/lele2025tmlr-rethinking/}
}