Proxsparse: Regularized Learning of Semi-Structured Sparsity Masks for Pretrained LLMs
Abstract
Large Language Models (LLMs) have demonstrated exceptional performance in natural language processing tasks, yet their massive size makes serving them inefficient and costly. Semi-structured pruning has emerged as an effective method for model acceleration, but existing approaches are suboptimal because they focus on local, layer-wise optimizations using heuristic rules, failing to leverage global feedback. We present ProxSparse, a learning-based framework for mask selection enabled by regularized optimization. ProxSparse transforms the rigid, non-differentiable mask selection process into a smoother optimization procedure, allowing gradual mask exploration with flexibility. ProxSparse does not involve additional weight updates once the mask is determined. Our extensive evaluations on 7 widely used models show that ProxSparse consistently outperforms previously proposed semi-structured mask selection methods with significant improvement, demonstrating the effectiveness of our learned approach towards semi-structured pruning.
Cite
Text
Liu et al. "Proxsparse: Regularized Learning of Semi-Structured Sparsity Masks for Pretrained LLMs." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Liu et al. "Proxsparse: Regularized Learning of Semi-Structured Sparsity Masks for Pretrained LLMs." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/liu2025icml-proxsparse/)BibTeX
@inproceedings{liu2025icml-proxsparse,
title = {{Proxsparse: Regularized Learning of Semi-Structured Sparsity Masks for Pretrained LLMs}},
author = {Liu, Hongyi and Saha, Rajarshi and Jia, Zhen and Park, Youngsuk and Huang, Jiaji and Sabach, Shoham and Wang, Yu-Xiang and Karypis, George},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {39405-39419},
volume = {267},
url = {https://mlanthology.org/icml/2025/liu2025icml-proxsparse/}
}