Layer Collapse Can Be Induced by Unstructured Pruning

Abstract

Unstructured pruning is a popular compression method for efficiently reducing model parameters. However, while it effectively decreases the number of parameters, it is commonly believed that unstructured pruning cannot shorten the computational critical path, i.e., the maximum number of layers traversed during forward propagation. In this paper, we study when and how unstructured pruning can yield structural effects. For rectifier-activated networks, we introduce the notion of neuron entropy, which quantifies the degree of nonlinearity utilization. We show that magnitude-based pruning naturally lowers this entropy, sometimes down to zero-entropy layers that become linearizable and can thus be removed. Building on this insight, we propose a method that leverages "unstructured" pruning to favor sparsity in low-entropy layers, enabling their complete removal. We validate the phenomenon across CNNs, Vision Transformers, and NLP models: unstructured pruning can induce effective layer removal with little or no performance degradation in over-parameterized networks. Our code is available at https://github.com/ZhuLIAO001/NEPENTHE.git.

Cite

Text

Liao et al. "Layer Collapse Can Be Induced by Unstructured Pruning." Transactions on Machine Learning Research, 2026.

Markdown

[Liao et al. "Layer Collapse Can Be Induced by Unstructured Pruning." Transactions on Machine Learning Research, 2026.](https://mlanthology.org/tmlr/2026/liao2026tmlr-layer/)

BibTeX

@article{liao2026tmlr-layer,
  title     = {{Layer Collapse Can Be Induced by Unstructured Pruning}},
  author    = {Liao, Zhu and Quétu, Victor and Nguyen, Van-Tam and Tartaglione, Enzo},
  journal   = {Transactions on Machine Learning Research},
  year      = {2026},
  url       = {https://mlanthology.org/tmlr/2026/liao2026tmlr-layer/}
}