Prune Efficiently by Soft Pruning

Abstract

Embedded systems are power sensitive and have limited memory, hence inferencing large networks on such systems is difficult. Pruning techniques have been instrumental in enhancing the efficiency of state-of-the-art convolutional neural networks on embedded systems. Traditional algorithms tend to eliminate weights abruptly during training, which may not provide the best accuracy. The proposed approach, called Soft Pruning using Weight Blending algorithm (SPWB), is designed to retain critical information by incrementally reducing the network’s weights to zero. Additionally, our method of channel pruning is cognizant of connections, allowing for optimal pruning that renders the network compatible with various inference engines. The findings demonstrate that SPWB algorithm can reduce computational complexity(measured in FLOPs) to half for ResNet50, with only a minimal impact—a 0.65% decrease—in top-1 accuracy on the ImageNet dataset. We also present our pruning results for both unstructured weight sparsity as well as channel sparsity. Our method is easy to use and provides enhancement in the network’s performance and efficiency without compromising accuracy. The method is available as a python package and can be easily integrated to other training scripts. The code is publicly available here.

Cite

Text

Agarwal et al. "Prune Efficiently by Soft Pruning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024. doi:10.1109/CVPRW63382.2024.00226

Markdown

[Agarwal et al. "Prune Efficiently by Soft Pruning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024.](https://mlanthology.org/cvprw/2024/agarwal2024cvprw-prune/) doi:10.1109/CVPRW63382.2024.00226

BibTeX

@inproceedings{agarwal2024cvprw-prune,
  title     = {{Prune Efficiently by Soft Pruning}},
  author    = {Agarwal, Parakh and Mathew, Manu and Patel, Kunal Ranjan and Tripathi, Varun and Swami, Pramod},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2024},
  pages     = {2210-2217},
  doi       = {10.1109/CVPRW63382.2024.00226},
  url       = {https://mlanthology.org/cvprw/2024/agarwal2024cvprw-prune/}
}