ICE-Pick: Iterative Cost-Efficient Pruning for DNNs

Abstract

Pruning is one of the main compression methods for Deep Neural Networks (DNNs), where less relevant parameters are removed from a DNN model to reduce its memory footprint. To get better final accuracy, pruning is often performed iteratively with increasing amounts of parameters being removed in each step, and fine-tuning (i.e., additional training epochs) being applied to the remaining parameters. However, this process can be very time-consuming, since the finetuning process is applied after every pruning step and calculates gradients for the whole model. Motivated by these overheads, in this paper we propose ICE-Pick, a novel threshold-guided finetuning method which freezes less sensitive layers and leverages a custom pruning-aware learning rate scheduler. We evaluate ICE-Pick using ResNet-110, ResNet-152, and MobileNetV2 (all defined for CIFAR-10), and show that it can save up to 87.6% of the pruning time while maintaining accuracy.

Cite

Text

Hu et al. "ICE-Pick: Iterative Cost-Efficient Pruning for DNNs." ICML 2023 Workshops: NCW, 2023.

Markdown

[Hu et al. "ICE-Pick: Iterative Cost-Efficient Pruning for DNNs." ICML 2023 Workshops: NCW, 2023.](https://mlanthology.org/icmlw/2023/hu2023icmlw-icepick/)

BibTeX

@inproceedings{hu2023icmlw-icepick,
  title     = {{ICE-Pick: Iterative Cost-Efficient Pruning for DNNs}},
  author    = {Hu, Wenhao and Gibson, Perry and Cano, José},
  booktitle = {ICML 2023 Workshops: NCW},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/hu2023icmlw-icepick/}
}