ICE-Pick: Iterative Cost-Efficient Pruning for DNNs
Abstract
Pruning is one of the main compression methods for Deep Neural Networks (DNNs), where less relevant parameters are removed from a DNN model to reduce its memory footprint. To get better final accuracy, pruning is often performed iteratively with increasing amounts of parameters being removed in each step, and fine-tuning (i.e., additional training epochs) being applied to the remaining parameters. However, this process can be very time-consuming, since the finetuning process is applied after every pruning step and calculates gradients for the whole model. Motivated by these overheads, in this paper we propose ICE-Pick, a novel threshold-guided finetuning method which freezes less sensitive layers and leverages a custom pruning-aware learning rate scheduler. We evaluate ICE-Pick using ResNet-110, ResNet-152, and MobileNetV2 (all defined for CIFAR-10), and show that it can save up to 87.6% of the pruning time while maintaining accuracy.
Cite
Text
Hu et al. "ICE-Pick: Iterative Cost-Efficient Pruning for DNNs." ICML 2023 Workshops: NCW, 2023.Markdown
[Hu et al. "ICE-Pick: Iterative Cost-Efficient Pruning for DNNs." ICML 2023 Workshops: NCW, 2023.](https://mlanthology.org/icmlw/2023/hu2023icmlw-icepick/)BibTeX
@inproceedings{hu2023icmlw-icepick,
title = {{ICE-Pick: Iterative Cost-Efficient Pruning for DNNs}},
author = {Hu, Wenhao and Gibson, Perry and Cano, José},
booktitle = {ICML 2023 Workshops: NCW},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/hu2023icmlw-icepick/}
}