PDP: Parameter-Free Differentiable Pruning Is All You Need
Abstract
In this paper, we propose an efficient yet effective train-time pruning scheme, Parameter-free Differentiable Pruning (PDP), which offers state-of-the-art qualities in model size, accuracy, and training cost. PDP uses a dynamic function of weights during training to generate soft pruning masks for the weights in a parameter-free manner for a given pruning target. While differentiable, the simplicity and efficiency of PDP make it universal enough to deliver state-of-the-art random/structured/channel pruning results on various vision models. For example, for MobileNet-v1, PDP can achieve 68.2% top-1 ImageNet1k accuracy at 86.6% sparsity, which is 1.7% higher accuracy than those from the state-of-the-art algorithms. PDP also improved the top-1 ImageNet1k accuracy of ResNet18 by over 3.6% and reduced the top-1 ImageNet1k accuracy of ResNet50 by 0.6% from the state-of-the-art.
Cite
Text
Cho et al. "PDP: Parameter-Free Differentiable Pruning Is All You Need." ICML 2023 Workshops: Differentiable_Almost_Everything, 2023.Markdown
[Cho et al. "PDP: Parameter-Free Differentiable Pruning Is All You Need." ICML 2023 Workshops: Differentiable_Almost_Everything, 2023.](https://mlanthology.org/icmlw/2023/cho2023icmlw-pdp/)BibTeX
@inproceedings{cho2023icmlw-pdp,
title = {{PDP: Parameter-Free Differentiable Pruning Is All You Need}},
author = {Cho, Minsik and Adya, Saurabh and Naik, Devang},
booktitle = {ICML 2023 Workshops: Differentiable_Almost_Everything},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/cho2023icmlw-pdp/}
}