Practical Network Acceleration with Tiny Sets

Abstract

Due to data privacy issues, accelerating networks with tiny training sets has become a critical need in practice. Previous methods mainly adopt filter-level pruning to accelerate networks with scarce training samples. In this paper, we reveal that dropping blocks is a fundamentally superior approach in this scenario. It enjoys a higher acceleration ratio and results in a better latency-accuracy performance under the few-shot setting. To choose which blocks to drop, we propose a new concept namely recoverability to measure the difficulty of recovering the compressed network. Our recoverability is efficient and effective for choosing which blocks to drop. Finally, we propose an algorithm named PRACTISE to accelerate networks using only tiny sets of training images. PRACTISE outperforms previous methods by a significant margin. For 22% latency reduction, PRACTISE surpasses previous methods by on average 7% on ImageNet-1k. It also enjoys high generalization ability, working well under data-free or out-of-domain data settings, too. Our code is at https://github.com/DoctorKey/Practise.

Cite

Text

Wang and Wu. "Practical Network Acceleration with Tiny Sets." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.01947

Markdown

[Wang and Wu. "Practical Network Acceleration with Tiny Sets." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/wang2023cvpr-practical-b/) doi:10.1109/CVPR52729.2023.01947

BibTeX

@inproceedings{wang2023cvpr-practical-b,
  title     = {{Practical Network Acceleration with Tiny Sets}},
  author    = {Wang, Guo-Hua and Wu, Jianxin},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {20331-20340},
  doi       = {10.1109/CVPR52729.2023.01947},
  url       = {https://mlanthology.org/cvpr/2023/wang2023cvpr-practical-b/}
}