COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level Pruning
Abstract
Neural network compression empowers the effective yet unwieldy deep convolutional neural networks (CNN) to be deployed in resource-constrained scenarios. Most state-of-the-art approaches prune the model in filter-level according to the "importance" of filters. Despite their success, we notice they suffer from at least two of the following problems: 1) The redundancy among filters is not considered because the importance is evaluated independently. 2) Cross-layer filter comparison is unachievable since the importance is defined locally within each layer. Consequently, we must manually specify layer-wise pruning ratios. 3) They are prone to generate sub-optimal solutions because they neglect the inequality between reducing parameters and reducing computational cost. Reducing the same number of parameters in different positions in the network may reduce different computational cost. To address the above problems, we develop a novel algorithm named as COP (correlation-based pruning), which can detect the redundant filters efficiently. We enable the cross-layer filter comparison through global normalization. We add parameter-quantity and computational-cost regularization terms to the importance, which enables the users to customize the compression according to their preference (smaller or faster). Extensive experiments have shown COP outperforms the others significantly. The code is released at https://github.com/ZJULearning/COP.
Cite
Text
Wang et al. "COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level Pruning." International Joint Conference on Artificial Intelligence, 2019. doi:10.24963/IJCAI.2019/525Markdown
[Wang et al. "COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level Pruning." International Joint Conference on Artificial Intelligence, 2019.](https://mlanthology.org/ijcai/2019/wang2019ijcai-cop/) doi:10.24963/IJCAI.2019/525BibTeX
@inproceedings{wang2019ijcai-cop,
title = {{COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level Pruning}},
author = {Wang, Wenxiao and Fu, Cong and Guo, Jishun and Cai, Deng and He, Xiaofei},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2019},
pages = {3785-3791},
doi = {10.24963/IJCAI.2019/525},
url = {https://mlanthology.org/ijcai/2019/wang2019ijcai-cop/}
}