Channel Pruning for Accelerating Very Deep Neural Networks

Abstract

In this paper, we introduce a new channel pruning method to accelerate very deep convolutional neural networks.Given a trained CNN model, we propose an iterative two-step algorithm to effectively prune each layer, by a LASSO regression based channel selection and least square reconstruction. We further generalize this algorithm to multi-layer and multi-branch cases. Our method reduces the accumulated error and enhance the compatibility with various architectures. Our pruned VGG-16 achieves the state-of-the-art results by 5x speed-up along with only 0.3% increase of error. More importantly, our method is able to accelerate modern networks like ResNet, Xception and suffers only 1.4%, 1.0% accuracy loss under 2x speed-up respectively, which is significant.

Cite

Text

He et al. "Channel Pruning for Accelerating Very Deep Neural Networks." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.155

Markdown

[He et al. "Channel Pruning for Accelerating Very Deep Neural Networks." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/he2017iccv-channel/) doi:10.1109/ICCV.2017.155

BibTeX

@inproceedings{he2017iccv-channel,
  title     = {{Channel Pruning for Accelerating Very Deep Neural Networks}},
  author    = {He, Yihui and Zhang, Xiangyu and Sun, Jian},
  booktitle = {International Conference on Computer Vision},
  year      = {2017},
  doi       = {10.1109/ICCV.2017.155},
  url       = {https://mlanthology.org/iccv/2017/he2017iccv-channel/}
}