Exploring the Granularity of Sparsity in Convolutional Neural Networks
Abstract
Sparsity helps reducing the computation complexity of DNNs by skipping the multiplication with zeros. The granularity of sparsity affects the efficiency of hardware architecture and the prediction accuracy. In this paper we quantitatively measure the accuracy-sparsity relationship with different granularity. Coarse-grained sparsity brings more regular sparsity pattern, making it easier for hardware acceleration, and our experimental results show that coarsegrained sparsity have very small impact on the sparsity ratio given no loss of accuracy. Moreover, due to the index saving effect, coarse-grained sparsity is able to obtain similar or even better compression rates than fine-grained sparsity at the same accuracy threshold. Our analysis, which is based on the framework of a recent sparse convolutional neural network (SCNN) accelerator, further demonstrates that it saves 30% – 35% of memory references compared with fine-grained sparsity.
Cite
Text
Mao et al. "Exploring the Granularity of Sparsity in Convolutional Neural Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017. doi:10.1109/CVPRW.2017.241Markdown
[Mao et al. "Exploring the Granularity of Sparsity in Convolutional Neural Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017.](https://mlanthology.org/cvprw/2017/mao2017cvprw-exploring/) doi:10.1109/CVPRW.2017.241BibTeX
@inproceedings{mao2017cvprw-exploring,
title = {{Exploring the Granularity of Sparsity in Convolutional Neural Networks}},
author = {Mao, Huizi and Han, Song and Pool, Jeff and Li, Wenshuo and Liu, Xingyu and Wang, Yu and Dally, William J.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2017},
pages = {1927-1934},
doi = {10.1109/CVPRW.2017.241},
url = {https://mlanthology.org/cvprw/2017/mao2017cvprw-exploring/}
}