DSC: Dense-Sparse Convolution for Vectorized Inference of Convolutional Neural Networks
Abstract
The efficient applications of Convolutional Neural Networks (CNNs) in automotive-rated and safety critical hardware-accelerators require an interplay of DNN design optimization, programming techniques and hardware resources. Ad-hoc pruning would result in irregular sparsity and compression leading in very inefficient real world applications. Therefore, the proposed methodology, called Dense-Sparse Convolution, makes use of the right balance between pruning regularity, quantization and the underlying vectorized hardware. Different word length compute units, e.g. CPU, are used for low latency inference of the spares CNNs. The proposed open source CPU-kernel scales along with the vector word length and the number of cores.
Cite
Text
Frickenstein et al. "DSC: Dense-Sparse Convolution for Vectorized Inference of Convolutional Neural Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019. doi:10.1109/CVPRW.2019.00175Markdown
[Frickenstein et al. "DSC: Dense-Sparse Convolution for Vectorized Inference of Convolutional Neural Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/frickenstein2019cvprw-dsc/) doi:10.1109/CVPRW.2019.00175BibTeX
@inproceedings{frickenstein2019cvprw-dsc,
title = {{DSC: Dense-Sparse Convolution for Vectorized Inference of Convolutional Neural Networks}},
author = {Frickenstein, Alexander and Vemparala, Manoj Rohit and Unger, Christian and Ayar, Fatih and Stechele, Walter},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2019},
pages = {1353-1360},
doi = {10.1109/CVPRW.2019.00175},
url = {https://mlanthology.org/cvprw/2019/frickenstein2019cvprw-dsc/}
}