Decoupled Convolutions for CNNs
Abstract
In this paper, we are interested in designing small CNNs by decoupling the convolution along the spatial and channel domains. Most existing decoupling techniques focus on approximating the filter matrix through decomposition. In contrast, we provide a two-step interpretation of the standard convolution from the filter at a single location to all locations, which is exactly equivalent to the standard convolution. Motivated by the observations in our decoupling view, we propose an effective approach to relax the sparsity of the filter in spatial aggregation by learning a spatial configuration, and reduce the redundancy by reducing the number of intermediate channels. Our approach achieves comparable classification performance with the standard uncoupled convolution, but with a smaller model size over CIFAR-100, CIFAR-10 and ImageNet.
Cite
Text
Xie et al. "Decoupled Convolutions for CNNs." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.11638Markdown
[Xie et al. "Decoupled Convolutions for CNNs." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/xie2018aaai-decoupled/) doi:10.1609/AAAI.V32I1.11638BibTeX
@inproceedings{xie2018aaai-decoupled,
title = {{Decoupled Convolutions for CNNs}},
author = {Xie, Guotian and Zhang, Ting and Yang, Kuiyuan and Lai, Jianhuang and Wang, Jingdong},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2018},
pages = {4284-4291},
doi = {10.1609/AAAI.V32I1.11638},
url = {https://mlanthology.org/aaai/2018/xie2018aaai-decoupled/}
}