Scaling-Translation-Equivariant Networks with Decomposed Convolutional Filters
Abstract
Encoding the scale information explicitly into the representation learned by a convolutional neural network (CNN) is beneficial for many computer vision tasks especially when dealing with multiscale inputs. We study, in this paper, a scaling-translation-equivariant ($\mathcal{ST}$-equivariant) CNN with joint convolutions across the space and the scaling group, which is shown to be both sufficient and necessary to achieve equivariance for the regular representation of the scaling-translation group $\mathcal{ST}$. To reduce the model complexity and computational burden, we decompose the convolutional filters under two pre-fixed separable bases and truncate the expansion to low-frequency components. A further benefit of the truncated filter expansion is the improved deformation robustness of the equivariant representation, a property which is theoretically analyzed and empirically verified. Numerical experiments demonstrate that the proposed scaling-translation-equivariant network with decomposed convolutional filters (ScDCFNet) achieves significantly improved performance in multiscale image classification and better interpretability than regular CNNs at a reduced model size.
Cite
Text
Zhu et al. "Scaling-Translation-Equivariant Networks with Decomposed Convolutional Filters." Journal of Machine Learning Research, 2022.Markdown
[Zhu et al. "Scaling-Translation-Equivariant Networks with Decomposed Convolutional Filters." Journal of Machine Learning Research, 2022.](https://mlanthology.org/jmlr/2022/zhu2022jmlr-scalingtranslationequivariant/)BibTeX
@article{zhu2022jmlr-scalingtranslationequivariant,
title = {{Scaling-Translation-Equivariant Networks with Decomposed Convolutional Filters}},
author = {Zhu, Wei and Qiu, Qiang and Calderbank, Robert and Sapiro, Guillermo and Cheng, Xiuyuan},
journal = {Journal of Machine Learning Research},
year = {2022},
pages = {1-45},
volume = {23},
url = {https://mlanthology.org/jmlr/2022/zhu2022jmlr-scalingtranslationequivariant/}
}