Truly Scale-Equivariant Deep Nets with Fourier Layers
Abstract
In computer vision, models must be able to adapt to changes in image resolution to effectively carry out tasks such as image segmentation; This is known as scale-equivariance. Recent works have made progress in developing scale-equivariant convolutional neural networks, e.g., through weight-sharing and kernel resizing. However, these networks are not truly scale-equivariant in practice. Specifically, they do not consider anti-aliasing as they formulate the down-scaling operation in the continuous domain. To address this shortcoming, we directly formulate down-scaling in the discrete domain with consideration of anti-aliasing. We then propose a novel architecture based on Fourier layers to achieve truly scale-equivariant deep nets, i.e., absolute zero equivariance-error. Following prior works, we test this model on MNIST-scale and STL-10 datasets. Our proposed model achieves competitive classification performance while maintaining zero equivariance-error.
Cite
Text
Rahman and Yeh. "Truly Scale-Equivariant Deep Nets with Fourier Layers." Neural Information Processing Systems, 2023.Markdown
[Rahman and Yeh. "Truly Scale-Equivariant Deep Nets with Fourier Layers." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/rahman2023neurips-truly/)BibTeX
@inproceedings{rahman2023neurips-truly,
title = {{Truly Scale-Equivariant Deep Nets with Fourier Layers}},
author = {Rahman, Md Ashiqur and Yeh, Raymond A.},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/rahman2023neurips-truly/}
}