Truly Scale-Equivariant Deep Nets with Fourier Layers

Abstract

In computer vision, models must be able to adapt to changes in image resolution to effectively carry out tasks such as image segmentation; This is known as scale-equivariance. Recent works have made progress in developing scale-equivariant convolutional neural networks, e.g., through weight-sharing and kernel resizing. However, these networks are not truly scale-equivariant in practice. Specifically, they do not consider anti-aliasing as they formulate the down-scaling operation in the continuous domain. To address this shortcoming, we directly formulate down-scaling in the discrete domain with consideration of anti-aliasing. We then propose a novel architecture based on Fourier layers to achieve truly scale-equivariant deep nets, i.e., absolute zero equivariance-error. Following prior works, we test this model on MNIST-scale and STL-10 datasets. Our proposed model achieves competitive classification performance while maintaining zero equivariance-error.

Cite

Text

Rahman and Yeh. "Truly Scale-Equivariant Deep Nets with Fourier Layers." Neural Information Processing Systems, 2023.

Markdown

[Rahman and Yeh. "Truly Scale-Equivariant Deep Nets with Fourier Layers." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/rahman2023neurips-truly/)

BibTeX

@inproceedings{rahman2023neurips-truly,
  title     = {{Truly Scale-Equivariant Deep Nets with Fourier Layers}},
  author    = {Rahman, Md Ashiqur and Yeh, Raymond A.},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/rahman2023neurips-truly/}
}