Dilated Convolution with Learnable Spacings: Beyond Bilinear Interpolation

Abstract

Dilated Convolution with Learnable Spacings (DCLS) is a recently proposed variation of the dilated convolution in which the spacings between the non-zero elements in the kernel, or equivalently their positions, are learnable. Non-integer positions are handled via interpolation. Thanks to this trick, positions have well-defined gradients. The original DCLS used bilinear interpolation, and thus only considered the four nearest pixels. Yet here we show that longer range interpolations, and in particular a Gaussian interpolation, allow improving performance on ImageNet1k classification on two state-of-the-art convolutional architectures (ConvNeXt and Conv\-Former), without increasing the number of parameters. The method code is based on PyTorch and is available at https://github.com/K-H-Ismail/Dilated-Convolution-with-Learnable-Spacings-PyTorch.

Cite

Text

Khalfaoui-Hassani et al. "Dilated Convolution with Learnable Spacings: Beyond Bilinear Interpolation." ICML 2023 Workshops: Differentiable_Almost_Everything, 2023.

Markdown

[Khalfaoui-Hassani et al. "Dilated Convolution with Learnable Spacings: Beyond Bilinear Interpolation." ICML 2023 Workshops: Differentiable_Almost_Everything, 2023.](https://mlanthology.org/icmlw/2023/khalfaouihassani2023icmlw-dilated/)

BibTeX

@inproceedings{khalfaouihassani2023icmlw-dilated,
  title     = {{Dilated Convolution with Learnable Spacings: Beyond Bilinear Interpolation}},
  author    = {Khalfaoui-Hassani, Ismail and Pellegrini, Thomas and Masquelier, Timothée},
  booktitle = {ICML 2023 Workshops: Differentiable_Almost_Everything},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/khalfaouihassani2023icmlw-dilated/}
}