Tight and Efficient Upper Bound on Spectral Norm of Convolutional Layers

Abstract

Controlling the spectral norm of the Jacobian matrix, which is related to the convolution operation, has been shown to improve generalization, training stability and robustness in CNNs. Existing methods for computing the norm either tend to overestimate it or their performance may deteriorate quickly with increasing the input and kernel sizes. In this paper, we demonstrate that the tensor version of the spectral norm of a four-dimensional convolution kernel, up to a constant factor, serves as an upper bound for the spectral norm of the Jacobian matrix associated with the convolution operation. This new upper bound is independent of the input image resolution, differentiable and can be efficiently calculated during training. Through experiments, we demonstrate how this new bound can be used to improve the performance of convolutional architectures.

Cite

Text

Grishina et al. "Tight and Efficient Upper Bound on Spectral Norm of Convolutional Layers." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-73024-5_2

Markdown

[Grishina et al. "Tight and Efficient Upper Bound on Spectral Norm of Convolutional Layers." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/grishina2024eccv-tight/) doi:10.1007/978-3-031-73024-5_2

BibTeX

@inproceedings{grishina2024eccv-tight,
  title     = {{Tight and Efficient Upper Bound on Spectral Norm of Convolutional Layers}},
  author    = {Grishina, Ekaterina and Gorbunov, Mikhail and Rakhuba, Maxim},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2024},
  doi       = {10.1007/978-3-031-73024-5_2},
  url       = {https://mlanthology.org/eccv/2024/grishina2024eccv-tight/}
}