Kernel Transformer Networks for Compact Spherical Convolution

Abstract

Ideally, 360deg imagery could inherit the convolutional neural networks (CNNs) already trained with great success on perspective projection images. However, existing methods to transfer CNNs from perspective to spherical images introduce significant computational costs and/or degradations in accuracy. We present the Kernel Transformer Network (KTN) to efficiently transfer convolution kernels from perspective images to the equirectangular projection of 360deg images. Given a source CNN for perspective images as input, the KTN produces a function parameterized by a polar angle and kernel as output. Given a novel 360deg image, that function in turn can compute convolutions for arbitrary layers and kernels as would the source CNN on the corresponding tangent plane projections. Distinct from all existing methods, KTNs allow model transfer: the same model can be applied to different source CNNs with the same base architecture. KTNs successfully preserve the source CNN's accuracy, while offering transferability, scalability to typical image resolutions, and, in many cases, a substantially lower memory footprint.

Cite

Text

Su and Grauman. "Kernel Transformer Networks for Compact Spherical Convolution." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.

Markdown

[Su and Grauman. "Kernel Transformer Networks for Compact Spherical Convolution." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/su2019cvprw-kernel/)

BibTeX

@inproceedings{su2019cvprw-kernel,
  title     = {{Kernel Transformer Networks for Compact Spherical Convolution}},
  author    = {Su, Yu-Chuan and Grauman, Kristen},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2019},
  pages     = {11-15},
  url       = {https://mlanthology.org/cvprw/2019/su2019cvprw-kernel/}
}