RepVGG: Making VGG-Style ConvNets Great Again

Abstract

We present a simple but powerful architecture of convolutional neural network, which has a VGG-like inference-time body composed of nothing but a stack of 3x3 convolution and ReLU, while the training-time model has a multi-branch topology. Such decoupling of the training-time and inference-time architecture is realized by a structural re-parameterization technique so that the model is named RepVGG. On ImageNet, RepVGG reaches over 80% top-1 accuracy, which is the first time for a plain model, to the best of our knowledge. On NVIDIA 1080Ti GPU, RepVGG models run 83% faster than ResNet-50 or 101% faster than ResNet-101 with higher accuracy and show favorable accuracy-speed trade-off compared to the state-of-the-art models like EfficientNet and RegNet. The code and trained models are available at https://github.com/megvii-model/RepVGG.

Cite

Text

Ding et al. "RepVGG: Making VGG-Style ConvNets Great Again." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.01352

Markdown

[Ding et al. "RepVGG: Making VGG-Style ConvNets Great Again." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/ding2021cvpr-repvgg/) doi:10.1109/CVPR46437.2021.01352

BibTeX

@inproceedings{ding2021cvpr-repvgg,
  title     = {{RepVGG: Making VGG-Style ConvNets Great Again}},
  author    = {Ding, Xiaohan and Zhang, Xiangyu and Ma, Ningning and Han, Jungong and Ding, Guiguang and Sun, Jian},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {13733-13742},
  doi       = {10.1109/CVPR46437.2021.01352},
  url       = {https://mlanthology.org/cvpr/2021/ding2021cvpr-repvgg/}
}