ParameterNet: Parameters Are All You Need for Large-Scale Visual Pretraining of Mobile Networks

Abstract

The large-scale visual pretraining has significantly improve the performance of large vision models. However we observe the low FLOPs pitfall that the existing low-FLOPs models cannot benefit from large-scale pretraining. In this paper we introduce a novel design principle termed ParameterNet aimed at augmenting the number of parameters in large-scale visual pretraining models while minimizing the increase in FLOPs. We leverage dynamic convolutions to incorporate additional parameters into the networks with only a marginal rise in FLOPs. The ParameterNet approach allows low-FLOPs networks to take advantage of large-scale visual pretraining. Furthermore we extend the ParameterNet concept to the language domain to enhance inference results while preserving inference speed. Experiments on the large-scale ImageNet-22K have shown the superiority of our ParameterNet scheme. For example ParameterNet-600M can achieve higher accuracy than the widely-used Swin Transformer (81.6% vs. 80.9%) and has much lower FLOPs (0.6G vs. 4.5G). The code will be released at https://parameternet.github.io/.

Cite

Text

Han et al. "ParameterNet: Parameters Are All You Need for Large-Scale Visual Pretraining of Mobile Networks." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.01491

Markdown

[Han et al. "ParameterNet: Parameters Are All You Need for Large-Scale Visual Pretraining of Mobile Networks." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/han2024cvpr-parameternet/) doi:10.1109/CVPR52733.2024.01491

BibTeX

@inproceedings{han2024cvpr-parameternet,
  title     = {{ParameterNet: Parameters Are All You Need for Large-Scale Visual Pretraining of Mobile Networks}},
  author    = {Han, Kai and Wang, Yunhe and Guo, Jianyuan and Wu, Enhua},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {15751-15761},
  doi       = {10.1109/CVPR52733.2024.01491},
  url       = {https://mlanthology.org/cvpr/2024/han2024cvpr-parameternet/}
}