Designing Concise ConvNets with Columnar Stages
Abstract
In the era of vision Transformers, the recent success of VanillaNet shows the huge potential of simple and concise convolutional neural networks (ConvNets). Where such models mainly focus on runtime, it is also crucial to simultaneously focus on other aspects, e.g., FLOPs, parameters, etc, to strengthen their utility further. To this end, we introduce a refreshing ConvNet macro design called Columnar Stage Network (CoSNet). CoSNet has a systematically developed simple and concise structure, smaller depth, low parameter count, low FLOPs, and attention- less operations, well suited for resource-constrained deployment. The key novelty of CoSNet is deploying parallel convolutions with fewer kernels fed by input replication, using columnar stacking of these convolutions, and minimizing the use of 1×1 convolution layers. Our comprehensive evaluations show that CoSNet rivals many renowned ConvNets and Transformer designs under resource-constrained scenarios. Pretrained models shall be open-sourced.
Cite
Text
Kumar and Park. "Designing Concise ConvNets with Columnar Stages." International Conference on Learning Representations, 2025.Markdown
[Kumar and Park. "Designing Concise ConvNets with Columnar Stages." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/kumar2025iclr-designing/)BibTeX
@inproceedings{kumar2025iclr-designing,
title = {{Designing Concise ConvNets with Columnar Stages}},
author = {Kumar, Ashish and Park, Jaesik},
booktitle = {International Conference on Learning Representations},
year = {2025},
url = {https://mlanthology.org/iclr/2025/kumar2025iclr-designing/}
}