Any-Width Networks

Abstract

Despite remarkable improvements in speed and accuracy, convolutional neural networks (CNNs) still typically operate as monolithic entities at inference time. This poses a challenge for resource-constrained practical applications, where both computational budgets and performance needs can vary with the situation. To address these constraints, we propose the Any-Width Network (AWN), an adjustable-width CNN architecture and associated training routine that allow for fine-grained control over speed and accuracy during inference. Our key innovation is the use of lower-triangular weight matrices which explicitly address width-varying batch statistics while being naturally suited for multi-width operations. We also show that this design facilitates an efficient training routine based on random width sampling. We empirically demonstrate that our proposed AWNs compare favorably to existing methods while providing maximally granular control during inference.

Cite

Text

Vu et al. "Any-Width Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020. doi:10.1109/CVPRW50498.2020.00360

Markdown

[Vu et al. "Any-Width Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020.](https://mlanthology.org/cvprw/2020/vu2020cvprw-anywidth/) doi:10.1109/CVPRW50498.2020.00360

BibTeX

@inproceedings{vu2020cvprw-anywidth,
  title     = {{Any-Width Networks}},
  author    = {Vu, Thanh and Eder, Marc and Price, true and Frahm, Jan-Michael},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2020},
  pages     = {3018-3026},
  doi       = {10.1109/CVPRW50498.2020.00360},
  url       = {https://mlanthology.org/cvprw/2020/vu2020cvprw-anywidth/}
}