ISBNet: Instance-Aware Selective Branching Networks

Abstract

Recent years have witnessed growing interests in designing efficient neural networks and neural architecture search (NAS). Although remarkable efficiency and accuracy have been achieved, existing expert designed and NAS models neglect the fact that input instances are of varying complexity and thus different amounts of computation are required. Inference with a fixed model that processes all instances through the same transformations would incur computational resources unnecessarily. Customizing the model capacity in an instance-aware manner is required to alleviate such a problem. In this paper, we propose a novel Instance-aware Selective Branching Network-ISBNet to support efficient instance-level inference by selectively bypassing transformation branches of insignificant importance weight. These weights are dynamically determined by a lightweight hypernetwork SelectionNet and recalibrated by gumbel-softmax for sparse branch selection. Extensive experiments show that ISBNet achieves extremely efficient inference in terms of parameter size and FLOPs comparing to existing networks. For example, ISBNet takes only 8.70% parameters and 31.01% FLOPs of the efficient network MobileNetV2 with comparable accuracy on CIFAR-10.

Cite

Text

Cai et al. "ISBNet: Instance-Aware Selective Branching Networks." International Conference on Learning Representations, 2020.

Markdown

[Cai et al. "ISBNet: Instance-Aware Selective Branching Networks." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/cai2020iclr-isbnet/)

BibTeX

@inproceedings{cai2020iclr-isbnet,
  title     = {{ISBNet: Instance-Aware Selective Branching Networks}},
  author    = {Cai, Shaofeng and Shu, Yao and Wang, Wei and Chen, Gang and Ooi, Beng Chin},
  booktitle = {International Conference on Learning Representations},
  year      = {2020},
  url       = {https://mlanthology.org/iclr/2020/cai2020iclr-isbnet/}
}