SMASH: One-Shot Model Architecture Search Through HyperNetworks

Abstract

Designing architectures for deep neural networks requires expert knowledge and substantial computation time. We propose a technique to accelerate architecture selection by learning an auxiliary HyperNet that generates the weights of a main model conditioned on that model's architecture. By comparing the relative validation performance of networks with HyperNet-generated weights, we can effectively search over a wide range of architectures at the cost of a single training run. To facilitate this search, we develop a flexible mechanism based on memory read-writes that allows us to define a wide range of network connectivity patterns, with ResNet, DenseNet, and FractalNet blocks as special cases. We validate our method (SMASH) on CIFAR-10 and CIFAR-100, STL-10, ModelNet10, and Imagenet32x32, achieving competitive performance with similarly-sized hand-designed networks.

Cite

Text

Brock et al. "SMASH: One-Shot Model Architecture Search Through HyperNetworks." International Conference on Learning Representations, 2018.

Markdown

[Brock et al. "SMASH: One-Shot Model Architecture Search Through HyperNetworks." International Conference on Learning Representations, 2018.](https://mlanthology.org/iclr/2018/brock2018iclr-smash/)

BibTeX

@inproceedings{brock2018iclr-smash,
  title     = {{SMASH: One-Shot Model Architecture Search Through HyperNetworks}},
  author    = {Brock, Andrew and Lim, Theo and Ritchie, J.M. and Weston, Nick},
  booktitle = {International Conference on Learning Representations},
  year      = {2018},
  url       = {https://mlanthology.org/iclr/2018/brock2018iclr-smash/}
}