GENNAPE: Towards Generalized Neural Architecture Performance Estimators
Abstract
Predicting neural architecture performance is a challenging task and is crucial to neural architecture design and search. Existing approaches either rely on neural performance predictors which are limited to modeling architectures in a predefined design space involving specific sets of operators and connection rules, and cannot generalize to unseen architectures, or resort to Zero-Cost Proxies which are not always accurate. In this paper, we propose GENNAPE, a Generalized Neural Architecture Performance Estimator, which is pretrained on open neural architecture benchmarks, and aims to generalize to completely unseen architectures through combined innovations in network representation, contrastive pretraining, and a fuzzy clustering-based predictor ensemble. Specifically, GENNAPE represents a given neural network as a Computation Graph (CG) of atomic operations which can model an arbitrary architecture. It first learns a graph encoder via Contrastive Learning to encourage network separation by topological features, and then trains multiple predictor heads, which are soft-aggregated according to the fuzzy membership of a neural network. Experiments show that GENNAPE pretrained on NAS-Bench-101 can achieve superior transferability to 5 different public neural network benchmarks, including NAS-Bench-201, NAS-Bench-301, MobileNet and ResNet families under no or minimum fine-tuning. We further introduce 3 challenging newly labelled neural network benchmarks: HiAML, Inception and Two-Path, which can concentrate in narrow accuracy ranges. Extensive experiments show that GENNAPE can correctly discern high-performance architectures in these families. Finally, when paired with a search algorithm, GENNAPE can find architectures that improve accuracy while reducing FLOPs on three families.
Cite
Text
Mills et al. "GENNAPE: Towards Generalized Neural Architecture Performance Estimators." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I8.26102Markdown
[Mills et al. "GENNAPE: Towards Generalized Neural Architecture Performance Estimators." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/mills2023aaai-gennape/) doi:10.1609/AAAI.V37I8.26102BibTeX
@inproceedings{mills2023aaai-gennape,
title = {{GENNAPE: Towards Generalized Neural Architecture Performance Estimators}},
author = {Mills, Keith G. and Han, Fred X. and Zhang, Jialin and Chudak, Fabian and Mamaghani, Ali Safari and Salameh, Mohammad and Lu, Wei and Jui, Shangling and Niu, Di},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {9190-9199},
doi = {10.1609/AAAI.V37I8.26102},
url = {https://mlanthology.org/aaai/2023/mills2023aaai-gennape/}
}