Supervising the Multi-Fidelity Race of Hyperparameter Configurations

Abstract

Multi-fidelity (gray-box) hyperparameter optimization techniques (HPO) have recently emerged as a promising direction for tuning Deep Learning methods. However, existing methods suffer from a sub-optimal allocation of the HPO budget to the hyperparameter configurations. In this work, we introduce DyHPO, a Bayesian Optimization method that learns to decide which hyperparameter configuration to train further in a dynamic race among all feasible configurations. We propose a new deep kernel for Gaussian Processes that embeds the learning curve dynamics, and an acquisition function that incorporates multi-budget information. We demonstrate the significant superiority of DyHPO against state-of-the-art hyperparameter optimization methods through large-scale experiments comprising 50 datasets (Tabular, Image, NLP) and diverse architectures (MLP, CNN/NAS, RNN).

Cite

Text

Wistuba et al. "Supervising the Multi-Fidelity Race of Hyperparameter Configurations." Neural Information Processing Systems, 2022.

Markdown

[Wistuba et al. "Supervising the Multi-Fidelity Race of Hyperparameter Configurations." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/wistuba2022neurips-supervising/)

BibTeX

@inproceedings{wistuba2022neurips-supervising,
  title     = {{Supervising the Multi-Fidelity Race of Hyperparameter Configurations}},
  author    = {Wistuba, Martin and Kadra, Arlind and Grabocka, Josif},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/wistuba2022neurips-supervising/}
}