Put CASH on Bandits: A Max K-Armed Problem for Automated Machine Learning

Abstract

The Combined Algorithm Selection and Hyperparameter optimization (CASH) is a challenging resource allocation problem in the field of AutoML. We propose MaxUCB, a max $k$-armed bandit method to trade off exploring different model classes and conducting hyperparameter optimization. MaxUCB is specifically designed for the light-tailed and bounded reward distributions arising in this setting and, thus, provides an efficient alternative compared to classic max $k$-armed bandit methods assuming heavy-tailed reward distributions. We theoretically and empirically evaluate our method on four standard AutoML benchmarks, demonstrating superior performance over prior approaches. We make our code and data available at https://github.com/amirbalef/CASH_with_Bandits

Cite

Text

Balef et al. "Put CASH on Bandits: A Max K-Armed Problem for Automated Machine Learning." Advances in Neural Information Processing Systems, 2025.

Markdown

[Balef et al. "Put CASH on Bandits: A Max K-Armed Problem for Automated Machine Learning." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/balef2025neurips-put/)

BibTeX

@inproceedings{balef2025neurips-put,
  title     = {{Put CASH on Bandits: A Max K-Armed Problem for Automated Machine Learning}},
  author    = {Balef, Amir Rezaei and Vernade, Claire and Eggensperger, Katharina},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/balef2025neurips-put/}
}