A Method for Evaluating Hyperparameter Sensitivity in Reinforcement Learning

Abstract

The performance of modern reinforcement learning algorithms critically relieson tuning ever increasing numbers of hyperparameters. Often, small changes ina hyperparameter can lead to drastic changes in performance, and different environments require very different hyperparameter settings to achieve state-of-the-artperformance reported in the literature. We currently lack a scalable and widelyaccepted approach to characterizing these complex interactions. This work proposes a new empirical methodology for studying, comparing, and quantifying thesensitivity of an algorithm’s performance to hyperparameter tuning for a given setof environments. We then demonstrate the utility of this methodology by assessingthe hyperparameter sensitivity of several commonly used normalization variants ofPPO. The results suggest that several algorithmic performance improvements may,in fact, be a result of an increased reliance on hyperparameter tuning.

Cite

Text

Adkins et al. "A Method for Evaluating Hyperparameter Sensitivity in Reinforcement Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-3964

Markdown

[Adkins et al. "A Method for Evaluating Hyperparameter Sensitivity in Reinforcement Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/adkins2024neurips-method/) doi:10.52202/079017-3964

BibTeX

@inproceedings{adkins2024neurips-method,
  title     = {{A Method for Evaluating Hyperparameter Sensitivity in Reinforcement Learning}},
  author    = {Adkins, Jacob and Bowling, Michael and White, Adam},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-3964},
  url       = {https://mlanthology.org/neurips/2024/adkins2024neurips-method/}
}