Can Learned Optimization Make Reinforcement Learning Less Difficult?

Abstract

While reinforcement learning (RL) holds great potential for decision making in the real world, it suffers from a number of unique difficulties which often need specific consideration. In particular: it is highly non-stationary; suffers from high degrees of plasticity loss; and requires exploration to prevent premature convergence to local optima and maximize return. In this paper, we consider whether learned optimization can help overcome these problems. Our method, Learned **O**ptimization for **P**lasticity, **E**xploration and **N**on-stationarity (*OPEN*), meta-learns an update rule whose input features and output structure are informed by previously proposed solutions to these difficulties. We show that our parameterization is flexible enough to enable meta-learning in diverse learning contexts, including the ability to use stochasticity for exploration. Our experiments demonstrate that when meta-trained on single and small sets of environments, *OPEN* outperforms or equals traditionally used optimizers. Furthermore, *OPEN* shows strong generalization across a *distribution of environments* and a range of agent architectures.

Cite

Text

Goldie et al. "Can Learned Optimization Make Reinforcement Learning Less Difficult?." ICML 2024 Workshops: AutoRL, 2024.

Markdown

[Goldie et al. "Can Learned Optimization Make Reinforcement Learning Less Difficult?." ICML 2024 Workshops: AutoRL, 2024.](https://mlanthology.org/icmlw/2024/goldie2024icmlw-learned/)

BibTeX

@inproceedings{goldie2024icmlw-learned,
  title     = {{Can Learned Optimization Make Reinforcement Learning Less Difficult?}},
  author    = {Goldie, Alexander D. and Lu, Chris and Jackson, Matthew Thomas and Whiteson, Shimon and Foerster, Jakob Nicolaus},
  booktitle = {ICML 2024 Workshops: AutoRL},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/goldie2024icmlw-learned/}
}