Fast TRAC: A Parameter-Free Optimizer for Lifelong Reinforcement Learning

Abstract

A key challenge in lifelong reinforcement learning (RL) is the loss of plasticity, where previous learning progress hinders an agent's adaptation to new tasks. While regularization and resetting can help, they require precise hyperparameter selection at the outset and environment-dependent adjustments. Building on the principled theory of online convex optimization, we present a parameter-free optimizer for lifelong RL, called TRAC, which requires no tuning or prior knowledge about the distribution shifts. Extensive experiments on Procgen, Atari, and Gym Control environments show that TRAC works surprisingly well—mitigating loss of plasticity and rapidly adapting to challenging distribution shifts—despite the underlying optimization problem being nonconvex and nonstationary.

Cite

Text

Muppidi et al. "Fast TRAC: A Parameter-Free Optimizer for Lifelong Reinforcement Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-1619

Markdown

[Muppidi et al. "Fast TRAC: A Parameter-Free Optimizer for Lifelong Reinforcement Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/muppidi2024neurips-fast/) doi:10.52202/079017-1619

BibTeX

@inproceedings{muppidi2024neurips-fast,
  title     = {{Fast TRAC: A Parameter-Free Optimizer for Lifelong Reinforcement Learning}},
  author    = {Muppidi, Aneesh and Zhang, Zhiyu and Yang, Heng},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-1619},
  url       = {https://mlanthology.org/neurips/2024/muppidi2024neurips-fast/}
}