Real-Time Recurrent Learning Using Trace Units in Reinforcement Learning
Abstract
Recurrent Neural Networks (RNNs) are used to learn representations in partially observable environments. For agents that learn online and continually interact with the environment, it is desirable to train RNNs with real-time recurrent learning (RTRL); unfortunately, RTRL is prohibitively expensive for standard RNNs. A promising direction is to use linear recurrent architectures (LRUs), where dense recurrent weights are replaced with a complex-valued diagonal, making RTRL efficient. In this work, we build on these insights to provide a lightweight but effective approach for training RNNs in online RL. We introduce Recurrent Trace Units (RTUs), a small modification on LRUs that we nonetheless find to have significant performance benefits over LRUs when trained with RTRL. We find RTUs significantly outperform GRUs and Transformers across several partially observable environments while using significantly less computation.
Cite
Text
Elelimy et al. "Real-Time Recurrent Learning Using Trace Units in Reinforcement Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-0541Markdown
[Elelimy et al. "Real-Time Recurrent Learning Using Trace Units in Reinforcement Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/elelimy2024neurips-realtime/) doi:10.52202/079017-0541BibTeX
@inproceedings{elelimy2024neurips-realtime,
title = {{Real-Time Recurrent Learning Using Trace Units in Reinforcement Learning}},
author = {Elelimy, Esraa and White, Adam and Bowling, Michael and White, Martha},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-0541},
url = {https://mlanthology.org/neurips/2024/elelimy2024neurips-realtime/}
}