Federated Ensemble-Directed Offline Reinforcement Learning
Abstract
We consider the problem of federated offline reinforcement learning (RL), a scenario under which distributed learning agents must collaboratively learn a high-quality control policy only using small pre-collected datasets generated according to different unknown behavior policies. Na\"ively combining a standard offline RL approach with a standard federated learning approach to solve this problem can lead to poorly performing policies. In response, we develop the Federated Ensemble-Directed Offline Reinforcement Learning Algorithm (FEDORA), which distills the collective wisdom of the clients using an ensemble learning approach. We develop the FEDORA codebase to utilize distributed compute resources on a federated learning platform. We show that FEDORA significantly outperforms other approaches, including offline RL over the combined data pool, in various complex continuous control environments and real-world datasets. Finally, we demonstrate the performance of FEDORA in the real-world on a mobile robot. We provide our code and a video of our experiments at \url{https://github.com/DesikRengarajan/FEDORA}.
Cite
Text
Rengarajan et al. "Federated Ensemble-Directed Offline Reinforcement Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-0199Markdown
[Rengarajan et al. "Federated Ensemble-Directed Offline Reinforcement Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/rengarajan2024neurips-federated/) doi:10.52202/079017-0199BibTeX
@inproceedings{rengarajan2024neurips-federated,
title = {{Federated Ensemble-Directed Offline Reinforcement Learning}},
author = {Rengarajan, Desik and Ragothaman, Nitin and Kalathil, Dileep and Shakkottai, Srinivas},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-0199},
url = {https://mlanthology.org/neurips/2024/rengarajan2024neurips-federated/}
}