PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice
Abstract
Common simplifying assumptions often cause standard reinforcement learning (RL) methods to fail on complex, open-ended environments. Creating a new wrapper for each environment and learning library can help alleviate these limitations, but building them is labor-intensive and error-prone. This practical tooling gap restricts the applicability of RL as a whole. To address this challenge, PufferLib transforms complex environments into a broadly compatible, vectorized format that eliminates the need for bespoke conversion layers and enables rigorous cross-environment testing. PufferLib does this without deviating from standard reinforcement learning APIs, significantly reducing the technical overhead. We release PufferLib's complete source code under the MIT license, a pip module, a containerized setup, comprehensive documentation, and example integrations. We also maintain a community Discord channel to facilitate support and discussion.
Cite
Text
Suarez. "PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice." NeurIPS 2023 Workshops: ALOE, 2023.Markdown
[Suarez. "PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice." NeurIPS 2023 Workshops: ALOE, 2023.](https://mlanthology.org/neuripsw/2023/suarez2023neuripsw-pufferlib/)BibTeX
@inproceedings{suarez2023neuripsw-pufferlib,
title = {{PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice}},
author = {Suarez, Joseph},
booktitle = {NeurIPS 2023 Workshops: ALOE},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/suarez2023neuripsw-pufferlib/}
}