PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice

Abstract

Common simplifying assumptions often cause standard reinforcement learning (RL) methods to fail on complex, open-ended environments. Creating a new wrapper for each environment and learning library can help alleviate these limitations, but building them is labor-intensive and error-prone. This practical tooling gap restricts the applicability of RL as a whole. To address this challenge, PufferLib transforms complex environments into a broadly compatible, vectorized format that eliminates the need for bespoke conversion layers and enables rigorous cross-environment testing. PufferLib does this without deviating from standard reinforcement learning APIs, significantly reducing the technical overhead. We release PufferLib's complete source code under the MIT license, a pip module, a containerized setup, comprehensive documentation, and example integrations. We also maintain a community Discord channel to facilitate support and discussion.

Cite

Text

Suarez. "PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice." NeurIPS 2023 Workshops: ALOE, 2023.

Markdown

[Suarez. "PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice." NeurIPS 2023 Workshops: ALOE, 2023.](https://mlanthology.org/neuripsw/2023/suarez2023neuripsw-pufferlib/)

BibTeX

@inproceedings{suarez2023neuripsw-pufferlib,
  title     = {{PufferLib: Making Reinforcement Learning Libraries and Environments Play Nice}},
  author    = {Suarez, Joseph},
  booktitle = {NeurIPS 2023 Workshops: ALOE},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/suarez2023neuripsw-pufferlib/}
}