Automata Conditioned Reinforcement Learning with Experience Replay
Abstract
We explore the problem of goal-conditioned reinforcement learning (RL) where goals are represented using deterministic finite state automata (DFAs). Due to the sparse and binary nature of automata-based goals, we hypothesize that experience replay can help an RL agent learn more quickly and consistently in this setting. To enable the use of experience replay, we introduce a novel end-to-end neural architecture, including a graph neural network (GNN) to encode the DFA goal before passing it to a feed-forward policy network. Experimental results in a gridworld domain demonstrate the efficacy of the model architecture and highlight the significant role of experience replay in enhancing the learning speed and reducing the variance of RL agents for DFA tasks.
Cite
Text
Yalcinkaya et al. "Automata Conditioned Reinforcement Learning with Experience Replay." NeurIPS 2023 Workshops: GCRL, 2023.Markdown
[Yalcinkaya et al. "Automata Conditioned Reinforcement Learning with Experience Replay." NeurIPS 2023 Workshops: GCRL, 2023.](https://mlanthology.org/neuripsw/2023/yalcinkaya2023neuripsw-automata/)BibTeX
@inproceedings{yalcinkaya2023neuripsw-automata,
title = {{Automata Conditioned Reinforcement Learning with Experience Replay}},
author = {Yalcinkaya, Beyazit and Lauffer, Niklas and Vazquez-Chanlatte, Marcell and Seshia, Sanjit},
booktitle = {NeurIPS 2023 Workshops: GCRL},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/yalcinkaya2023neuripsw-automata/}
}