Is Feedback All You Need? Leveraging Natural Language Feedback in Goal-Conditioned RL

Abstract

Despite numerous successes, the field of reinforcement learning (RL) remains far from matching the impressive generalisation power of human behaviour learning. One way to help bridge this gap may be to provide RL agents with richer, more human-like feedback expressed in natural language. First, we extend BabyAI to automatically generate language feedback from the environment dynamics and goal condition success. Then, we modify the Decision Transformer architecture to take advantage of this additional signal. We find that training with language feedback either in place of or in addition to the return-to-go or goal descriptions improves agents’ generalisation performance, and that agents can benefit from feedback even when this is only available during training, but not at inference.

Cite

Text

McCallum et al. "Is Feedback All You Need? Leveraging Natural Language Feedback in Goal-Conditioned RL." NeurIPS 2023 Workshops: GCRL, 2023.

Markdown

[McCallum et al. "Is Feedback All You Need? Leveraging Natural Language Feedback in Goal-Conditioned RL." NeurIPS 2023 Workshops: GCRL, 2023.](https://mlanthology.org/neuripsw/2023/mccallum2023neuripsw-feedback/)

BibTeX

@inproceedings{mccallum2023neuripsw-feedback,
  title     = {{Is Feedback All You Need? Leveraging Natural Language Feedback in Goal-Conditioned RL}},
  author    = {McCallum, Sabrina and Taylor-Davies, Max and Albrecht, Stefano and Suglia, Alessandro},
  booktitle = {NeurIPS 2023 Workshops: GCRL},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/mccallum2023neuripsw-feedback/}
}