Bridging Natural Language and Emergent Representation in Hierarchical Reinforcement Learning

Abstract

Hierarchical Reinforcement Learning (HRL) breaks down complex tasks into manageable subtasks, but faces challenges with efficiency and generalization in high-dimensional, open-ended environments. Human In The Loop approaches offer a potential solution to these limitations. In this work, we propose the integration of large language models (LLMs) with HRL, leveraging LLMs' natural language and reasoning capabilities and study how to bridge the gap between human instructions and HRL's learned abstract representations. By translating human demonstrations into actionable reinforcement learning signals, LLMs can improve task abstraction and planning within HRL. Our approach builds upon the Spatial-Temporal Abstraction via Reachability (STAR) algorithm, using a LLM to optimize the hierarchical planning process. Empirical results obtained on continuous control tasks illustrate the potential of LLMs to enhance HRL particularly in environments requiring spatial reasoning and hierarchical control.

Cite

Text

Ji et al. "Bridging Natural Language and Emergent Representation in Hierarchical Reinforcement Learning." NeurIPS 2024 Workshops: IMOL, 2024.

Markdown

[Ji et al. "Bridging Natural Language and Emergent Representation in Hierarchical Reinforcement Learning." NeurIPS 2024 Workshops: IMOL, 2024.](https://mlanthology.org/neuripsw/2024/ji2024neuripsw-bridging/)

BibTeX

@inproceedings{ji2024neuripsw-bridging,
  title     = {{Bridging Natural Language and Emergent Representation in Hierarchical Reinforcement Learning}},
  author    = {Ji, Zihe and Nguyen, Sao Mai and Zadem, Mehdi},
  booktitle = {NeurIPS 2024 Workshops: IMOL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/ji2024neuripsw-bridging/}
}