Conditional Meta-Reinforcement Learning with State Representation

Abstract

Reinforcement Learning (RL) has achieved remarkable success in diverse areas, yet its sample inefficiency—requiring extensive interactions to develop optimal policies—remains a challenge. Meta-Reinforcement Learning (Meta-RL) addresses this by leveraging previously acquired knowledge, often integrating contextual information into learning. This study delves into conditional Meta-RL, investigating how context influences learning efficiency. We introduce a novel theoretical framework for both unconditional and conditional Meta-RL scenarios, with a focus on approximating the value function using state representations in environments where the transition kernel is known. This framework lays the groundwork for understanding the advantages of conditional Meta-RL over unconditional Meta-RL approaches. Furthermore, we present a conditional Meta-RL algorithm that shown to offer more than 50 percent increase in the average mean than unconditional setting in MiniGrid environments.

Cite

Text

Sun et al. "Conditional Meta-Reinforcement Learning with State Representation." ICML 2024 Workshops: AutoRL, 2024.

Markdown

[Sun et al. "Conditional Meta-Reinforcement Learning with State Representation." ICML 2024 Workshops: AutoRL, 2024.](https://mlanthology.org/icmlw/2024/sun2024icmlw-conditional/)

BibTeX

@inproceedings{sun2024icmlw-conditional,
  title     = {{Conditional Meta-Reinforcement Learning with State Representation}},
  author    = {Sun, Yuxuan and Toni, Laura and Andreopoulos, Yiannis},
  booktitle = {ICML 2024 Workshops: AutoRL},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/sun2024icmlw-conditional/}
}