Learning State Reachability as a Graph in Translation Invariant Goal-Based Reinforcement Learning Tasks
Abstract
Deep Reinforcement Learning proved efficient at learning universal control policies when the goal state is close enough to the starting state, or when the value function features few discontinuities. But reaching goals that require long action sequences in complex environments remains difficult. Drawing inspiration from the cognitive process which reuses learned atomic skills in a global planning procedure, we propose an algorithm which encodes reachability between abstract goals as a graph, and produces plans in this goal space. Transitions between goals rely on the exploitation of a learned policy which enjoys a property we call \emph{translation invariant local optimality}, which encodes the intuition that goal-reaching skills can be reused throughout the state space. Overall, our contribution permits solving large and difficult navigation tasks, outperforming related methods from the literature.
Cite
Text
Bonnavaud et al. "Learning State Reachability as a Graph in Translation Invariant Goal-Based Reinforcement Learning Tasks." Transactions on Machine Learning Research, 2024.Markdown
[Bonnavaud et al. "Learning State Reachability as a Graph in Translation Invariant Goal-Based Reinforcement Learning Tasks." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/bonnavaud2024tmlr-learning/)BibTeX
@article{bonnavaud2024tmlr-learning,
title = {{Learning State Reachability as a Graph in Translation Invariant Goal-Based Reinforcement Learning Tasks}},
author = {Bonnavaud, Hedwin and Albore, Alexandre and Rachelson, Emmanuel},
journal = {Transactions on Machine Learning Research},
year = {2024},
url = {https://mlanthology.org/tmlr/2024/bonnavaud2024tmlr-learning/}
}