Hierarchical Skills for Efficient Exploration
Abstract
In reinforcement learning, pre-trained low-level skills have the potential to greatly facilitate exploration. However, prior knowledge of the downstream task is required to strike the right balance between generality (fine-grained control) and specificity (faster learning) in skill design. In previous work on continuous control, the sensitivity of methods to this trade-off has not been addressed explicitly, as locomotion provides a suitable prior for navigation tasks, which have been of foremost interest. In this work, we analyze this trade-off for low-level policy pre-training with a new benchmark suite of diverse, sparse-reward tasks for bipedal robots. We alleviate the need for prior knowledge by proposing a hierarchical skill learning framework that acquires skills of varying complexity in an unsupervised manner. For utilization on downstream tasks, we present a three-layered hierarchical learning algorithm to automatically trade off between general and specific skills as required by the respective task. In our experiments, we show that our approach performs this trade-off effectively and achieves better results than current state-of-the-art methods for end-to-end hierarchical reinforcement learning and unsupervised skill discovery.
Cite
Text
Gehring et al. "Hierarchical Skills for Efficient Exploration." Neural Information Processing Systems, 2021.Markdown
[Gehring et al. "Hierarchical Skills for Efficient Exploration." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/gehring2021neurips-hierarchical/)BibTeX
@inproceedings{gehring2021neurips-hierarchical,
title = {{Hierarchical Skills for Efficient Exploration}},
author = {Gehring, Jonas and Synnaeve, Gabriel and Krause, Andreas and Usunier, Nicolas},
booktitle = {Neural Information Processing Systems},
year = {2021},
url = {https://mlanthology.org/neurips/2021/gehring2021neurips-hierarchical/}
}