Sub-Policy Adaptation for Hierarchical Reinforcement Learning

Abstract

Hierarchical reinforcement learning is a promising approach to tackle long-horizon decision-making problems with sparse rewards. Unfortunately, most methods still decouple the lower-level skill acquisition process and the training of a higher level that controls the skills in a new task. Leaving the skills fixed can lead to significant sub-optimality in the transfer setting. In this work, we propose a novel algorithm to discover a set of skills, and continuously adapt them along with the higher level even when training on a new task. Our main contributions are two-fold. First, we derive a new hierarchical policy gradient with an unbiased latent-dependent baseline, and we introduce Hierarchical Proximal Policy Optimization (HiPPO), an on-policy method to efficiently train all levels of the hierarchy jointly. Second, we propose a method of training time-abstractions that improves the robustness of the obtained skills to environment changes. Code and videos are available at sites.google.com/view/hippo-rl.

Cite

Text

Li et al. "Sub-Policy Adaptation for Hierarchical Reinforcement Learning." International Conference on Learning Representations, 2020.

Markdown

[Li et al. "Sub-Policy Adaptation for Hierarchical Reinforcement Learning." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/li2020iclr-subpolicy/)

BibTeX

@inproceedings{li2020iclr-subpolicy,
  title     = {{Sub-Policy Adaptation for Hierarchical Reinforcement Learning}},
  author    = {Li, Alexander C. and Florensa, Carlos and Clavera, Ignasi and Abbeel, Pieter},
  booktitle = {International Conference on Learning Representations},
  year      = {2020},
  url       = {https://mlanthology.org/iclr/2020/li2020iclr-subpolicy/}
}