Hierarchical Policy Gradient Algorithms

Abstract

Hierarchical reinforcement learning is a general framework which attempts to accelerate policy learning in large domains. On the other hand, policy gradient reinforcement learning (PGRL) methods have received recent attention as a means to solve problems with continuous state spaces. However, they suffer from slow convergence. In this paper, we combine these two approaches and propose a family of hierarchical policy gradient algorithms for problems with continuous state and/or action spaces. We also introduce a class of hierarchical hybrid algorithms, in which a group of subtasks, usually at the higher-levels of the hierarchy, are formulated as value function-based RL (VFRL) problems and the others as PGRL problems. We demonstrate the performance of our proposed algorithms using a simple taxi-fuel problem and a complex continuous state and action ship steering domain. ICML Proceedings of the Twentieth International Conference on Machine Learning

Cite

Text

Ghavamzadeh and Mahadevan. "Hierarchical Policy Gradient Algorithms." International Conference on Machine Learning, 2003.

Markdown

[Ghavamzadeh and Mahadevan. "Hierarchical Policy Gradient Algorithms." International Conference on Machine Learning, 2003.](https://mlanthology.org/icml/2003/ghavamzadeh2003icml-hierarchical/)

BibTeX

@inproceedings{ghavamzadeh2003icml-hierarchical,
  title     = {{Hierarchical Policy Gradient Algorithms}},
  author    = {Ghavamzadeh, Mohammad and Mahadevan, Sridhar},
  booktitle = {International Conference on Machine Learning},
  year      = {2003},
  pages     = {226-233},
  url       = {https://mlanthology.org/icml/2003/ghavamzadeh2003icml-hierarchical/}
}