Reinforcement Learning with a Hierarchy of Abstract Models

Abstract

Reinforcement learning (RL) algorithms have traditionally been thought of as trial and error learning methods that use actual control experience to incrementally improve a control policy. Sutton's DYNA architecture demonstrated that RL algorithms can work as well using simulated experience from an environment model, and that the resulting computation was similar to doing one-step lookahead planning. Inspired by the literature on hierarchical planning, I propose learning a hierarchy of models of the environment that abstract temporal detail as a means of improving the scalability of RL algorithms. I present H-DYNA (Hierarchical DYNA), an extension to Sutton's DYNA architecture that is able to learn such a hierarchy of abstract models. H-DYNA differs from hierarchical planners in two ways: first, the abstract models are learned using experience gained while...

Cite

Text

Singh. "Reinforcement Learning with a Hierarchy of Abstract Models." AAAI Conference on Artificial Intelligence, 1992.

Markdown

[Singh. "Reinforcement Learning with a Hierarchy of Abstract Models." AAAI Conference on Artificial Intelligence, 1992.](https://mlanthology.org/aaai/1992/singh1992aaai-reinforcement/)

BibTeX

@inproceedings{singh1992aaai-reinforcement,
  title     = {{Reinforcement Learning with a Hierarchy of Abstract Models}},
  author    = {Singh, Satinder P.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {1992},
  pages     = {202-207},
  url       = {https://mlanthology.org/aaai/1992/singh1992aaai-reinforcement/}
}