Reinforcement Learning with a Hierarchy of Abstract Models
Abstract
Reinforcement learning (RL) algorithms have traditionally been thought of as trial and error learning methods that use actual control experience to incrementally improve a control policy. Sutton's DYNA architecture demonstrated that RL algorithms can work as well using simulated experience from an environment model, and that the resulting computation was similar to doing one-step lookahead planning. Inspired by the literature on hierarchical planning, I propose learning a hierarchy of models of the environment that abstract temporal detail as a means of improving the scalability of RL algorithms. I present H-DYNA (Hierarchical DYNA), an extension to Sutton's DYNA architecture that is able to learn such a hierarchy of abstract models. H-DYNA differs from hierarchical planners in two ways: first, the abstract models are learned using experience gained while...
Cite
Text
Singh. "Reinforcement Learning with a Hierarchy of Abstract Models." AAAI Conference on Artificial Intelligence, 1992.Markdown
[Singh. "Reinforcement Learning with a Hierarchy of Abstract Models." AAAI Conference on Artificial Intelligence, 1992.](https://mlanthology.org/aaai/1992/singh1992aaai-reinforcement/)BibTeX
@inproceedings{singh1992aaai-reinforcement,
title = {{Reinforcement Learning with a Hierarchy of Abstract Models}},
author = {Singh, Satinder P.},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {1992},
pages = {202-207},
url = {https://mlanthology.org/aaai/1992/singh1992aaai-reinforcement/}
}