Accelerating Reinforcement Learning by Composing Solutions of Automatically Identified Subtasks

Abstract

This paper discusses a system that accelerates reinforcement learning by using transfer from related tasks. Without such transfer, even if two tasks are very similar at some abstract level, an extensive re-learning effort is required. The system achieves much of its power by transferring parts of previously learned solutions rather than a single complete solution. The system exploits strong features in the multi-dimensional function produced by reinforcement learning in solving a particular task. These features are stable and easy to recognize early in the learning process. They generate a partitioning of the state space and thus the function. The partition is represented as a graph. This is used to index and compose functions stored in a case base to form a close approximation to the solution of the new task. Experiments demonstrate that function composition often produces more than an order of magnitude increase in learning rate compared to a basic reinforcement learning algorithm.

Cite

Text

Drummond. "Accelerating Reinforcement Learning by Composing Solutions of Automatically Identified Subtasks." Journal of Artificial Intelligence Research, 2002. doi:10.1613/JAIR.904

Markdown

[Drummond. "Accelerating Reinforcement Learning by Composing Solutions of Automatically Identified Subtasks." Journal of Artificial Intelligence Research, 2002.](https://mlanthology.org/jair/2002/drummond2002jair-accelerating/) doi:10.1613/JAIR.904

BibTeX

@article{drummond2002jair-accelerating,
  title     = {{Accelerating Reinforcement Learning by Composing Solutions of Automatically Identified Subtasks}},
  author    = {Drummond, Chris},
  journal   = {Journal of Artificial Intelligence Research},
  year      = {2002},
  pages     = {59-104},
  doi       = {10.1613/JAIR.904},
  volume    = {16},
  url       = {https://mlanthology.org/jair/2002/drummond2002jair-accelerating/}
}