Composing Functions to Speed up Reinforcement Learning in a Changing World
Abstract
This paper presents a system that transfers the results of prior learning to speed up reinforcement learning in a changing world. Often, even when the change to the world is relatively small an extensive relearning effort is required. The new system exploits strong features in the multi-dimensional function produced by reinforcement learning. The features generate a partitioning of the state space. The partition is represented as a graph. This is used to index and compose functions stored in a case base to form a close approximation to the solution of the new task. The experimental results investigate one important example of a changing world, a new goal position. In this situation, there is close to a two orders of magnitude increase in learning rate over using a basic reinforcement learning algorithm.
Cite
Text
Drummond. "Composing Functions to Speed up Reinforcement Learning in a Changing World." European Conference on Machine Learning, 1998. doi:10.1007/BFB0026708Markdown
[Drummond. "Composing Functions to Speed up Reinforcement Learning in a Changing World." European Conference on Machine Learning, 1998.](https://mlanthology.org/ecmlpkdd/1998/drummond1998ecml-composing/) doi:10.1007/BFB0026708BibTeX
@inproceedings{drummond1998ecml-composing,
title = {{Composing Functions to Speed up Reinforcement Learning in a Changing World}},
author = {Drummond, Chris},
booktitle = {European Conference on Machine Learning},
year = {1998},
pages = {370-381},
doi = {10.1007/BFB0026708},
url = {https://mlanthology.org/ecmlpkdd/1998/drummond1998ecml-composing/}
}