A New Way to Introduce Knowledge into Reinforcement Learning
Abstract
We present in this paper a method to introduce a priori knowledge into reinforcement learning using temporally extended actions. The aim of our work is to reduce the learning time of the Q-learning algorithm. This introduction of initial knowledge is done by constraining the set of available actions in some states. But at the same time, we can formulate that if the agent is in some particular states (called exception states), we have to relax those constraints. We define a mechanism called the propagation mechanism to get out of blocked situations induced by the initial knowledge constraints. We give some formal properties of our method and test it on a complex grid-world task. On this task, we compare our method with Q-learning and show that the learning time is drastically reduced for a very simple initial knowledge which would not be sufficient, by itself, to solve the task without the definition of exception situations and the propagation mechanism.
Cite
Text
Garcia. "A New Way to Introduce Knowledge into Reinforcement Learning." European Conference on Machine Learning, 2003. doi:10.1007/978-3-540-39857-8_16Markdown
[Garcia. "A New Way to Introduce Knowledge into Reinforcement Learning." European Conference on Machine Learning, 2003.](https://mlanthology.org/ecmlpkdd/2003/garcia2003ecml-new/) doi:10.1007/978-3-540-39857-8_16BibTeX
@inproceedings{garcia2003ecml-new,
title = {{A New Way to Introduce Knowledge into Reinforcement Learning}},
author = {Garcia, Pascal},
booktitle = {European Conference on Machine Learning},
year = {2003},
pages = {157-168},
doi = {10.1007/978-3-540-39857-8_16},
url = {https://mlanthology.org/ecmlpkdd/2003/garcia2003ecml-new/}
}