Transfer of Samples in Batch Reinforcement Learning
Abstract
The main objective of transfer learning is to reduce the complexity of learning the solution of a target task by effectively reusing the knowledge retained from solving one or more source tasks. In this paper, we introduce a novel algorithm that transfers samples (i.e., experience tuples ) from source to target tasks. Under the assumption that tasks defined on the same environment often have similar transition models and reward functions, we propose a method to select samples from the source tasks that are mostly similar to the target task, and, then, to use them as input for batch reinforcement learning algorithms. As a result, the number of samples that the agent needs to collect from the target task to learn its solution is reduced. We empirically show that, following the proposed approach, the transfer of samples is effective in reducing the learning complexity, even when the source tasks are significantly different from the target task.
Cite
Text
Lazaric et al. "Transfer of Samples in Batch Reinforcement Learning." International Conference on Machine Learning, 2008. doi:10.1145/1390156.1390225Markdown
[Lazaric et al. "Transfer of Samples in Batch Reinforcement Learning." International Conference on Machine Learning, 2008.](https://mlanthology.org/icml/2008/lazaric2008icml-transfer/) doi:10.1145/1390156.1390225BibTeX
@inproceedings{lazaric2008icml-transfer,
title = {{Transfer of Samples in Batch Reinforcement Learning}},
author = {Lazaric, Alessandro and Restelli, Marcello and Bonarini, Andrea},
booktitle = {International Conference on Machine Learning},
year = {2008},
pages = {544-551},
doi = {10.1145/1390156.1390225},
url = {https://mlanthology.org/icml/2008/lazaric2008icml-transfer/}
}