Batch Reinforcement Learning with State Importance
Abstract
We investigate the problem of using function approximation in reinforcement learning where the agent’s policy is represented as a classifier mapping states to actions. High classification accuracy is usually deemed to correlate with high policy quality. But this is not necessarily the case as increasing classification accuracy can actually decrease the policy’s quality. This phenomenon takes place when the learning process begins to focus on classifying less “important” states. In this paper, we introduce a measure of state’s decision-making importance that can be used to improve policy learning. As a result, the focused learning process is shown to converge faster to better policies.
Cite
Text
Li et al. "Batch Reinforcement Learning with State Importance." European Conference on Machine Learning, 2004. doi:10.1007/978-3-540-30115-8_53Markdown
[Li et al. "Batch Reinforcement Learning with State Importance." European Conference on Machine Learning, 2004.](https://mlanthology.org/ecmlpkdd/2004/li2004ecml-batch/) doi:10.1007/978-3-540-30115-8_53BibTeX
@inproceedings{li2004ecml-batch,
title = {{Batch Reinforcement Learning with State Importance}},
author = {Li, Lihong and Bulitko, Vadim and Greiner, Russell},
booktitle = {European Conference on Machine Learning},
year = {2004},
pages = {566-568},
doi = {10.1007/978-3-540-30115-8_53},
url = {https://mlanthology.org/ecmlpkdd/2004/li2004ecml-batch/}
}