Semi-Markov Reinforcement Learning for Stochastic Resource Collection

Abstract

We show that the task of collecting stochastic, spatially distributed resources (Stochastic Resource Collection, SRC) may be considered as a Semi-Markov-Decision-Process. Our Deep-Q-Network (DQN) based approach uses a novel scalable and transferable artificial neural network architecture. The concrete use-case of the SRC is an officer (single agent) trying to maximize the amount of fined parking violations in his area. We evaluate our approach on a environment based on the real-world parking data of the city of Melbourne. In small, hence simple, settings with short distances between resources and few simultaneous violations, our approach is comparable to previous work. When the size of the network grows (and hence the amount of resources) our solution significantly outperforms preceding methods. Moreover, applying a trained agent to a non-overlapping new area outperforms existing approaches.

Cite

Text

Schmoll and Schubert. "Semi-Markov Reinforcement Learning for Stochastic Resource Collection." International Joint Conference on Artificial Intelligence, 2020. doi:10.24963/IJCAI.2020/463

Markdown

[Schmoll and Schubert. "Semi-Markov Reinforcement Learning for Stochastic Resource Collection." International Joint Conference on Artificial Intelligence, 2020.](https://mlanthology.org/ijcai/2020/schmoll2020ijcai-semi/) doi:10.24963/IJCAI.2020/463

BibTeX

@inproceedings{schmoll2020ijcai-semi,
  title     = {{Semi-Markov Reinforcement Learning for Stochastic Resource Collection}},
  author    = {Schmoll, Sebastian and Schubert, Matthias},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {3349-3355},
  doi       = {10.24963/IJCAI.2020/463},
  url       = {https://mlanthology.org/ijcai/2020/schmoll2020ijcai-semi/}
}