Structured Kernel-Based Reinforcement Learning

Abstract

Kernel-based reinforcement learning (KBRL) is a popular approach to learning non-parametric value function approximations. In this paper, we present structured KBRL, a paradigm for kernel-based RL that allows for modeling independencies in the transition and reward models of problems. Real-world problems often exhibit this structure and can be solved more efficiently when it is modeled. We make three contributions. First, we motivate our work, define a structured backup operator, and prove that it is a contraction. Second, we show how to evaluate our operator efficiently. Our analysis reveals that the fixed point of the operator is the optimal value function in a special factored MDP. Finally, we evaluate our method on a synthetic problem and compare it to two KBRL baselines. In most experiments, we learn better policies than the baselines from an order of magnitude less training data.

Cite

Text

Kveton and Theocharous. "Structured Kernel-Based Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2013. doi:10.1609/AAAI.V27I1.8669

Markdown

[Kveton and Theocharous. "Structured Kernel-Based Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2013.](https://mlanthology.org/aaai/2013/kveton2013aaai-structured/) doi:10.1609/AAAI.V27I1.8669

BibTeX

@inproceedings{kveton2013aaai-structured,
  title     = {{Structured Kernel-Based Reinforcement Learning}},
  author    = {Kveton, Branislav and Theocharous, Georgios},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2013},
  pages     = {569-575},
  doi       = {10.1609/AAAI.V27I1.8669},
  url       = {https://mlanthology.org/aaai/2013/kveton2013aaai-structured/}
}