Exploration in Relational Worlds

Abstract

One of the key problems in model-based reinforcement learning is balancing exploration and exploitation. Another is learning and acting in large relational domains, in which there is a varying number of objects and relations between them. We provide one of the first solutions to exploring large relational Markov decision processes by developing relational extensions of the concepts of the Explicit Explore or Exploit ( E ^3) algorithm. A key insight is that the inherent generalization of learnt knowledge in the relational representation has profound implications also on the exploration strategy: what in a propositional setting would be considered a novel situation and worth exploration may in the relational setting be an instance of a well-known context in which exploitation is promising. Our experimental evaluation shows the effectiveness and benefit of relational exploration over several propositional benchmark approaches on noisy 3D simulated robot manipulation problems.

Cite

Text

Lang et al. "Exploration in Relational Worlds." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2010. doi:10.1007/978-3-642-15883-4_12

Markdown

[Lang et al. "Exploration in Relational Worlds." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2010.](https://mlanthology.org/ecmlpkdd/2010/lang2010ecmlpkdd-exploration/) doi:10.1007/978-3-642-15883-4_12

BibTeX

@inproceedings{lang2010ecmlpkdd-exploration,
  title     = {{Exploration in Relational Worlds}},
  author    = {Lang, Tobias and Toussaint, Marc and Kersting, Kristian},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2010},
  pages     = {178-194},
  doi       = {10.1007/978-3-642-15883-4_12},
  url       = {https://mlanthology.org/ecmlpkdd/2010/lang2010ecmlpkdd-exploration/}
}