Relational Reinforcement Learning for Planning with Exogenous Effects

Abstract

Probabilistic planners have improved recently to the point that they can solve difficult tasks with complex and expressive models. In contrast, learners cannot tackle yet the expressive models that planners do, which forces complex models to be mostly handcrafted. We propose a new learning approach that can learn relational probabilistic models with both action effects and exogenous effects. The proposed learning approach combines a multi-valued variant of inductive logic programming for the generation of candidate models, with an optimization method to select the best set of planning operators to model a problem. We also show how to combine this learner with reinforcement learning algorithms to solve complete problems. Finally, experimental validation is provided that shows improvements over previous work in both simulation and a robotic task. The robotic task involves a dynamic scenario with several agents where a manipulator robot has to clear the tableware on a table. We show that the exogenous effects learned by our approach allowed the robot to clear the table in a more efficient way.

Cite

Text

Martínez et al. "Relational Reinforcement Learning for Planning with Exogenous Effects." Journal of Machine Learning Research, 2017.

Markdown

[Martínez et al. "Relational Reinforcement Learning for Planning with Exogenous Effects." Journal of Machine Learning Research, 2017.](https://mlanthology.org/jmlr/2017/martinez2017jmlr-relational/)

BibTeX

@article{martinez2017jmlr-relational,
  title     = {{Relational Reinforcement Learning for Planning with Exogenous Effects}},
  author    = {Martínez, David and Alenyà, Guillem and Ribeiro, Tony and Inoue, Katsumi and Torras, Carme},
  journal   = {Journal of Machine Learning Research},
  year      = {2017},
  pages     = {1-44},
  volume    = {18},
  url       = {https://mlanthology.org/jmlr/2017/martinez2017jmlr-relational/}
}