Generalised Task Planning with First-Order Function Approximation

Abstract

Real world robotics often operates in uncertain and dynamic environments where generalisation over different scenarios is of practical interest. In the absence of a model, value-based reinforcement learning can be used to learn a goal-directed policy. Typically, the interaction between robots and the objects in the environment exhibit a first-order structure. We introduce first-order, or relational, features to represent an approximation of the Q-function so that it can induce a generalised policy. Empirical results for a service robot domain show that our online relational reinforcement learning method is scalable to large scale problems and enables transfer learning between different problems and simulation environments with dissimilar transition dynamics.

Cite

Text

Ng and Petrick. "Generalised Task Planning with First-Order Function Approximation." Conference on Robot Learning, 2021.

Markdown

[Ng and Petrick. "Generalised Task Planning with First-Order Function Approximation." Conference on Robot Learning, 2021.](https://mlanthology.org/corl/2021/ng2021corl-generalised/)

BibTeX

@inproceedings{ng2021corl-generalised,
  title     = {{Generalised Task Planning with First-Order Function Approximation}},
  author    = {Ng, Jun Hao Alvin and Petrick, Ronald P.A.},
  booktitle = {Conference on Robot Learning},
  year      = {2021},
  pages     = {1595-1610},
  volume    = {164},
  url       = {https://mlanthology.org/corl/2021/ng2021corl-generalised/}
}