Dynamically Constructed (PO)MDPs for Adaptive Robot Planning

Abstract

To operate in human-robot coexisting environments, intelligent robots need to simultaneously reason with commonsense knowledge and plan under uncertainty. Markov decision processes (MDPs) and partially observable MDPs (POMDPs), are good at planning under uncertainty toward maximizing long-term rewards; P-LOG, a declarative programming language under Answer Set semantics, is strong in commonsense reasoning. In this paper, we present a novel algorithm called iCORPP to dynamically reason about, and construct (PO)MDPs using P-LOG. iCORPP successfully shields exogenous domain attributes from (PO)MDPs, which limits computational complexity and enables (PO)MDPs to adapt to the value changes these attributes produce. We conduct a number of experimental trials using two example problems in simulation and demonstrate iCORPP on a real robot. Results show significant improvements compared to competitive baselines.

Cite

Text

Zhang et al. "Dynamically Constructed (PO)MDPs for Adaptive Robot Planning." AAAI Conference on Artificial Intelligence, 2017. doi:10.1609/AAAI.V31I1.11042

Markdown

[Zhang et al. "Dynamically Constructed (PO)MDPs for Adaptive Robot Planning." AAAI Conference on Artificial Intelligence, 2017.](https://mlanthology.org/aaai/2017/zhang2017aaai-dynamically/) doi:10.1609/AAAI.V31I1.11042

BibTeX

@inproceedings{zhang2017aaai-dynamically,
  title     = {{Dynamically Constructed (PO)MDPs for Adaptive Robot Planning}},
  author    = {Zhang, Shiqi and Khandelwal, Piyush and Stone, Peter},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {3855-3863},
  doi       = {10.1609/AAAI.V31I1.11042},
  url       = {https://mlanthology.org/aaai/2017/zhang2017aaai-dynamically/}
}