Dynamically Constructed (PO)MDPs for Adaptive Robot Planning
Abstract
To operate in human-robot coexisting environments, intelligent robots need to simultaneously reason with commonsense knowledge and plan under uncertainty. Markov decision processes (MDPs) and partially observable MDPs (POMDPs), are good at planning under uncertainty toward maximizing long-term rewards; P-LOG, a declarative programming language under Answer Set semantics, is strong in commonsense reasoning. In this paper, we present a novel algorithm called iCORPP to dynamically reason about, and construct (PO)MDPs using P-LOG. iCORPP successfully shields exogenous domain attributes from (PO)MDPs, which limits computational complexity and enables (PO)MDPs to adapt to the value changes these attributes produce. We conduct a number of experimental trials using two example problems in simulation and demonstrate iCORPP on a real robot. Results show significant improvements compared to competitive baselines.
Cite
Text
Zhang et al. "Dynamically Constructed (PO)MDPs for Adaptive Robot Planning." AAAI Conference on Artificial Intelligence, 2017. doi:10.1609/AAAI.V31I1.11042Markdown
[Zhang et al. "Dynamically Constructed (PO)MDPs for Adaptive Robot Planning." AAAI Conference on Artificial Intelligence, 2017.](https://mlanthology.org/aaai/2017/zhang2017aaai-dynamically/) doi:10.1609/AAAI.V31I1.11042BibTeX
@inproceedings{zhang2017aaai-dynamically,
title = {{Dynamically Constructed (PO)MDPs for Adaptive Robot Planning}},
author = {Zhang, Shiqi and Khandelwal, Piyush and Stone, Peter},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2017},
pages = {3855-3863},
doi = {10.1609/AAAI.V31I1.11042},
url = {https://mlanthology.org/aaai/2017/zhang2017aaai-dynamically/}
}