Structured Solution Methods for Non-Markovian Decision Processes

Abstract

Markov Decision Processes (MDPs), currently a popular method for modeling and solving decision theoretic planning problems, are limited by the Markovian assumption: rewards and dynamics depend on the current state only, and not on previous history. Non-Markovian decision processes (NMDPs) can also be defined, but then the more tractable solution techniques developed for MDP's cannot be directly applied. In this paper, we show how an NMDP, in which temporal logic is used to specify history dependence, can be automatically converted into an equivalent MDP by adding appropriate temporal variables. The resulting MDP can be represented in a structured fashion and solved using structured policy construction methods. In many cases, this offers significant computational advantagesover previous proposals for solving NMDPs. 1 Introduction Markov decision processes (MDPs) have proven to be an effective modeling and computational paradigm for decisiontheoretic planning (DTP). MDPs allow one to d...

Cite

Text

Bacchus et al. "Structured Solution Methods for Non-Markovian Decision Processes." AAAI Conference on Artificial Intelligence, 1997.

Markdown

[Bacchus et al. "Structured Solution Methods for Non-Markovian Decision Processes." AAAI Conference on Artificial Intelligence, 1997.](https://mlanthology.org/aaai/1997/bacchus1997aaai-structured/)

BibTeX

@inproceedings{bacchus1997aaai-structured,
  title     = {{Structured Solution Methods for Non-Markovian Decision Processes}},
  author    = {Bacchus, Fahiem and Boutilier, Craig and Grove, Adam J.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {1997},
  pages     = {112-117},
  url       = {https://mlanthology.org/aaai/1997/bacchus1997aaai-structured/}
}