Online Planning in MDPs with Stochastic Durative Actions

Abstract

Stochastic planning problems are typically modeled as Markov Decision Processes, in which actions are assumed to be instantaneous and applied sequentially. Yet, real-world actions often have durations and are applied concurrently. This paper presents an online planning approach that can deal with durative actions with stochastic outcomes. Our approach relies on Monte Carlo Tree Search with a new backpropagation procedure and temporal reasoning techniques that address the need to not only choose which action to execute, but also when to execute it. We also introduce a novel heuristic that combines reasoning about time and probabilities. Overall, we present the first online planner for stochastic temporal planning, solving a richer problem representation than previous work while achieving state-of-the-art empirical results.

Cite

Text

Berman et al. "Online Planning in MDPs with Stochastic Durative Actions." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/941

Markdown

[Berman et al. "Online Planning in MDPs with Stochastic Durative Actions." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/berman2025ijcai-online/) doi:10.24963/IJCAI.2025/941

BibTeX

@inproceedings{berman2025ijcai-online,
  title     = {{Online Planning in MDPs with Stochastic Durative Actions}},
  author    = {Berman, Tal and Brafman, Ronen I. and Karpas, Erez},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {8465-8473},
  doi       = {10.24963/IJCAI.2025/941},
  url       = {https://mlanthology.org/ijcai/2025/berman2025ijcai-online/}
}