Monte-Carlo Exploration for Deterministic Planning

Abstract

Search methods based on Monte-Carlo simulation have recently led to breakthrough performance improvements in difficult game-playing domains such as Go and General Game Playing. Monte-Carlo Random Walk (MRW) planning applies Monte-Carlo ideas to deterministic classical planning. In the forward chaining planner Arvand, Monte-Carlo random walks are used to explore the local neighborhood of a search state for action selection. In contrast to the stochastic local search approach used in the recent planner Identidem, random walks yield a larger and unbiased sample of the search neighborhood, and require state evaluations only at the endpoints of each walk. On IPC-4 competition problems, the performance of Arvand is competitive with state of the art systems. Hootan Nakhost, Martin M�ller

Cite

Text

Nakhost and Müller. "Monte-Carlo Exploration for Deterministic Planning." International Joint Conference on Artificial Intelligence, 2009.

Markdown

[Nakhost and Müller. "Monte-Carlo Exploration for Deterministic Planning." International Joint Conference on Artificial Intelligence, 2009.](https://mlanthology.org/ijcai/2009/nakhost2009ijcai-monte/)

BibTeX

@inproceedings{nakhost2009ijcai-monte,
  title     = {{Monte-Carlo Exploration for Deterministic Planning}},
  author    = {Nakhost, Hootan and Müller, Martin},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2009},
  pages     = {1766-1771},
  url       = {https://mlanthology.org/ijcai/2009/nakhost2009ijcai-monte/}
}