Markovian State and Action Abstractions for MDPs via Hierarchical MCTS

Abstract

State abstraction is an important technique for scaling MD Palgorithms. As is well known, however, it introduces difficulties due to the non-Markovian nature of state-abstracted models. Whereas prior approaches rely upon ad hoc fixes for this issue, we propose instead to view the state-abstracted model as a POMDP and show that we can thereby take advantage of state abstraction without sacrificing the Markov property. We further exploit the hierarchical structure introduced by state abstraction by extending the theory of options to a POMDP setting. In this context we propose a hierarchical Monte Carlo tree search algorithm and show that it converges to a recursively optimal hierarchical policy. Both theoretical and empirical results suggest that abstracting an MDP into a POMDP yields a scalable solution approach. PDF

Cite

Text

Bai et al. "Markovian State and Action Abstractions for MDPs via Hierarchical MCTS." International Joint Conference on Artificial Intelligence, 2016.

Markdown

[Bai et al. "Markovian State and Action Abstractions for MDPs via Hierarchical MCTS." International Joint Conference on Artificial Intelligence, 2016.](https://mlanthology.org/ijcai/2016/bai2016ijcai-markovian/)

BibTeX

@inproceedings{bai2016ijcai-markovian,
  title     = {{Markovian State and Action Abstractions for MDPs via Hierarchical MCTS}},
  author    = {Bai, Aijun and Srivastava, Siddharth and Russell, Stuart},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {3029-3039},
  url       = {https://mlanthology.org/ijcai/2016/bai2016ijcai-markovian/}
}