Looping Suffix Tree-Based Inference of Partially Observable Hidden State

Abstract

We present a solution for inferring hidden state from sensorimotor experience when the environment takes the form of a POMDP with deterministic transition and observation functions. Such environments can appear to be arbitrarily complex and non-deterministic on the surface, but are actually deterministic with respect to the unobserved underlying state. We show that there always exists a finite history-based representation that fully captures the unobserved world state, allowing for perfect prediction of action effects. This representation takes the form of a looping prediction suffix tree (PST). We derive a sound and complete algorithm for learning a looping PST from a sufficient sample of sensorimotor experience. We also give empirical illustrations of the advantages conferred by this approach, and characterize the approximations to the looping PST that are made by existing algorithms such as Variable Length Markov Models, Utile Suffix Memory and Causal State Splitting Reconstruction.

Cite

Text

Holmes and Jr.. "Looping Suffix Tree-Based Inference of Partially Observable Hidden State." International Conference on Machine Learning, 2006. doi:10.1145/1143844.1143896

Markdown

[Holmes and Jr.. "Looping Suffix Tree-Based Inference of Partially Observable Hidden State." International Conference on Machine Learning, 2006.](https://mlanthology.org/icml/2006/holmes2006icml-looping/) doi:10.1145/1143844.1143896

BibTeX

@inproceedings{holmes2006icml-looping,
  title     = {{Looping Suffix Tree-Based Inference of Partially Observable Hidden State}},
  author    = {Holmes, Michael P. and Jr., Charles Lee Isbell},
  booktitle = {International Conference on Machine Learning},
  year      = {2006},
  pages     = {409-416},
  doi       = {10.1145/1143844.1143896},
  url       = {https://mlanthology.org/icml/2006/holmes2006icml-looping/}
}