Looping Suffix Tree-Based Inference of Partially Observable Hidden State
Abstract
We present a solution for inferring hidden state from sensorimotor experience when the environment takes the form of a POMDP with deterministic transition and observation functions. Such environments can appear to be arbitrarily complex and non-deterministic on the surface, but are actually deterministic with respect to the unobserved underlying state. We show that there always exists a finite history-based representation that fully captures the unobserved world state, allowing for perfect prediction of action effects. This representation takes the form of a looping prediction suffix tree (PST). We derive a sound and complete algorithm for learning a looping PST from a sufficient sample of sensorimotor experience. We also give empirical illustrations of the advantages conferred by this approach, and characterize the approximations to the looping PST that are made by existing algorithms such as Variable Length Markov Models, Utile Suffix Memory and Causal State Splitting Reconstruction.
Cite
Text
Holmes and Jr.. "Looping Suffix Tree-Based Inference of Partially Observable Hidden State." International Conference on Machine Learning, 2006. doi:10.1145/1143844.1143896Markdown
[Holmes and Jr.. "Looping Suffix Tree-Based Inference of Partially Observable Hidden State." International Conference on Machine Learning, 2006.](https://mlanthology.org/icml/2006/holmes2006icml-looping/) doi:10.1145/1143844.1143896BibTeX
@inproceedings{holmes2006icml-looping,
title = {{Looping Suffix Tree-Based Inference of Partially Observable Hidden State}},
author = {Holmes, Michael P. and Jr., Charles Lee Isbell},
booktitle = {International Conference on Machine Learning},
year = {2006},
pages = {409-416},
doi = {10.1145/1143844.1143896},
url = {https://mlanthology.org/icml/2006/holmes2006icml-looping/}
}