A Direct Approximation of AIXI Using Logical State Abstractions
Abstract
We propose a practical integration of logical state abstraction with AIXI, a Bayesian optimality notion for reinforcement learning agents, to significantly expand the model class that AIXI agents can be approximated over to complex history-dependent and structured environments. The state representation and reasoning framework is based on higher-order logic, which can be used to define and enumerate complex features on non-Markovian and structured environments. We address the problem of selecting the right subset of features to form state abstractions by adapting the $\Phi$-MDP optimisation criterion from state abstraction theory. Exact Bayesian model learning is then achieved using a suitable generalisation of Context Tree Weighting over abstract state sequences. The resultant architecture can be integrated with different planning algorithms. Experimental results on controlling epidemics on large-scale contact networks validates the agent's performance.
Cite
Text
Yang-Zhao et al. "A Direct Approximation of AIXI Using Logical State Abstractions." Neural Information Processing Systems, 2022.Markdown
[Yang-Zhao et al. "A Direct Approximation of AIXI Using Logical State Abstractions." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/yangzhao2022neurips-direct/)BibTeX
@inproceedings{yangzhao2022neurips-direct,
title = {{A Direct Approximation of AIXI Using Logical State Abstractions}},
author = {Yang-Zhao, Samuel and Wang, Tianyu and Ng, Kee Siong},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/yangzhao2022neurips-direct/}
}