A Direct Approximation of AIXI Using Logical State Abstractions

Abstract

We propose a practical integration of logical state abstraction with AIXI, a Bayesian optimality notion for reinforcement learning agents, to significantly expand the model class that AIXI agents can be approximated over to complex history-dependent and structured environments. The state representation and reasoning framework is based on higher-order logic, which can be used to define and enumerate complex features on non-Markovian and structured environments. We address the problem of selecting the right subset of features to form state abstractions by adapting the $\Phi$-MDP optimisation criterion from state abstraction theory. Exact Bayesian model learning is then achieved using a suitable generalisation of Context Tree Weighting over abstract state sequences. The resultant architecture can be integrated with different planning algorithms. Experimental results on controlling epidemics on large-scale contact networks validates the agent's performance.

Cite

Text

Yang-Zhao et al. "A Direct Approximation of AIXI Using Logical State Abstractions." Neural Information Processing Systems, 2022.

Markdown

[Yang-Zhao et al. "A Direct Approximation of AIXI Using Logical State Abstractions." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/yangzhao2022neurips-direct/)

BibTeX

@inproceedings{yangzhao2022neurips-direct,
  title     = {{A Direct Approximation of AIXI Using Logical State Abstractions}},
  author    = {Yang-Zhao, Samuel and Wang, Tianyu and Ng, Kee Siong},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/yangzhao2022neurips-direct/}
}