Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability
Abstract
Interesting real-world datasets often exhibit nonlinear, noisy, continuous-valued states that are unexplorable, are poorly described by first principles, and are only partially observable. If partial observability can be overcome, these constraints suggest the use of model-based reinforcement learning. We experiment with manifold embeddings as the reconstructed observable state-space of an off-line, model-based reinforcement learning approach to control. We demonstrate the embedding of a system changes as a result of learning and that the best performing embeddings well-represent the dynamics of both the uncontrolled and adaptively controlled system. We apply this approach in simulation to learn a neurostimulation policy that is more efficient in treating epilepsy than conventional policies. We then demonstrate the learned policy completely suppressing seizures in real-world neurostimulation experiments on actual animal brain slices.
Cite
Text
Bush and Pineau. "Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability." Neural Information Processing Systems, 2009.Markdown
[Bush and Pineau. "Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability." Neural Information Processing Systems, 2009.](https://mlanthology.org/neurips/2009/bush2009neurips-manifold/)BibTeX
@inproceedings{bush2009neurips-manifold,
title = {{Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability}},
author = {Bush, Keith and Pineau, Joelle},
booktitle = {Neural Information Processing Systems},
year = {2009},
pages = {189-197},
url = {https://mlanthology.org/neurips/2009/bush2009neurips-manifold/}
}