Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability

Abstract

Interesting real-world datasets often exhibit nonlinear, noisy, continuous-valued states that are unexplorable, are poorly described by first principles, and are only partially observable. If partial observability can be overcome, these constraints suggest the use of model-based reinforcement learning. We experiment with manifold embeddings as the reconstructed observable state-space of an off-line, model-based reinforcement learning approach to control. We demonstrate the embedding of a system changes as a result of learning and that the best performing embeddings well-represent the dynamics of both the uncontrolled and adaptively controlled system. We apply this approach in simulation to learn a neurostimulation policy that is more efficient in treating epilepsy than conventional policies. We then demonstrate the learned policy completely suppressing seizures in real-world neurostimulation experiments on actual animal brain slices.

Cite

Text

Bush and Pineau. "Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability." Neural Information Processing Systems, 2009.

Markdown

[Bush and Pineau. "Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability." Neural Information Processing Systems, 2009.](https://mlanthology.org/neurips/2009/bush2009neurips-manifold/)

BibTeX

@inproceedings{bush2009neurips-manifold,
  title     = {{Manifold Embeddings for Model-Based Reinforcement Learning Under Partial Observability}},
  author    = {Bush, Keith and Pineau, Joelle},
  booktitle = {Neural Information Processing Systems},
  year      = {2009},
  pages     = {189-197},
  url       = {https://mlanthology.org/neurips/2009/bush2009neurips-manifold/}
}