CORL: A Continuous-State Offset-Dynamics Reinforcement Learner

Abstract

Continuous state spaces and stochastic, switching dynamics characterize a number of rich, real-world domains, such as robot navigation across varying terrain. We describe a reinforcement-learning algorithm for learning in these domains and prove for certain environments the algorithm is probably approximately correct with a sample complexity that scales polynomially with the state-space dimension. Unfortunately, no optimal planning techniques exist in general for such problems; instead we use fitted value iteration to solve the learned MDP, and include the error due to approximate planning in our bounds. Finally, we report an experiment using a robotic car driving over varying terrain to demonstrate that these dynamics representations adequately capture real-world dynamics and that our algorithm can be used to efficiently solve such problems.

Cite

Text

Brunskill et al. "CORL: A Continuous-State Offset-Dynamics Reinforcement Learner." Conference on Uncertainty in Artificial Intelligence, 2008.

Markdown

[Brunskill et al. "CORL: A Continuous-State Offset-Dynamics Reinforcement Learner." Conference on Uncertainty in Artificial Intelligence, 2008.](https://mlanthology.org/uai/2008/brunskill2008uai-corl/)

BibTeX

@inproceedings{brunskill2008uai-corl,
  title     = {{CORL: A Continuous-State Offset-Dynamics Reinforcement Learner}},
  author    = {Brunskill, Emma and Leffler, Bethany R. and Li, Lihong and Littman, Michael L. and Roy, Nicholas},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2008},
  pages     = {53-61},
  url       = {https://mlanthology.org/uai/2008/brunskill2008uai-corl/}
}