Applying Online Search Techniques to Continuous-State Reinforcement Learning

Abstract

In this paper, we describe methods for efficiently com-puting better solutions to control problems in contin-uous state spaces. We provide algorithms that exploit online search to boost the power of very approximate value functions discovered by traditional reinforcement learning techniques. We examine local searches, where the agent performs a finite-depth lookahead search, and global searches, where the agent performs a search for a trajectory all the way from the current state to a goal state. The key to the success of the local methods lies in tak-ing a value function, which gives a rough solution to the hard problem of finding good trajectories from ev-ery single state, and combining that with online search, which then gives an accurate solution to the easier problem of finding a good trajectory specifically from the current state. The key to the success of the global methods lies in using aggressive state-space search techniques such as uniform-cost search and A*, tamed into a tractable form by exploiting neighborhood relations and trajec-tory constraints that arise from continuous-space dy-namic control.

Cite

Text

Davies et al. "Applying Online Search Techniques to Continuous-State Reinforcement Learning." AAAI Conference on Artificial Intelligence, 1998.

Markdown

[Davies et al. "Applying Online Search Techniques to Continuous-State Reinforcement Learning." AAAI Conference on Artificial Intelligence, 1998.](https://mlanthology.org/aaai/1998/davies1998aaai-applying/)

BibTeX

@inproceedings{davies1998aaai-applying,
  title     = {{Applying Online Search Techniques to Continuous-State Reinforcement Learning}},
  author    = {Davies, Scott and Ng, Andrew Y. and Moore, Andrew W.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {1998},
  pages     = {753-760},
  url       = {https://mlanthology.org/aaai/1998/davies1998aaai-applying/}
}