Obstacle Avoidance Through Reinforcement Learning

Abstract

A method is described for generating plan-like. reflexive. obstacle avoidance behaviour in a mobile robot. The experiments reported here use a simulated vehicle with a primitive range sensor. Avoidance behaviour is encoded as a set of continuous functions of the perceptual input space. These functions are stored using CMACs and trained by a variant of Barto and Sutton's adaptive critic algorithm. As the vehicle explores its surroundings it adapts its responses to sensory stimuli so as to minimise the negative reinforcement arising from collisions. Strategies for local navigation are therefore acquired in an explicitly goal-driven fashion. The resulting trajectories form elegant collision(cid:173) free paths through the environment

Cite

Text

Prescott and Mayhew. "Obstacle Avoidance Through Reinforcement Learning." Neural Information Processing Systems, 1991.

Markdown

[Prescott and Mayhew. "Obstacle Avoidance Through Reinforcement Learning." Neural Information Processing Systems, 1991.](https://mlanthology.org/neurips/1991/prescott1991neurips-obstacle/)

BibTeX

@inproceedings{prescott1991neurips-obstacle,
  title     = {{Obstacle Avoidance Through Reinforcement Learning}},
  author    = {Prescott, Tony J. and Mayhew, John E. W.},
  booktitle = {Neural Information Processing Systems},
  year      = {1991},
  pages     = {523-530},
  url       = {https://mlanthology.org/neurips/1991/prescott1991neurips-obstacle/}
}