Robust Combination of Local Controllers

Abstract

Finding solutions to high dimensional Markov Decision Processes (MDPs) is a difficult problem, especially in the presence of uncertainty or if the actions and time measurements are continuous. Frequently this difficulty can be alleviated by the availability of problem-specific knowledge. For example, it may be relatively easy to design controllers that are good locally, though having no global guarantees. We propose a nonparametric method to combine these local controllers to obtain globally good solutions. We apply this formulation to two types of problems: motion planning (stochastic shortest path problems) and discounted-cost MDPs. For motion planning, we argue that only considering the expected cost of a path may be overly simplistic in the presence of uncertainty. We propose an alternative: finding the minimum cost path, subject to the constraint that the robot must reach the goal with high probability. For this problem, we prove that a polynomial number of samples is sufficient to obtain a high probability path. For discounted MDPs, we consider various problem formulations that explicitly deal with model uncertainty. We provide empirical evidence of the usefulness of these approaches using the control of a robot arm.

Cite

Text

Guestrin and Ormoneit. "Robust Combination of Local Controllers." Conference on Uncertainty in Artificial Intelligence, 2001.

Markdown

[Guestrin and Ormoneit. "Robust Combination of Local Controllers." Conference on Uncertainty in Artificial Intelligence, 2001.](https://mlanthology.org/uai/2001/guestrin2001uai-robust/)

BibTeX

@inproceedings{guestrin2001uai-robust,
  title     = {{Robust Combination of Local Controllers}},
  author    = {Guestrin, Carlos and Ormoneit, Dirk},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2001},
  pages     = {178-185},
  url       = {https://mlanthology.org/uai/2001/guestrin2001uai-robust/}
}