Risk-Aware Active Inverse Reinforcement Learning

Abstract

Active learning from demonstration allows a robot to query a human for specific types of input to achieve efficient learning. Existing work has explored a variety of active query strategies; however, to our knowledge, none of these strategies directly minimize the performance risk of the policy the robot is learning. Utilizing recent advances in performance bounds for inverse reinforcement learning, we propose a risk-aware active inverse reinforcement learning algorithm that focuses active queries on areas of the state space with the potential for large generalization error. We show that risk-aware active learning outperforms standard active IRL approaches on gridworld, simulated driving, and table setting tasks, while also providing a performance-based stopping criterion that allows a robot to know when it has received enough demonstrations to safely perform a task.

Cite

Text

Brown et al. "Risk-Aware Active Inverse Reinforcement Learning." Conference on Robot Learning, 2018.

Markdown

[Brown et al. "Risk-Aware Active Inverse Reinforcement Learning." Conference on Robot Learning, 2018.](https://mlanthology.org/corl/2018/brown2018corl-risk/)

BibTeX

@inproceedings{brown2018corl-risk,
  title     = {{Risk-Aware Active Inverse Reinforcement Learning}},
  author    = {Brown, Daniel S. and Cui, Yuchen and Niekum, Scott},
  booktitle = {Conference on Robot Learning},
  year      = {2018},
  pages     = {362-372},
  url       = {https://mlanthology.org/corl/2018/brown2018corl-risk/}
}