Risk-Aware Active Inverse Reinforcement Learning
Abstract
Active learning from demonstration allows a robot to query a human for specific types of input to achieve efficient learning. Existing work has explored a variety of active query strategies; however, to our knowledge, none of these strategies directly minimize the performance risk of the policy the robot is learning. Utilizing recent advances in performance bounds for inverse reinforcement learning, we propose a risk-aware active inverse reinforcement learning algorithm that focuses active queries on areas of the state space with the potential for large generalization error. We show that risk-aware active learning outperforms standard active IRL approaches on gridworld, simulated driving, and table setting tasks, while also providing a performance-based stopping criterion that allows a robot to know when it has received enough demonstrations to safely perform a task.
Cite
Text
Brown et al. "Risk-Aware Active Inverse Reinforcement Learning." Conference on Robot Learning, 2018.Markdown
[Brown et al. "Risk-Aware Active Inverse Reinforcement Learning." Conference on Robot Learning, 2018.](https://mlanthology.org/corl/2018/brown2018corl-risk/)BibTeX
@inproceedings{brown2018corl-risk,
title = {{Risk-Aware Active Inverse Reinforcement Learning}},
author = {Brown, Daniel S. and Cui, Yuchen and Niekum, Scott},
booktitle = {Conference on Robot Learning},
year = {2018},
pages = {362-372},
url = {https://mlanthology.org/corl/2018/brown2018corl-risk/}
}