A Bayesian Approach to Robust Inverse Reinforcement Learning
Abstract
We consider a Bayesian approach to offline model-based inverse reinforcement learning (IRL). The proposed framework differs from existing offline model-based IRL approaches by performing simultaneous estimation of the expert’s reward function and subjective model of environment dynamics. We make use of a class of prior distributions which parameterizes how accurate the expert’s model of the environment is to develop efficient algorithms to estimate the expert’s reward and subjective dynamics in high-dimensional settings. Our analysis reveals a novel insight that the estimated policy exhibits robust performance when the expert is believed (a priori) to have a highly accurate model of the environment. We verify this observation in the MuJoCo environments and show that our algorithms outperform state-of-the-art offline IRL algorithms.
Cite
Text
Wei et al. "A Bayesian Approach to Robust Inverse Reinforcement Learning." Conference on Robot Learning, 2023.Markdown
[Wei et al. "A Bayesian Approach to Robust Inverse Reinforcement Learning." Conference on Robot Learning, 2023.](https://mlanthology.org/corl/2023/wei2023corl-bayesian/)BibTeX
@inproceedings{wei2023corl-bayesian,
title = {{A Bayesian Approach to Robust Inverse Reinforcement Learning}},
author = {Wei, Ran and Zeng, Siliang and Li, Chenliang and Garcia, Alfredo and McDonald, Anthony D and Hong, Mingyi},
booktitle = {Conference on Robot Learning},
year = {2023},
pages = {2304-2322},
volume = {229},
url = {https://mlanthology.org/corl/2023/wei2023corl-bayesian/}
}