A Bayesian Approach to Robust Inverse Reinforcement Learning

Abstract

We consider a Bayesian approach to offline model-based inverse reinforcement learning (IRL). The proposed framework differs from existing offline model-based IRL approaches by performing simultaneous estimation of the expert’s reward function and subjective model of environment dynamics. We make use of a class of prior distributions which parameterizes how accurate the expert’s model of the environment is to develop efficient algorithms to estimate the expert’s reward and subjective dynamics in high-dimensional settings. Our analysis reveals a novel insight that the estimated policy exhibits robust performance when the expert is believed (a priori) to have a highly accurate model of the environment. We verify this observation in the MuJoCo environments and show that our algorithms outperform state-of-the-art offline IRL algorithms.

Cite

Text

Wei et al. "A Bayesian Approach to Robust Inverse Reinforcement Learning." Conference on Robot Learning, 2023.

Markdown

[Wei et al. "A Bayesian Approach to Robust Inverse Reinforcement Learning." Conference on Robot Learning, 2023.](https://mlanthology.org/corl/2023/wei2023corl-bayesian/)

BibTeX

@inproceedings{wei2023corl-bayesian,
  title     = {{A Bayesian Approach to Robust Inverse Reinforcement Learning}},
  author    = {Wei, Ran and Zeng, Siliang and Li, Chenliang and Garcia, Alfredo and McDonald, Anthony D and Hong, Mingyi},
  booktitle = {Conference on Robot Learning},
  year      = {2023},
  pages     = {2304-2322},
  volume    = {229},
  url       = {https://mlanthology.org/corl/2023/wei2023corl-bayesian/}
}