Risk-Sensitive Generative Adversarial Imitation Learning

Abstract

We study risk-sensitive imitation learning where the agent’s goal is to perform at least as well as the expert in terms of a risk profile. We first formulate our risk-sensitive imitation learning setting. We consider the generative adversarial approach to imitation learning (GAIL) and derive an optimization problem for our formulation, which we call it risk- sensitive GAIL (RS-GAIL). We then derive two different versions of our RS-GAIL optimization problem that aim at matching the risk profiles of the agent and the expert w.r.t. Jensen-Shannon (JS) divergence and Wasserstein distance, and develop risk-sensitive generative adversarial imitation learning algorithms based on these optimization problems. We evaluate the performance of our algorithms and compare them with GAIL and the risk-averse imitation learning (RAIL) algorithms in two MuJoCo and two OpenAI classical control tasks.

Cite

Text

Lacotte et al. "Risk-Sensitive Generative Adversarial Imitation Learning." Artificial Intelligence and Statistics, 2019.

Markdown

[Lacotte et al. "Risk-Sensitive Generative Adversarial Imitation Learning." Artificial Intelligence and Statistics, 2019.](https://mlanthology.org/aistats/2019/lacotte2019aistats-risksensitive/)

BibTeX

@inproceedings{lacotte2019aistats-risksensitive,
  title     = {{Risk-Sensitive Generative Adversarial Imitation Learning}},
  author    = {Lacotte, Jonathan and Ghavamzadeh, Mohammad and Chow, Yinlam and Pavone, Marco},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2019},
  pages     = {2154-2163},
  volume    = {89},
  url       = {https://mlanthology.org/aistats/2019/lacotte2019aistats-risksensitive/}
}