Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations

Abstract

Robust reinforcement learning (RL) seeks to train policies that can perform well under environment perturbations or adversarial attacks. Existing approaches typically assume that the space of possible perturbations remains the same across timesteps. However, in many settings, the space of possible perturbations at a given timestep depends on past perturbations. We formally introduce temporally-coupled perturbations, presenting a novel challenge for existing robust RL methods. To tackle this challenge, we propose GRAD, a novel game-theoretic approach that treats the temporally-coupled robust RL problem as a partially-observable two-player zero-sum game. By finding an approximate equilibrium in this game, GRAD ensures the agent's robustness against temporally-coupled perturbations. Empirical experiments on a variety of continuous control tasks demonstrate that our proposed approach exhibits significant robustness advantages compared to baselines against both standard and temporally-coupled attacks, in both state and action spaces.

Cite

Text

Liang et al. "Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations." ICML 2023 Workshops: AdvML-Frontiers, 2023.

Markdown

[Liang et al. "Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations." ICML 2023 Workshops: AdvML-Frontiers, 2023.](https://mlanthology.org/icmlw/2023/liang2023icmlw-gametheoretic/)

BibTeX

@inproceedings{liang2023icmlw-gametheoretic,
  title     = {{Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations}},
  author    = {Liang, Yongyuan and Sun, Yanchao and Zheng, Ruijie and Liu, Xiangyu and Sandholm, Tuomas and Huang, Furong and McAleer, Stephen Marcus},
  booktitle = {ICML 2023 Workshops: AdvML-Frontiers},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/liang2023icmlw-gametheoretic/}
}