Roping in Uncertainty: Robustness and Regularization in Markov Games

Abstract

We study robust Markov games (RMG) with $s$-rectangular uncertainty. We show a general equivalence between computing a robust Nash equilibrium (RNE) of a $s$-rectangular RMG and computing a Nash equilibrium (NE) of an appropriately constructed regularized MG. The equivalence result yields a planning algorithm for solving $s$-rectangular RMGs, as well as provable robustness guarantees for policies computed using regularized methods. However, we show that even for just reward-uncertain two-player zero-sum matrix games, computing an RNE is PPAD-hard. Consequently, we derive a special uncertainty structure called efficient player-decomposability and show that RNE for two-player zero-sum RMG in this class can be provably solved in polynomial time. This class includes commonly used uncertainty sets such as $L_1$ and $L_\infty$ ball uncertainty sets.

Cite

Text

Mcmahan et al. "Roping in Uncertainty: Robustness and Regularization in Markov Games." International Conference on Machine Learning, 2024.

Markdown

[Mcmahan et al. "Roping in Uncertainty: Robustness and Regularization in Markov Games." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/mcmahan2024icml-roping/)

BibTeX

@inproceedings{mcmahan2024icml-roping,
  title     = {{Roping in Uncertainty: Robustness and Regularization in Markov Games}},
  author    = {Mcmahan, Jeremy and Artiglio, Giovanni and Xie, Qiaomin},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {35267-35295},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/mcmahan2024icml-roping/}
}