Defending Against Unknown Corrupted Agents: Reinforcement Learning of Adversarially Robust Nash Equilibria
Abstract
We consider a Multi-agent Reinforcement Learning (MARL) setting, in which an attacker can arbitrarily corrupt any subset of up to $k$ out of $n$ agents at deployment. Our goal is to design agents that are robust against such an attack, by accounting for the presence of corrupted agents at test time. To that end, we introduce a novel solution concept, the Adversarially Robust Nash Equilibrium (ARNEQ), and provide theoretical proof of its existence in general-sum Markov games. Furthermore, we introduce a proof-of-concept model-based approach to computing it and theoretically prove its convergence under standard assumptions. We also present a practical approach called Adversarially Robust Training (ART), an independent learning algorithm based on stochastic gradient descent ascent. Our experiments in both cooperative and mixed cooperative-competitive environments demonstrate ART's effectiveness and practical value in enhancing MARL resilience against adversarial behavior.
Cite
Text
Nika et al. "Defending Against Unknown Corrupted Agents: Reinforcement Learning of Adversarially Robust Nash Equilibria." Transactions on Machine Learning Research, 2024.Markdown
[Nika et al. "Defending Against Unknown Corrupted Agents: Reinforcement Learning of Adversarially Robust Nash Equilibria." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/nika2024tmlr-defending/)BibTeX
@article{nika2024tmlr-defending,
title = {{Defending Against Unknown Corrupted Agents: Reinforcement Learning of Adversarially Robust Nash Equilibria}},
author = {Nika, Andi and Nöther, Jonathan and Singla, Adish and Radanovic, Goran},
journal = {Transactions on Machine Learning Research},
year = {2024},
url = {https://mlanthology.org/tmlr/2024/nika2024tmlr-defending/}
}