Policy Resilience to Environment Poisoning Attack on Reinforcement Learning

Abstract

This paper investigates policy resilience to training-environment poisoning attacks on reinforcement learning (RL) policies, with the goal of recovering the deployment performance of a poisoned RL policy. Due to the fact that policy resilience is an add-on concern to RL algorithms, it must be resource-efficient, time-conserving, and widely applicable without compromising the performance of RL algorithms. This paper proposes such a policy-resilience mechanism based on an idea of sharing the environment knowledge. We summarize the policy resilience as three stages: preparation, diagnosis, recovery. Specifically, we design the mechanism as a federated architecture coupled with a meta-learning approach, pursuing an efficient extraction and sharing of environment knowledge. With the shared knowledge, a poisoned agent can quickly identify the deployment condition and accordingly recover its policy performance. We empirically evaluate the resilience mechanism for both model-based and model-free RL algorithms, showing its effectiveness and efficiency in restoring the deployment performance of a poisoned policy.

Cite

Text

Xu et al. "Policy Resilience to Environment Poisoning Attack on Reinforcement Learning." NeurIPS 2022 Workshops: MLSW, 2022.

Markdown

[Xu et al. "Policy Resilience to Environment Poisoning Attack on Reinforcement Learning." NeurIPS 2022 Workshops: MLSW, 2022.](https://mlanthology.org/neuripsw/2022/xu2022neuripsw-policy/)

BibTeX

@inproceedings{xu2022neuripsw-policy,
  title     = {{Policy Resilience to Environment Poisoning Attack on Reinforcement Learning}},
  author    = {Xu, Hang and Qu, Xinghua and Rabinovich, Zinovi},
  booktitle = {NeurIPS 2022 Workshops: MLSW},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/xu2022neuripsw-policy/}
}