Policy Resilience to Environment Poisoning Attack on Reinforcement Learning
Abstract
This paper investigates policy resilience to training-environment poisoning attacks on reinforcement learning (RL) policies, with the goal of recovering the deployment performance of a poisoned RL policy. Due to the fact that policy resilience is an add-on concern to RL algorithms, it must be resource-efficient, time-conserving, and widely applicable without compromising the performance of RL algorithms. This paper proposes such a policy-resilience mechanism based on an idea of sharing the environment knowledge. We summarize the policy resilience as three stages: preparation, diagnosis, recovery. Specifically, we design the mechanism as a federated architecture coupled with a meta-learning approach, pursuing an efficient extraction and sharing of environment knowledge. With the shared knowledge, a poisoned agent can quickly identify the deployment condition and accordingly recover its policy performance. We empirically evaluate the resilience mechanism for both model-based and model-free RL algorithms, showing its effectiveness and efficiency in restoring the deployment performance of a poisoned policy.
Cite
Text
Xu et al. "Policy Resilience to Environment Poisoning Attack on Reinforcement Learning." NeurIPS 2022 Workshops: MLSW, 2022.Markdown
[Xu et al. "Policy Resilience to Environment Poisoning Attack on Reinforcement Learning." NeurIPS 2022 Workshops: MLSW, 2022.](https://mlanthology.org/neuripsw/2022/xu2022neuripsw-policy/)BibTeX
@inproceedings{xu2022neuripsw-policy,
title = {{Policy Resilience to Environment Poisoning Attack on Reinforcement Learning}},
author = {Xu, Hang and Qu, Xinghua and Rabinovich, Zinovi},
booktitle = {NeurIPS 2022 Workshops: MLSW},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/xu2022neuripsw-policy/}
}