Removing Adversarial Noise in Class Activation Feature Space

Abstract

Deep neural networks (DNNs) are vulnerable to adversarial noise. Pre-processing based defenses could largely remove adversarial noise by processing inputs. However, they are typically affected by the error amplification effect, especially in the front of continuously evolving attacks. To solve this problem, in this paper, we propose to remove adversarial noise by implementing a self-supervised adversarial training mechanism in a class activation feature space. To be specific, we first maximize the disruptions to class activation features of natural examples to craft adversarial examples. Then, we train a denoising model to minimize the distances between the adversarial examples and the natural examples in the class activation feature space. Empirical evaluations demonstrate that our method could significantly enhance adversarial robustness in comparison to previous state-of-the-art approaches, especially against unseen adversarial attacks and adaptive attacks.

Cite

Text

Zhou et al. "Removing Adversarial Noise in Class Activation Feature Space." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.00778

Markdown

[Zhou et al. "Removing Adversarial Noise in Class Activation Feature Space." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/zhou2021iccv-removing/) doi:10.1109/ICCV48922.2021.00778

BibTeX

@inproceedings{zhou2021iccv-removing,
  title     = {{Removing Adversarial Noise in Class Activation Feature Space}},
  author    = {Zhou, Dawei and Wang, Nannan and Peng, Chunlei and Gao, Xinbo and Wang, Xiaoyu and Yu, Jun and Liu, Tongliang},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {7878-7887},
  doi       = {10.1109/ICCV48922.2021.00778},
  url       = {https://mlanthology.org/iccv/2021/zhou2021iccv-removing/}
}