Demiguise Attack: Crafting Invisible Semantic Adversarial Perturbations with Perceptual Similarity
Abstract
Deep neural networks (DNNs) have been found to be vulnerable to adversarial examples. Adversarial examples are malicious images with visually imperceptible perturbations. While these carefully crafted perturbations restricted with tight Lp norm bounds are small, they are still easily perceivable by humans. These perturbations also have limited success rates when attacking black-box models or models with defenses like noise reduction filters. To solve these problems, we propose Demiguise Attack, crafting "unrestricted" perturbations with Perceptual Similarity. Specifically, we can create powerful and photorealistic adversarial examples by manipulating semantic information based on Perceptual Similarity. Adversarial examples we generate are friendly to the human visual system (HVS), although the perturbations are of large magnitudes. We extend widely-used attacks with our approach, enhancing adversarial effectiveness impressively while contributing to imperceptibility. Extensive experiments show that the proposed method not only outperforms various state-of-the-art attacks in terms of fooling rate, transferability, and robustness against defenses but can also improve attacks effectively. In addition, we also notice that our implementation can simulate illumination and contrast changes that occur in real-world scenarios, which will contribute to exposing the blind spots of DNNs.
Cite
Text
Wang et al. "Demiguise Attack: Crafting Invisible Semantic Adversarial Perturbations with Perceptual Similarity." International Joint Conference on Artificial Intelligence, 2021. doi:10.24963/IJCAI.2021/430Markdown
[Wang et al. "Demiguise Attack: Crafting Invisible Semantic Adversarial Perturbations with Perceptual Similarity." International Joint Conference on Artificial Intelligence, 2021.](https://mlanthology.org/ijcai/2021/wang2021ijcai-demiguise/) doi:10.24963/IJCAI.2021/430BibTeX
@inproceedings{wang2021ijcai-demiguise,
title = {{Demiguise Attack: Crafting Invisible Semantic Adversarial Perturbations with Perceptual Similarity}},
author = {Wang, Yajie and Wu, Shangbo and Jiang, Wenyi and Hao, Shengang and Tan, Yu-an and Zhang, Quanxin},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2021},
pages = {3125-3133},
doi = {10.24963/IJCAI.2021/430},
url = {https://mlanthology.org/ijcai/2021/wang2021ijcai-demiguise/}
}