Adversarial Defence by Diversified Simultaneous Training of Deep Ensembles

Abstract

Learning-based classifiers are susceptible to adversarial examples. Existing defence methods are mostly devised on individual classifiers. Recent studies showed that it is viable to increase adversarial robustness by promoting diversity over an ensemble of models. In this paper, we propose adversarial defence by encouraging ensemble diversity on learning high-level feature representations and gradient dispersion in simultaneous training of deep ensemble networks. We perform extensive evaluations under white-box and black-box attacks including transferred examples and adaptive attacks. Our approach achieves a significant gain of up to 52% in adversarial robustness, compared with the baseline and the state-of-the-art method on image benchmarks with complex data scenes. The proposed approach complements the defence paradigm of adversarial training, and can further boost the performance. The source code is available at https://github.com/ALIS-Lab/AAAI2021-PDD.

Cite

Text

Huang et al. "Adversarial Defence by Diversified Simultaneous Training of Deep Ensembles." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I9.16955

Markdown

[Huang et al. "Adversarial Defence by Diversified Simultaneous Training of Deep Ensembles." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/huang2021aaai-adversarial/) doi:10.1609/AAAI.V35I9.16955

BibTeX

@inproceedings{huang2021aaai-adversarial,
  title     = {{Adversarial Defence by Diversified Simultaneous Training of Deep Ensembles}},
  author    = {Huang, Bo and Ke, Zhiwei and Wang, Yi and Wang, Wei and Shen, Linlin and Liu, Feng},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {7823-7831},
  doi       = {10.1609/AAAI.V35I9.16955},
  url       = {https://mlanthology.org/aaai/2021/huang2021aaai-adversarial/}
}