Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models

Abstract

Neural networks are vulnerable to adversarial attacks - small visually imperceptible crafted noise which when added to the input drastically changes the output. The most effective method of defending against adversarial attacks is to use the methodology of adversarial training. We analyze the adversarially trained robust models to study their vulnerability against adversarial attacks at the level of the latent layers. Our analysis reveals that contrary to the input layer which is robust to adversarial attack, the latent layer of these robust models are highly susceptible to adversarial perturbations of small magnitude. Leveraging this information, we introduce a new technique Latent Adversarial Training (LAT) which comprises of fine-tuning the adversarially trained models to ensure the robustness at the feature layers. We also propose Latent Attack (LA), a novel algorithm for constructing adversarial examples. LAT results in a minor improvement in test accuracy and leads to a state-of-the-art adversarial accuracy against the universal first-order adversarial PGD attack which is shown for the MNIST, CIFAR-10, CIFAR-100, SVHN and Restricted ImageNet datasets.

Cite

Text

Kumari et al. "Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models." International Joint Conference on Artificial Intelligence, 2019. doi:10.24963/IJCAI.2019/385

Markdown

[Kumari et al. "Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models." International Joint Conference on Artificial Intelligence, 2019.](https://mlanthology.org/ijcai/2019/kumari2019ijcai-harnessing/) doi:10.24963/IJCAI.2019/385

BibTeX

@inproceedings{kumari2019ijcai-harnessing,
  title     = {{Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models}},
  author    = {Kumari, Nupur and Singh, Mayank and Sinha, Abhishek and Machiraju, Harshitha and Krishnamurthy, Balaji and Balasubramanian, Vineeth N.},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2019},
  pages     = {2779-2785},
  doi       = {10.24963/IJCAI.2019/385},
  url       = {https://mlanthology.org/ijcai/2019/kumari2019ijcai-harnessing/}
}