Checking Robustness of Representations Learned by Deep Neural Networks

Abstract

Recent works have shown the vulnerability of deep neural networks to adversarial or out-of-distribution examples. This weakness may come from the fact that training deep models often leads to extracting spurious correlations between image classes and some characteristics of images used for training. As demonstrated, popular, ready-to-use models like the ResNet or the EfficientNet may rely on the non-obvious and counterintuitive features. Detection of these weaknesses is often difficult as classification accuracy is excellent and does not indicate that the model is non-robust. To address this problem, we propose a new method and a measure called robustness score. The method allows indicating which classes are recognized by the deep model using non-robust representations, i.e. representations based on spurious correlations. Since the root of this problem lies in the quality of the training data, our method allows us to analyze the training dataset in terms of the existence of these non-obvious spurious correlations. This knowledge can be used to attack the model by finding adversarial images. Consequently, our method can expose threats to the model’s reliability, which should be addressed to increase the certainty of classification decisions. The method was verified using the ImageNet and Pascal VOC datasets, revealing many flaws that affect the final quality of deep models trained on these datasets.

Cite

Text

Szyc et al. "Checking Robustness of Representations Learned by Deep Neural Networks." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2021. doi:10.1007/978-3-030-86517-7_25

Markdown

[Szyc et al. "Checking Robustness of Representations Learned by Deep Neural Networks." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2021.](https://mlanthology.org/ecmlpkdd/2021/szyc2021ecmlpkdd-checking/) doi:10.1007/978-3-030-86517-7_25

BibTeX

@inproceedings{szyc2021ecmlpkdd-checking,
  title     = {{Checking Robustness of Representations Learned by Deep Neural Networks}},
  author    = {Szyc, Kamil and Walkowiak, Tomasz and Maciejewski, Henryk},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2021},
  pages     = {399-414},
  doi       = {10.1007/978-3-030-86517-7_25},
  url       = {https://mlanthology.org/ecmlpkdd/2021/szyc2021ecmlpkdd-checking/}
}