Learning Representational Invariance Instead of Categorization

Abstract

The current most accurate models of image object categorization are deep neural networks trained on large labeled data sets. Minimizing a classification loss between the predictions of the network and the true labels has proven an effective way to learn discriminative functions of the object classes. However, recent studies have suggested that such models learn highly discriminative features that are not aligned with visual perception and might be at the root of adversarial vulnerability. Here, we propose to replace the classification loss with the joint optimization of invariance to identity-preserving transformations of images (data augmentation invariance), and the invariance to objects of the same category (class-wise invariance). We hypothesize that optimizing these invariance objectives might yield features more aligned with visual perception, more robust to adversarial perturbations, while still suitable for accurate object categorization.

Cite

Text

Hernández-García and König. "Learning Representational Invariance Instead of Categorization." IEEE/CVF International Conference on Computer Vision Workshops, 2019. doi:10.1109/ICCVW.2019.00561

Markdown

[Hernández-García and König. "Learning Representational Invariance Instead of Categorization." IEEE/CVF International Conference on Computer Vision Workshops, 2019.](https://mlanthology.org/iccvw/2019/hernandezgarcia2019iccvw-learning/) doi:10.1109/ICCVW.2019.00561

BibTeX

@inproceedings{hernandezgarcia2019iccvw-learning,
  title     = {{Learning Representational Invariance Instead of Categorization}},
  author    = {Hernández-García, Alex and König, Peter},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2019},
  pages     = {4587-4590},
  doi       = {10.1109/ICCVW.2019.00561},
  url       = {https://mlanthology.org/iccvw/2019/hernandezgarcia2019iccvw-learning/}
}