Intriguing Properties of Generative Classifiers
Abstract
What is the best paradigm to recognize objects---discriminative inference (fast but potentially prone to shortcut learning) or using a generative model (slow but potentially more robust)? We build on recent advances in generative modeling that turn text-to-image models into classifiers. This allows us to study their behavior and to compare them against discriminative models and human psychophysical data. We report four intriguing emergent properties of generative classifiers: they show a record-breaking human-like shape bias (99% for Imagen), near human-level out-of-distribution accuracy, state-of-the-art alignment with human classification errors, and they understand certain perceptual illusions. Our results indicate that while the current dominant paradigm for modeling human object recognition is discriminative inference, zero-shot generative models approximate human object recognition data surprisingly well.
Cite
Text
Jaini et al. "Intriguing Properties of Generative Classifiers." International Conference on Learning Representations, 2024.Markdown
[Jaini et al. "Intriguing Properties of Generative Classifiers." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/jaini2024iclr-intriguing/)BibTeX
@inproceedings{jaini2024iclr-intriguing,
title = {{Intriguing Properties of Generative Classifiers}},
author = {Jaini, Priyank and Clark, Kevin and Geirhos, Robert},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/jaini2024iclr-intriguing/}
}