Understanding Visual Concepts Across Models
Abstract
Large multimodal models such as Stable Diffusion can generate, detect, and classify new visual concepts after fine-tuning just a single word embedding. Do models learn similar words for the same concepts (i.e. \texttt{<orange-cat>} = orange + cat)? We conduct a large-scale analysis on three state-of-the-art models in text-to-image generation, open-set object detection, and zero-shot classification, and find that new word embeddings are model-specific and non-transferable. Across 4,800 new embeddings trained for 40 diverse visual concepts on four standard datasets, we find perturbations within an $\epsilon$-ball to any prior embedding that generate, detect, and classify an arbitrary concept. When these new embeddings are spliced into new models, fine-tuning that targets the original model is lost. We show popular soft prompt-tuning approaches find these perturbative solutions when applied to visual concept learning tasks, and embeddings for visual concepts are not transferable. Code for reproducing our work is available at: \href{https://anonymous-visual-words.github.io}anonymous-visual-words.github.io.
Cite
Text
Trabucco et al. "Understanding Visual Concepts Across Models." NeurIPS 2024 Workshops: AFM, 2024.Markdown
[Trabucco et al. "Understanding Visual Concepts Across Models." NeurIPS 2024 Workshops: AFM, 2024.](https://mlanthology.org/neuripsw/2024/trabucco2024neuripsw-understanding/)BibTeX
@inproceedings{trabucco2024neuripsw-understanding,
title = {{Understanding Visual Concepts Across Models}},
author = {Trabucco, Brandon and Gurinas, Max A and Doherty, Kyle and Salakhutdinov, Russ},
booktitle = {NeurIPS 2024 Workshops: AFM},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/trabucco2024neuripsw-understanding/}
}