In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic Segmentation
Abstract
We present Lazy Visual Grounding for open-vocabulary semantic segmentation, which decouples unsupervised object mask discovery from object grounding. Plenty of the previous art casts this task as pixel-to-text classification without object-level comprehension, leveraging the image-to-text classification capability of pretrained vision-and-language models. We argue that visual objects are distinguishable without the prior text information as segmentation is essentially a visual understanding task. Lazy visual grounding first discovers object masks covering an image with iterative Normalized cuts and then later assigns text on the discovered objects in a late interaction manner. Our model requires no additional training yet shows great performance on five public datasets: Pascal VOC, Pascal Context, COCO-object, COCO-stuff, and ADE 20K. Especially, the visually appealing segmentation results demonstrate the model capability to localize objects precisely.
Cite
Text
Kang and Cho. "In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic Segmentation." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-72940-9_9Markdown
[Kang and Cho. "In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic Segmentation." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/kang2024eccv-defense/) doi:10.1007/978-3-031-72940-9_9BibTeX
@inproceedings{kang2024eccv-defense,
title = {{In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic Segmentation}},
author = {Kang, Dahyun and Cho, Minsu},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2024},
doi = {10.1007/978-3-031-72940-9_9},
url = {https://mlanthology.org/eccv/2024/kang2024eccv-defense/}
}