WinoViz: Probing Visual Properties of Objects Under Different States
Abstract
Humans interpret visual aspects of objects based on contexts. For example, a banana appears brown when rotten and green when unripe. Previous studies focused on language models' grasp of typical object properties. We introduce WinoViz, a text-only dataset with 1,380 examples of probing language models' reasoning about diverse visual properties under different contexts. Our task demands pragmatic and visual knowledge reasoning. We also present multi-hop data, a more challenging version requiring multi-step reasoning chains. Experimental findings include: a) GPT-4 excels overall but struggles with multi-hop data. b) Large models perform well in pragmatic reasoning but struggle with visual knowledge reasoning. c) Vision-language models outperform language-only models.
Cite
Text
Jin et al. "WinoViz: Probing Visual Properties of Objects Under Different States." ICLR 2024 Workshops: SeT_LLM, 2024.Markdown
[Jin et al. "WinoViz: Probing Visual Properties of Objects Under Different States." ICLR 2024 Workshops: SeT_LLM, 2024.](https://mlanthology.org/iclrw/2024/jin2024iclrw-winoviz/)BibTeX
@inproceedings{jin2024iclrw-winoviz,
title = {{WinoViz: Probing Visual Properties of Objects Under Different States}},
author = {Jin, Woojeong and Srinivasan, Tejas and Thomason, Jesse and Ren, Xiang},
booktitle = {ICLR 2024 Workshops: SeT_LLM},
year = {2024},
url = {https://mlanthology.org/iclrw/2024/jin2024iclrw-winoviz/}
}