Vision-language models (VLMs) have shown remarkable performance on visual reasoning tasks (e.g. attributes, location). While such tasks measure the requisite knowledge to ground and reason over a given visual instance, they do not, however, measure the ability of VLMs to retain and generalize such knowledge. In this work, we evaluate VLMs' ability to acquire “visible” physical knowledge - the information that is easily accessible from images of static scenes, particularly along the dimensions of object color, size, and space. We build an automatic pipeline to derive a comprehensive knowledge resource for calibrating and probing these models. Our results indicate a severe gap between model and human performance across all three dimensions. Furthermore, we demonstrate that an LM tuned on the captions significantly outperforms VLMs on both size and spatial tasks - highlighting that despite sufficient access to ground language with visual modality, they struggle to retain such knowledge. The dataset and code are available at https://github.com/luka-group/ViPhy.
CITATION STYLE
Singh, S., Qasemi, E., & Chen, M. (2023). VIPHY: Probing “Visible” Physical Commonsense Knowledge. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 7113–7128). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.473
Mendeley helps you to discover research relevant for your work.