QuTI! Quantifying Text-Image Consistency in Multimodal Documents

8Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The World Wide Web and social media platforms have become popular sources for news and information. Typically, multimodal information, e.g., image and text is used to convey information more effectively and to attract attention. While in most cases image content is decorative or depicts additional information, it has also been leveraged to spread misinformation and rumors in recent years. In this paper, we present a web-based demo application that automatically quantifies the cross-modal relations of entities∼(persons, locations, and events) in image and text. The applications are manifold. For example, the system can help users to explore multimodal articles more efficiently, or can assist human assessors and fact-checking efforts in the verification of the credibility of news stories, tweets, or other multimodal documents.

Cite

CITATION STYLE

APA

Springstein, M., Müller-Budack, E., & Ewerth, R. (2021). QuTI! Quantifying Text-Image Consistency in Multimodal Documents. In SIGIR 2021 - Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 2575–2579). Association for Computing Machinery, Inc. https://doi.org/10.1145/3404835.3462796

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free