Twitter-COMMs: Detecting Climate, COVID, and Military Multimodal Misinformation

19Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.

Abstract

Detecting out-of-context media, such as “miscaptioned” images on Twitter, is a relevant problem, especially in domains of high public significance. In this work we aim to develop defenses against such misinformation for the topics of Climate Change, COVID-19, and Military Vehicles. We first present a large-scale multimodal dataset with over 884k tweets relevant to these topics. Next, we propose a detection method, based on the state-of-the-art CLIP model, that leverages automatically generated hard image-text mismatches. While this approach works well on our automatically constructed out-of-context tweets, we aim to validate its usefulness on data representative of the real world. Thus, we test it on a set of human-generated fakes created by mimicking in-the-wild misinformation. We achieve an 11% detection improvement in a high precision regime over a strong baseline. Finally, we share insights about our best model design and analyze the challenges of this emerging threat.

Cite

CITATION STYLE

APA

Biamby, G., Luo, G., Darrell, T., & Rohrbach, A. (2022). Twitter-COMMs: Detecting Climate, COVID, and Military Multimodal Misinformation. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1530–1549). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.110

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free