AMBIFC: Fact-Checking Ambiguous Claims with Evidence

7Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

Automated fact-checking systems verify claims against evidence to predict their verac-ity. In real-world scenarios, the retrieved evidence may not unambiguously support or refute the claim and yield conflicting but valid interpretations. Existing fact-checking data-sets assume that the models developed with them predict a single veracity label for each claim, thus discouraging the handling of such ambiguity. To address this issue we present AMBIFC,1 a fact-checking dataset with 10k claims derived from real-world information needs. It contains fine-grained evidence annotations of 50k passages from 5k Wikipedia pages. We analyze the disagreements aris-ing from ambiguity when comparing claims against evidence in AMBIFC, observing a strong correlation of annotator disagreement with linguistic phenomena such as underspec-ification and probabilistic reasoning. We de-velop models for predicting veracity handling this ambiguity via soft labels, and find that a pipeline that learns the label distribution for sentence-level evidence selection and veracity prediction yields the best performance. We compare models trained on different subsets of AMBIFC and show that models trained on the ambiguous instances perform better when faced with the identified linguistic phenomena.

Cite

CITATION STYLE

APA

Glockner, M., Staliūnaitė, I., Thorne, J., Vallejo, G., Vlachos, A., & Gurevych, I. (2024). AMBIFC: Fact-Checking Ambiguous Claims with Evidence. Transactions of the Association for Computational Linguistics, 12, 1–18. https://doi.org/10.1162/tacl_a_00629

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free