Leveraging Community and Author Context to Explain the Performance and Bias of Text-Based Deception Detection Models

0Citations
Citations of this article
50Readers
Mendeley users who have this article in their library.

Abstract

Deceptive news posts shared in online communities can be detected with NLP models, and much recent research has focused on the development of such models. In this work, we use characteristics of online communities and authors — the context of how and where content is posted — to explain the performance of a neural network deception detection model and identify sub-populations who are disproportionately affected by model accuracy or failure. We examine who is posting the content, and where the content is posted to. We find that while author characteristics are better predictors of deceptive content than community characteristics, both characteristics are strongly correlated with model performance. Traditional performance metrics such as F1 score may fail to capture poor model performance on isolated sub-populations such as specific authors, and as such, more nuanced evaluation of deception detection models is critical.

Cite

CITATION STYLE

APA

Weld, G., Ayton, E., Althoff, T., & Glenski, M. (2021). Leveraging Community and Author Context to Explain the Performance and Bias of Text-Based Deception Detection Models. In NLP4IF 2021 - NLP for Internet Freedom: Censorship, Disinformation, and Propaganda, Proceedings of the 4th Workshop (pp. 29–38). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.nlp4if-1.5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free