Enriching Abusive Language Detection with Community Context

3Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Uses of pejorative expressions can be benign or actively empowering. When models for abuse detection misclassify these expressions as derogatory, they inadvertently censor productive conversations held by marginalized groups. One way to engage with non-dominant perspectives is to add context around conversations. Previous research has leveraged user- and thread-level features, but it often neglects the spaces within which productive conversations take place. Our paper highlights how community context can improve classification outcomes in abusive language detection. We make two main contributions to this end. First, we demonstrate that online communities cluster by the nature of their support towards victims of abuse. Second, we establish how community context improves accuracy and reduces the false positive rates of state-of-the-art abusive language classifiers. These findings suggest a promising direction for context-aware models in abusive language research.

Cite

CITATION STYLE

APA

Kurrek, J., Saleem, H. M., & Ruths, D. (2022). Enriching Abusive Language Detection with Community Context. In WOAH 2022 - 6th Workshop on Online Abuse and Harms, Proceedings of the Workshop (pp. 131–142). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.woah-1.13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free