Bias in Hate Speech and Toxicity Detection

2Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Many Artificial Intelligence (AI) systems rely on finding patterns in large datasets, which are prone to bias and exacerbate existing segregation and inequalities of marginalised communities. Due to their socio-Technical impact, bias in AI has become a pressing issue. In this work, we investigate discrimination prevention methods on the assumption that disparities of specific populations in the training samples are reproduced or even amplified in the AI system outcomes. We aim to identify the information from vulnerable groups in the training data, uncover potential inequalities in how data capture these groups and provide additional information about them to alleviate inequalities, e.g., stereotypical and generalised views that lead to learning discriminatory associations. We develop data preprocessing techniques in automated moderation (AI systems to flag or filter online abuse) due to its substantial social implications and existing challenges common to many AI applications.

Cite

CITATION STYLE

APA

Lobo, P. R. (2022). Bias in Hate Speech and Toxicity Detection. In AIES 2022 - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society (p. 910). Association for Computing Machinery, Inc. https://doi.org/10.1145/3514094.3539519

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free