AI-assisted and explainable hate speech detection for social media moderators - A design science approach

25Citations
Citations of this article
95Readers
Mendeley users who have this article in their library.

Abstract

To date, the detection of hate speech is still primarily carried out by humans, yet there is great potential for combining human expertise with automated approaches. However, identified challenges include low levels of agreement between humans and machines due to the algorithms' missing expertise of, e.g., cultural, and social structures. In this work, a design science approach is used to derive design knowledge and develop an artifact, through which humans are integrated in the process of detecting and evaluating hate speech. For this purpose, explainable artificial intelligence (XAI) is utilized: the artifact will provide explanative information, why the deep learning model predicted whether a text contains hate. Results show that the instantiated design knowledge in form of a dashboard is perceived as valuable and that XAI features increase the perception of the artifact's usefulness, ease of use, trustworthiness as well as the intention to use it.

Cite

CITATION STYLE

APA

Bunde, E. (2021). AI-assisted and explainable hate speech detection for social media moderators - A design science approach. In Proceedings of the Annual Hawaii International Conference on System Sciences (Vol. 2020-January, pp. 1264–1273). IEEE Computer Society. https://doi.org/10.24251/hicss.2021.154

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free