Measuring Harmful Sentence Completion in Language Models for LGBTQIA+ Individuals

34Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

Abstract

Current language technology is ubiquitous and directly influences individuals' lives worldwide. Given the recent trend in AI on training and constantly releasing new and powerful large language models (LLMs), there is a need to assess their biases and potential concrete consequences. While some studies have highlighted the shortcomings of these models, there is only little on the negative impact of LLMs on LGBTQIA+ individuals. In this paper, we investigated a state-of-the-art template-based approach for measuring the harmfulness of English LLMs sentence completion when the subjects belong to the LGBTQIA+ community. Our findings show that, on average, the most likely LLM-generated completion is an identity attack 13% of the time. Our results raise serious concerns about the applicability of these models in production environments.

Cite

CITATION STYLE

APA

Nozza, D., Bianchi, F., Lauscher, A., & Hovy, D. (2022). Measuring Harmful Sentence Completion in Language Models for LGBTQIA+ Individuals. In LTEDI 2022 - 2nd Workshop on Language Technology for Equality, Diversity and Inclusion, Proceedings of the Workshop (pp. 26–34). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.ltedi-1.4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free