Stepmothers are mean and academics are pretentious: What do pretrained language models learn about you?

17Citations
Citations of this article
70Readers
Mendeley users who have this article in their library.

Abstract

Warning: this paper contains content that may be offensive or upsetting. In this paper, we investigate what types of stereotypical information are captured by pretrained language models. We present the first dataset comprising stereotypical attributes of a range of social groups and propose a method to elicit stereotypes encoded by pretrained language models in an unsupervised fashion. Moreover, we link the emergent stereotypes to their manifestation as basic emotions as a means to study their emotional effects in a more generalized manner. To demonstrate how our methods can be used to analyze emotion and stereotype shifts due to linguistic experience, we use fine-tuning on news sources as a case study. Our experiments expose how attitudes towards different social groups vary across models and how quickly emotions and stereotypes can shift at the fine-tuning stage.

Cite

CITATION STYLE

APA

Choenni, R., Shutova, E., & van Rooij, R. (2021). Stepmothers are mean and academics are pretentious: What do pretrained language models learn about you? In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1477–1491). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.111

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free