Measuring Fairness with Biased Rulers: A Comparative Study on Bias Metrics for Pre-trained Language Models

60Citations
Citations of this article
64Readers
Mendeley users who have this article in their library.

Abstract

An increasing awareness of biased patterns in natural language processing resources such as BERT has motivated many metrics to quantify 'bias' and 'fairness' in these resources. However, comparing the results of different metrics and the works that evaluate with such metrics remains difficult, if not outright impossible. We survey the literature on fairness metrics for pre-trained language models and experimentally evaluate compatibility, including both biases in language models and in their downstream tasks. We do this by combining traditional literature survey, correlation analysis and empirical evaluations. We find that many metrics are not compatible with each other and highly depend on (i) templates, (ii) attribute and target seeds and (iii) the choice of embeddings. We also see no tangible evidence of intrinsic bias relating to extrinsic bias. These results indicate that fairness or bias evaluation remains challenging for contextualized language models, among other reasons because these choices remain subjective. To improve future comparisons and fairness evaluations, we recommend to avoid embedding-based metrics and focus on fairness evaluations in downstream tasks.

References Powered by Scopus

GloVe: Global vectors for word representation

26918Citations
N/AReaders
Get full text

Measuring individual differences in implicit cognition: The implicit association test

8410Citations
N/AReaders
Get full text

On the dangers of stochastic parrots: Can language models be too big?

2957Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Evaluating Large Language Models in Generating Synthetic HCI Research Data: a Case Study

112Citations
N/AReaders
Get full text

On Second Thought, Let's Not Think Step by Step! Bias and Toxicity in Zero-Shot Reasoning

55Citations
N/AReaders
Get full text

Bias and Fairness in Large Language Models: A Survey

52Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Delobelle, P., Tokpo, E. K., Calders, T., & Berendt, B. (2022). Measuring Fairness with Biased Rulers: A Comparative Study on Bias Metrics for Pre-trained Language Models. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1693–1706). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.122

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 14

67%

Researcher 6

29%

Lecturer / Post doc 1

5%

Readers' Discipline

Tooltip

Computer Science 18

78%

Medicine and Dentistry 2

9%

Linguistics 2

9%

Neuroscience 1

4%

Save time finding and organizing research with Mendeley

Sign up for free