Balancing out Bias: Achieving Fairness Through Balanced Training

17Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.

Abstract

Group bias in natural language processing tasks manifests as disparities in system error rates across texts authorized by different demographic groups, typically disadvantaging minority groups. Dataset balancing has been shown to be effective at mitigating bias, however existing approaches do not directly account for correlations between author demographics and linguistic variables, limiting their effectiveness. To achieve Equal Opportunity fairness, such as equal job opportunity without regard to demographics, this paper introduces a simple, but highly effective, objective for countering bias using balanced training. We extend the method in the form of a gated model, which incorporates protected attributes as input, and show that it is effective at reducing bias in predictions through demographic input perturbation, outperforming all other bias mitigation techniques when combined with balanced training.

References Powered by Scopus

Survey of multi-objective optimization methods for engineering

3953Citations
N/AReaders
Get full text

Fairness through awareness

2449Citations
N/AReaders
Get full text

Class-balanced loss based on effective number of samples

1972Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Bias and Fairness in Large Language Models: A Survey

71Citations
N/AReaders
Get full text

FairLib: A Unified Framework for Assessing and Improving Fairness

8Citations
N/AReaders
Get full text

FAIRFLOW: Mitigating Dataset Biases through Undecided Learning for Natural Language Understanding

1Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Han, X., Baldwin, T., & Cohn, T. (2022). Balancing out Bias: Achieving Fairness Through Balanced Training. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 11335–11350). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.779

Readers over time

‘21‘22‘23‘24‘2505101520

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 7

58%

Researcher 3

25%

Professor / Associate Prof. 1

8%

Lecturer / Post doc 1

8%

Readers' Discipline

Tooltip

Computer Science 13

81%

Medicine and Dentistry 1

6%

Linguistics 1

6%

Neuroscience 1

6%

Save time finding and organizing research with Mendeley

Sign up for free
0