Existing methods for debiasing word embeddings often do so only superficially, in that words that are stereotypically associated with, e.g., a particular gender in the original embedding space can still be clustered together in the debiased space. However, there has yet to be a study that explores why this residual clustering exists, and how it might be addressed. The present work fills this gap. We identify two potential reasons for which residual bias exists and develop a new pipeline, MDR Cluster-Debias, to mitigate this bias. We explore the strengths and weaknesses of our method, finding that it significantly outperforms other existing debiasing approaches on a variety of upstream bias tests but achieves limited improvement on decreasing gender bias in a downstream task. This indicates that word embeddings encode gender bias in still other ways, not necessarily captured by upstream tests.
CITATION STYLE
Du, Y., & Joseph, K. (2020). MDR Cluster-Debias: A Nonlinear Word Embedding Debiasing Pipeline. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12268 LNCS, pp. 45–54). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-61255-9_5
Mendeley helps you to discover research relevant for your work.