Word Embeddings via Causal Inference: Gender Bias Reducing and Semantic Information Preserving

28Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.

Abstract

With widening deployments of natural language processing (NLP) in daily life, inherited social biases from NLP models have become more severe and problematic. Previous studies have shown that word embeddings trained on human-generated corpora have strong gender biases that can produce discriminative results in downstream tasks. Previous debiasing methods focus mainly on modeling bias and only implicitly consider semantic information while completely overlooking the complex underlying causal structure among bias and semantic components. To address these issues, we propose a novel methodology that leverages a causal inference framework to effectively remove gender bias. The proposed method allows us to construct and analyze the complex causal mechanisms facilitating gender information flow while retaining oracle semantic information within word embeddings. Our comprehensive experiments show that the proposed method achieves state-of-the-art results in gender-debiasing tasks. In addition, our methods yield better performance in word similarity evaluation and various extrinsic downstream NLP tasks.

Cite

CITATION STYLE

APA

Ding, L., Yu, D., Xie, J., Guo, W., Hu, S., Liu, M., … Jiang, B. (2022). Word Embeddings via Causal Inference: Gender Bias Reducing and Semantic Information Preserving. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 11864–11872). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i11.21443

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free