Challenges in Applying Explainability Methods to Improve the Fairness of NLP Models

16Citations
Citations of this article
71Readers
Mendeley users who have this article in their library.

Abstract

Motivations for methods in explainable artificial intelligence (XAI) often include detecting, quantifying and mitigating bias, and contributing to making machine learning models fairer. However, exactly how an XAI method can help in combating biases is often left unspecified. In this paper, we briefly review trends in explainability and fairness in NLP research, identify the current practices in which explainability methods are applied to detect and mitigate bias, and investigate the barriers preventing XAI methods from being used more widely in tackling fairness issues.

Cite

CITATION STYLE

APA

Balkır, E., Kiritchenko, S., Nejadgholi, I., & Fraser, K. C. (2022). Challenges in Applying Explainability Methods to Improve the Fairness of NLP Models. In TrustNLP 2022 - 2nd Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop (pp. 80–92). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.trustnlp-1.8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free