A Training-Free Debiasing Framework with Counterfactual Reasoning for Conversational Emotion Detection

9Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Unintended dataset biases typically exist in existing Emotion Recognition in Conversations (ERC) datasets, including label bias, where models favor the majority class due to imbalanced training data, as well as the speaker and neutral word bias, where models make unfair predictions because of excessive correlations between specific neutral words or speakers and classes. However, previous studies in ERC generally focus on capturing context-sensitive and speaker-sensitive dependencies, ignoring the unintended dataset biases of data, which hampers the generalization and fairness in ERC. To address this issue, we propose a Training-Free Debiasing framework (TFD) that operates during prediction without additional training. To ensure compatibility with various ERC models, it does not balance data or modify the model structure. Instead, TFD extracts biases from the model by generating counterfactual utterances and contexts and mitigates them using simple yet empirically robust element-wise subtraction operations. Extensive experiments on three public datasets demonstrate that TFD effectively improves generalization ability and fairness across different ERC models.

Cite

CITATION STYLE

APA

Tu, G., Jing, R., Liang, B., Yang, M., Wong, K. F., & Xu, R. (2023). A Training-Free Debiasing Framework with Counterfactual Reasoning for Conversational Emotion Detection. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 15639–15650). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.967

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free