Generating Realistic Natural Language Counterfactuals

30Citations
Citations of this article
54Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Counterfactuals are a valuable means for understanding decisions made by ML systems. However, the counterfactuals generated by the methods currently available for natural language text are either unrealistic or introduce imperceptible changes. We propose CounterfactualGAN: a method that combines a conditional GAN and the embeddings of a pretrained BERT encoder to model-agnostically generate realistic natural language text counterfactuals for explaining regression and classification tasks. Experimental results show that our method produces perceptibly distinguishable counterfactuals, while outperforming four baseline methods on fidelity and human judgments of naturalness, across multiple datasets and multiple predictive models.

Cite

CITATION STYLE

APA

Robeer, M., Bex, F., & Feelders, A. (2021). Generating Realistic Natural Language Counterfactuals. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 3611–3625). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.306

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free