Learning Model-Agnostic Counterfactual Explanations for Tabular Data

123Citations
Citations of this article
72Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Counterfactual explanations can be obtained by identifying the smallest change made to an input vector to influence a prediction in a positive way from a user's viewpoint; for example, from 'loan rejected' to 'awarded' or from 'high risk of cardiovascular disease' to 'low risk'. Previous approaches would not ensure that the produced counterfactuals be proximate (i.e., not local outliers) and connected to regions with substantial data density (i.e., close to correctly classified observations), two requirements known as counterfactual faithfulness. Our contribution is twofold. First, drawing ideas from the manifold learning literature, we develop a framework, called C-CHVAE, that generates faithful counterfactuals. Second, we suggest to complement the catalog of counterfactual quality measures using a criterion to quantify the degree of difficulty for a certain counterfactual suggestion. Our real world experiments suggest that faithful counterfactuals come at the cost of higher degrees of difficulty.

Cite

CITATION STYLE

APA

Pawelczyk, M., Broelemann, K., & Kasneci, G. (2020). Learning Model-Agnostic Counterfactual Explanations for Tabular Data. In The Web Conference 2020 - Proceedings of the World Wide Web Conference, WWW 2020 (pp. 3126–3132). Association for Computing Machinery, Inc. https://doi.org/10.1145/3366423.3380087

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free