Achieving Diversity in Counterfactual Explanations: a Review and Discussion

11Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In the field of Explainable Artificial Intelligence (XAI), counterfactual examples explain to a user the predictions of a trained decision model by indicating the modifications to be made to the instance so as to change its associated prediction. These counterfactual examples are generally defined as solutions to an optimization problem whose cost function combines several criteria that quantify desiderata for a good explanation meeting user needs. A large variety of such appropriate properties can be considered, as the user needs are generally unknown and differ from one user to another; their selection and formalization is difficult. To circumvent this issue, several approaches propose to generate, rather than a single one, a set of diverse counterfactual examples to explain a prediction. This paper proposes a review of the numerous, sometimes conflicting, definitions that have been proposed for this notion of diversity. It discusses their underlying principles as well as the hypotheses on the user needs they rely on and proposes to categorize them along several dimensions (explicit vs implicit, universe in which they are defined, level at which they apply), leading to the identification of further research challenges on this topic.

Cite

CITATION STYLE

APA

Laugel, T., Jeyasothy, A., Lesot, M. J., Marsala, C., & Detyniecki, M. (2023). Achieving Diversity in Counterfactual Explanations: a Review and Discussion. In ACM International Conference Proceeding Series (pp. 1859–1869). Association for Computing Machinery. https://doi.org/10.1145/3593013.3594122

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free