Integrating Prior Knowledge in Post-hoc Explanations

3Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In the field of explainable artificial intelligence (XAI), post-hoc interpretability methods aim at explaining to a user the predictions of a trained decision model. Integrating prior knowledge into such interpretability methods aims at improving the explanation understandability and allowing for personalised explanations adapted to each user. In this paper, we propose to define a cost function that explicitly integrates prior knowledge into the interpretability objectives: we present a general framework for the optimization problem of post-hoc interpretability methods, and show that user knowledge can thus be integrated to any method by adding a compatibility term in the cost function. We instantiate the proposed formalization in the case of counterfactual explanations and propose a new interpretability method called Knowledge Integration in Counterfactual Explanation (KICE) to optimize it. The paper performs an experimental study on several benchmark data sets to characterize the counterfactual instances generated by KICE, as compared to reference methods.

Cite

CITATION STYLE

APA

Jeyasothy, A., Laugel, T., Lesot, M. J., Marsala, C., & Detyniecki, M. (2022). Integrating Prior Knowledge in Post-hoc Explanations. In Communications in Computer and Information Science (Vol. 1602 CCIS, pp. 707–719). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-08974-9_56

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free