Even If Explanations: Prior Work, Desiderata & Benchmarks for Semi-Factual XAI

15Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, eXplainable AI (XAI) research has focused on counterfactual explanations as post-hoc justifications for AI-system decisions (e.g., a customer refused a loan might be told “if you asked for a loan with a shorter term, it would have been approved”). Counterfactuals explain what changes to the input-features of an AI system change the output-decision. However, there is a sub-type of counterfactual, semi-factuals, that have received less attention in AI (though the Cognitive Sciences have studied them more). This paper surveys semifactual explanation, summarising historical and recent work. It defines key desiderata for semifactual XAI, reporting benchmark tests of historical algorithms (as well as a novel, naïve method) to provide a solid basis for future developments.

Cite

CITATION STYLE

APA

Aryal, S., & Keane, M. T. (2023). Even If Explanations: Prior Work, Desiderata & Benchmarks for Semi-Factual XAI. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2023-August, pp. 6526–6535). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2023/732

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free