What is Missing in XAI So Far?: An Interdisciplinary Perspective

13Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.
Get full text

Abstract

With the perspective on applications of AI-technology, especially data intensive deep learning approaches, the need for methods to control and understand such models has been recognized and gave rise to a new research domain labeled explainable artificial intelligence (XAI). In this overview paper we give an interim appraisal of what has been achieved so far and where there are still gaps in the research. We take an interdisciplinary perspective to identify challenges on XAI research and point to open questions with respect to the quality of the explanations regarding faithfulness and consistency of explanations. On the other hand we see a need regarding the interaction between XAI and user to allow for adaptability to specific information needs and explanatory dialog for informed decision making as well as the possibility to correct models and explanations by interaction. This endeavor requires an integrated interdisciplinary perspective and rigorous approaches to empirical evaluation based on psychological, linguistic and even sociological theories.

Cite

CITATION STYLE

APA

Schmid, U., & Wrede, B. (2022). What is Missing in XAI So Far?: An Interdisciplinary Perspective. KI - Kunstliche Intelligenz, 36(3–4), 303–315. https://doi.org/10.1007/s13218-022-00786-2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free