On the Ethical and Epistemological Utility of Explicable AI in Medicine

13Citations
Citations of this article
33Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

In this article, I will argue in favor of both the ethical and epistemological utility of explanations in artificial intelligence (AI)-based medical technology. I will build on the notion of “explicability” due to Floridi, which considers both the intelligibility and accountability of AI systems to be important for truly delivering AI-powered services that strengthen autonomy, beneficence, and fairness. I maintain that explicable algorithms do, in fact, strengthen these ethical principles in medicine, e.g., in terms of direct patient–physician contact, as well as on a longer-term epistemological level by facilitating scientific progress that is informed through practice. With this article, I will therefore attempt to counter arguments against demands for explicable AI in medicine that are based on a notion of “whatever heals is right.” I will elucidate my elaboration on the positive aspects of explicable AI in medicine as well as by pointing out risks of non-explicable AI.

Cite

CITATION STYLE

APA

Herzog, C. (2022). On the Ethical and Epistemological Utility of Explicable AI in Medicine. Philosophy and Technology, 35(2). https://doi.org/10.1007/s13347-022-00546-y

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free