SMACE: A New Method for the Interpretability of Composite Decision Systems

0Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Interpretability is a pressing issue for decision systems. Many post hoc methods have been proposed to explain the predictions of a single machine learning model. However, business processes and decision systems are rarely centered around a unique model. These systems combine multiple models that produce key predictions, and then apply rules to generate the final decision. To explain such decisions, we propose the Semi-Model-Agnostic Contextual Explainer (SMACE), a new interpretability method that combines a geometric approach for decision rules with existing interpretability methods for machine learning models to generate an intuitive feature ranking tailored to the end user. We show that established model-agnostic approaches produce poor results on tabular data in this setting, in particular giving the same importance to several features, whereas SMACE can rank them in a meaningful way.

Cite

CITATION STYLE

APA

Lopardo, G., Garreau, D., Precioso, F., & Ottosson, G. (2023). SMACE: A New Method for the Interpretability of Composite Decision Systems. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13713 LNAI, pp. 325–339). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-26387-3_20

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free