Desiderata for interpretability: Explaining decision tree predictions with counterfactuals

17Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

Explanations in machine learning come in many forms, but a consensus regarding their desired properties is still emerging. In our work we collect and organise these explainability desiderata and discuss how they can be used to systematically evaluate properties and quality of an explainable system using the case of class-contrastive counterfactual statements. This leads us to propose a novel method for explaining predictions of a decision tree with counterfactuals. We show that our model-specific approach exploits all the theoretical advantages of counterfactual explanations, hence improves decision tree interpretability by decoupling the quality of the interpretation from the depth and width of the tree.

Cite

CITATION STYLE

APA

Sokol, K., & Flach, P. (2019). Desiderata for interpretability: Explaining decision tree predictions with counterfactuals. In 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019 (pp. 10035–10036). AAAI Press. https://doi.org/10.1609/aaai.v33i01.330110035

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free