Characterizing Data Scientists' Mental Models of Local Feature Importance

1Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Feature importance is an approach that helps to explain machine learning model predictions. It works through assigning importance scores to input features of a particular model. Different techniques exist to derive these scores, with widely varying underlying assumptions of what importance means. Little research has been done to verify whether these assumptions match the expectations of the target user, which is imperative to ensure that feature importance values are not misinterpreted. In this work, we explore data scientists' mental models of (local) feature importance and compare these with the conceptual models of the techniques. We first identify several properties of local feature importance techniques that could potentially lead to misinterpretations. Subsequently, we explore the expectations data scientists have about local feature importance through an exploratory (qualitative and quantitative) survey of 34 data scientists in industry. We compare the identified expectations to the theory and assumptions behind the techniques and find that the two are not (always) in agreement.

Cite

CITATION STYLE

APA

Collaris, D., Weerts, H. J. P., Miedema, D., Van Wijk, J. J., & Pechenizkiy, M. (2022). Characterizing Data Scientists’ Mental Models of Local Feature Importance. In ACM International Conference Proceeding Series. Association for Computing Machinery. https://doi.org/10.1145/3546155.3546670

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free