Feature importance is an approach that helps to explain machine learning model predictions. It works through assigning importance scores to input features of a particular model. Different techniques exist to derive these scores, with widely varying underlying assumptions of what importance means. Little research has been done to verify whether these assumptions match the expectations of the target user, which is imperative to ensure that feature importance values are not misinterpreted. In this work, we explore data scientists' mental models of (local) feature importance and compare these with the conceptual models of the techniques. We first identify several properties of local feature importance techniques that could potentially lead to misinterpretations. Subsequently, we explore the expectations data scientists have about local feature importance through an exploratory (qualitative and quantitative) survey of 34 data scientists in industry. We compare the identified expectations to the theory and assumptions behind the techniques and find that the two are not (always) in agreement.
CITATION STYLE
Collaris, D., Weerts, H. J. P., Miedema, D., Van Wijk, J. J., & Pechenizkiy, M. (2022). Characterizing Data Scientists’ Mental Models of Local Feature Importance. In ACM International Conference Proceeding Series. Association for Computing Machinery. https://doi.org/10.1145/3546155.3546670
Mendeley helps you to discover research relevant for your work.