The Prediction-Explanation Fallacy: A Pervasive Problem in ScientificApplications of Machine Learning

2Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

I highlight a problem that has become ubiquitous in scientificapplications of machine learning and can lead to seriously distorted inferences. I call it the Prediction-Explanation Fallacy. The fallacy occurs when researchers use prediction-optimized models for explanatory purposes, without considering the relevant tradeoffs. This is a problem for at least two reasons. First, prediction-optimized models are often deliberately biased and unrealistic in order to prevent overfitting.In other cases, they have an exceedingly complex structure that is hard or impossible to interpret. Second, different predictive models trained on the same or similar data can be biased in different ways, so that they may predict equally well but suggest conflictingexplanations. Here I introduce the tradeoffs between prediction and explanation in a non-technical fashion, present illustrative examples from neuroscience, and end by discussing some mitigating factors and methods that can be used to limit the problem.

Cite

CITATION STYLE

APA

Del Giudice, M. (2024). The Prediction-Explanation Fallacy: A Pervasive Problem in ScientificApplications of Machine Learning. Methodology, 20(1), 22–46. https://doi.org/10.5964/meth.11235

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free