This paper considers the use of machine learning in medicine by focusing on the main problem that it has been aimed at solving or at least minimizing: uncertainty. However, we point out how uncertainty is so ingrained in medicine that it biases also the representation of clinical phenomena, that is the very input of this class of computational models, thus undermining the clinical significance of their output. Recognizing this can motivate researchers to pursue different ways to assess the value of these decision aids, as well as alternative techniques that do not “sweep uncertainty under the rug” within an objectivist fiction (which doctors can come up by trusting).
CITATION STYLE
Cabitza, F., Ciucci, D., & Rasoini, R. (2019). A giant with feet of clay: On the validity of the data that feed machine learning in medicine. In Lecture Notes in Information Systems and Organisation (Vol. 28, pp. 121–136). Springer Heidelberg. https://doi.org/10.1007/978-3-319-90503-7_10
Mendeley helps you to discover research relevant for your work.