Systematic Review of Approaches to Preserve Machine Learning Performance in the Presence of Temporal Dataset Shift in Clinical Medicine

34Citations
Citations of this article
37Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Objective The change in performance of machine learning models over time as a result of temporal dataset shift is a barrier to machine learning-derived models facilitating decision-making in clinical practice. Our aim was to describe technical procedures used to preserve the performance of machine learning models in the presence of temporal dataset shifts. Methods Studies were included if they were fully published articles that used machine learning and implemented a procedure to mitigate the effects of temporal dataset shift in a clinical setting. We described how dataset shift was measured, the procedures used to preserve model performance, and their effects. Results Of 4,457 potentially relevant publications identified, 15 were included. The impact of temporal dataset shift was primarily quantified using changes, usually deterioration, in calibration or discrimination. Calibration deterioration was more common (n = 11) than discrimination deterioration (n = 3). Mitigation strategies were categorized as model level or feature level. Model-level approaches (n = 15) were more common than feature-level approaches (n = 2), with the most common approaches being model refitting (n = 12), probability calibration (n = 7), model updating (n = 6), and model selection (n = 6). In general, all mitigation strategies were successful at preserving calibration but not uniformly successful in preserving discrimination. Conclusion There was limited research in preserving the performance of machine learning models in the presence of temporal dataset shift in clinical medicine. Future research could focus on the impact of dataset shift on clinical decision making, benchmark the mitigation strategies on a wider range of datasets and tasks, and identify optimal strategies for specific settings.

References Powered by Scopus

Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement

17664Citations
N/AReaders
Get full text

MIMIC-III, a freely accessible critical care database

5436Citations
N/AReaders
Get full text

The Unified Medical Language System (UMLS): Integrating biomedical terminology

3337Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Foundation models for generalist medical artificial intelligence

580Citations
N/AReaders
Get full text

Algorithmic fairness in artificial intelligence for medicine and healthcare

145Citations
N/AReaders
Get full text

Data drift in medical machine learning: implications and potential remedies

49Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Guo, L. L., Pfohl, S. R., Fries, J., Posada, J., Fleming, S. L., Aftandilian, C., … Sung, L. (2021, August 1). Systematic Review of Approaches to Preserve Machine Learning Performance in the Presence of Temporal Dataset Shift in Clinical Medicine. Applied Clinical Informatics. Georg Thieme Verlag. https://doi.org/10.1055/s-0041-1735184

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 7

58%

Researcher 3

25%

Professor / Associate Prof. 1

8%

Lecturer / Post doc 1

8%

Readers' Discipline

Tooltip

Computer Science 7

58%

Medicine and Dentistry 3

25%

Materials Science 1

8%

Pharmacology, Toxicology and Pharmaceut... 1

8%

Save time finding and organizing research with Mendeley

Sign up for free