Predicting patients’ future outcomes by analyzing Electronic health records (EHRs) is a hot topic in machine learning. The key challenge in this area is how to transform high dimensional, redundant, and heterogeneous EHRs into appropriate representations. In this paper, we argue for four desired properties of ideal patient representation learning, which are completeness, cross-modality invariance, anti-nuisance, and personality maintenance. To obtain such properties, We propose a Supervised Deep Patient Representation Learning Framework (SDPRL) to learn patients’ representations that incorporates complete semantics of health conditions by using multi-modal EHR data. Furthermore, we propose to maximize the mutual information (MI) among each pair of different modal representations, as well as minimizing the task-specific loss function. This not only keeps the task-relevant semantic information into the learned representations, but also makes the resulting representations to be relatively invariant across the modalities, anti-nuisance, and maintain the personality. With experiments conducted on the publicly available MIMIC-III dataset on the mortality prediction and forecasting the length of stay (LOS) tasks, we empirically demonstrate that the proposed SDPRL achieves higher prediction performance than baseline frameworks. Moreover, we demonstrate that SDPRL can yield the desired properties we argued. It can well-handle the modal-missing issue in the test phase, as well as getting advance to the goal of personalized medicine.
CITATION STYLE
Zhang, X., Qian, B., Li, Y., Liu, Y., Chen, X., Guan, C., & Li, C. (2021). Learning robust patient representations from multi-modal electronic health records: A supervised deep learning approach. In SIAM International Conference on Data Mining, SDM 2021 (pp. 585–593). Siam Society. https://doi.org/10.1137/1.9781611976700.66
Mendeley helps you to discover research relevant for your work.