Improving the Transferability of Clinical Note Section Classification Models with BERT and Large Language Model Ensembles

5Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Text in electronic health records is organized into sections, and classifying those sections into section categories is useful for downstream tasks. In this work, we attempt to improve the transferability of section classification models by combining the dataset-specific knowledge in supervised learning models with the world knowledge inside large language models (LLMs). Surprisingly, we find that zero-shot LLMs out-perform supervised BERT-based models applied to out-of-domain data. We also find that their strengths are synergistic, so that a simple ensemble technique leads to additional performance gains.

Cite

CITATION STYLE

APA

Zhou, W., Afshar, M., Gao, Y., Dligach, D., & Miller, T. A. (2023). Improving the Transferability of Clinical Note Section Classification Models with BERT and Large Language Model Ensembles. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 125–130). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.clinicalnlp-1.16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free