Transfer learning for classifying Spanish and english text by clinical specialties

2Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

Transfer learning has demonstrated its potential in natural language processing tasks, where models have been pre-trained on large corpora and then tuned to specific tasks. We applied pre-trained transfer models to a Spanish biomedical document classification task. The main goal is to analyze the performance of text classification by clinical specialties using state-of-the-art language models for Spanish, and compared them with the results using corresponding models in English and with the most important pre-trained model for the biomedical domain. The outcomes present interesting perspectives on the performance of language models that are pre-trained for a particular domain. In particular, we found that BioBERT achieved better results on Spanish texts translated into English than the general domain model in Spanish and the state-of-the-art multilingual model. © 2021 European Federation for Medical Informatics (EFMI) and IOS Press.

Cite

CITATION STYLE

APA

Pomares-Quimbaya, A., López-Úbeda, P., & Schulz, S. (2021). Transfer learning for classifying Spanish and english text by clinical specialties. In Public Health and Informatics: Proceedings of MIE 2021 (pp. 377–381). IOS Press. https://doi.org/10.3233/SHTI210184

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free