Abstract
Transfer learning has demonstrated its potential in natural language processing tasks, where models have been pre-trained on large corpora and then tuned to specific tasks. We applied pre-trained transfer models to a Spanish biomedical document classification task. The main goal is to analyze the performance of text classification by clinical specialties using state-of-the-art language models for Spanish, and compared them with the results using corresponding models in English and with the most important pre-trained model for the biomedical domain. The outcomes present interesting perspectives on the performance of language models that are pre-trained for a particular domain. In particular, we found that BioBERT achieved better results on Spanish texts translated into English than the general domain model in Spanish and the state-of-the-art multilingual model. © 2021 European Federation for Medical Informatics (EFMI) and IOS Press.
Author supplied keywords
Cite
CITATION STYLE
Pomares-Quimbaya, A., López-Úbeda, P., & Schulz, S. (2021). Transfer learning for classifying Spanish and english text by clinical specialties. In Public Health and Informatics: Proceedings of MIE 2021 (pp. 377–381). IOS Press. https://doi.org/10.3233/SHTI210184
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.