Pre-trained Language Model for Biomedical Question Answering

51Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The recent success of question answering systems is largely attributed to pre-trained language models. However, as language models are mostly pre-trained on general domain corpora such as Wikipedia, they often have difficulty in understanding biomedical questions. In this paper, we investigate the performance of BioBERT, a pre-trained biomedical language model, in answering biomedical questions including factoid, list, and yes/no type questions. BioBERT uses almost the same structure across various question types and achieved the best performance in the 7th BioASQ Challenge (Task 7b, Phase B). BioBERT pre-trained on SQuAD or SQuAD 2.0 easily outperformed previous state-of-the-art models. BioBERT obtains the best performance when it uses the appropriate pre-/post-processing strategies for questions, passages, and answers.

Cite

CITATION STYLE

APA

Yoon, W., Lee, J., Kim, D., Jeong, M., & Kang, J. (2020). Pre-trained Language Model for Biomedical Question Answering. In Communications in Computer and Information Science (Vol. 1168 CCIS, pp. 727–740). Springer. https://doi.org/10.1007/978-3-030-43887-6_64

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free