exBERT: Extending pre-trained models with domain-specific vocabulary under constrained training resources

64Citations
Citations of this article
146Readers
Mendeley users who have this article in their library.

Abstract

We introduce exBERT, a training method to extend BERT pre-trained models from a general domain to a new pre-trained model for a specific domain with a new additive vocabulary under constrained training resources (i.e., constrained computation and data). exBERT uses a small extension module to learn to adapt an augmenting embedding for the new domain in the context of the original BERT’s embedding of a general vocabulary. The exBERT training method is novel in learning the new vocabulary and the extension module while keeping the weights of the original BERT model fixed, resulting in a substantial reduction in required training resources. We pre-train exBERT with biomedical articles from ClinicalKey and PubMed Central, and study its performance on biomedical downstream benchmark tasks using the MTLBioinformatics-2016 dataset. We demonstrate that exBERT consistently outperforms prior approaches when using limited corpus and pretraining computation resources.

Cite

CITATION STYLE

APA

Tai, W., Kung, H. T., Dong, X., Comiter, M., & Kuo, C. F. (2020). exBERT: Extending pre-trained models with domain-specific vocabulary under constrained training resources. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 1433–1439). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.129

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free