VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding

11Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.

Cite

CITATION STYLE

APA

Hu, D., Hou, X., Du, X., Zhou, M., Jiang, L., Mo, Y., & Shi, X. (2022). VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 6305–6315). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.489

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free