Abstract
Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.
Cite
CITATION STYLE
Hu, D., Hou, X., Du, X., Zhou, M., Jiang, L., Mo, Y., & Shi, X. (2022). VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 6305–6315). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.489
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.