Unsupervised domain adaptation of a pretrained cross-lingual language model

20Citations
Citations of this article
49Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recent research indicates that pretraining cross-lingual language models on large-scale unlabeled texts yields significant performance improvements over various cross-lingual and low-resource tasks. Through training on one hundred languages and terabytes of texts, cross-lingual language models have proven to be effective in leveraging high-resource languages to enhance low-resource language processing and outperform monolingual models. In this paper, we further investigate the cross-lingual and cross-domain (CLCD) setting when a pretrained cross-lingual language model needs to adapt to new domains. Specifically, we propose a novel unsupervised feature decomposition method that can automatically extract domain-specific features and domain-invariant features from the entangled pretrained cross-lingual representations, given unlabeled raw texts in the source language. Our proposed model leverages mutual information estimation to decompose the representations computed by a cross-lingual model into domain-invariant and domain-specific parts. Experimental results show that our proposed method achieves significant performance improvements over the state-of-the-art pretrained cross-lingual language model in the CLCD setting.

Cite

CITATION STYLE

APA

Li, J., He, R., Ye, H., Ng, H. T., Bing, L., & Yan, R. (2020). Unsupervised domain adaptation of a pretrained cross-lingual language model. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 3672–3678). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/508

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free