XLM-K: Improving Cross-Lingual Language Model Pre-training with Multilingual Knowledge

24Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.

Abstract

Cross-lingual pre-training has achieved great successes using monolingual and bilingual plain text corpora. However, most pre-trained models neglect multilingual knowledge, which is language agnostic but comprises abundant crosslingual structure alignment. In this paper, we propose XLMK, a cross-lingual language model incorporating multilingual knowledge in pre-training. XLM-K augments existing multilingual pre-training with two knowledge tasks, namely Masked Entity Prediction Task and Object Entailment Task. We evaluate XLM-K on MLQA, NER and XNLI. Experimental results clearly demonstrate significant improvements over existing multilingual language models. The results on MLQA and NER exhibit the superiority of XLM-K in knowledge related tasks. The success in XNLI shows a better crosslingual transferability obtained in XLM-K. What is more, we provide a detailed probing analysis to confirm the desired knowledge captured in our pre-training regimen. The code is available at https://github.com/microsoft/Unicoder/ tree/master/pretraining/xlmk.

Cite

CITATION STYLE

APA

Jiang, X., Liang, Y., Chen, W., & Duan, N. (2022). XLM-K: Improving Cross-Lingual Language Model Pre-training with Multilingual Knowledge. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 10840–10848). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i10.21330

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free