Isotropic Representation Can Improve Zero-Shot Cross-Lingual Transfer on Multilingual Language Models

3Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

With the development of multilingual pretrained language models (mPLMs), zero-shot cross-lingual transfer shows great potential. To further improve the performance of cross-lingual transfer, many studies have explored representation misalignment caused by morphological differences but neglected the misalignment caused by the anisotropic distribution of contextual representations. In this work, we propose enhanced isotropy and constrained code-switching for zero-shot cross-lingual transfer to alleviate the problem of misalignment caused by the anisotropic representations and maintain syntactic structural knowledge. Extensive experiments on three zero-shot cross-lingual transfer tasks demonstrate that our method gains significant improvements over strong mPLM backbones and further improves the state-of-the-art methods.

Cite

CITATION STYLE

APA

Ji, Y., Wang, J., Li, J., Ye, H., & Zhang, M. (2023). Isotropic Representation Can Improve Zero-Shot Cross-Lingual Transfer on Multilingual Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 8104–8118). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.545

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free