Zero-shot Cross-lingual Transfer With Learned Projections Using Unlabeled Target-Language Data

3Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Adapters have emerged as a parameter-efficient Transformer-based framework for cross-lingual transfer by inserting lightweight language-specific modules (language adapters) and task-specific modules (task adapters) within pretrained multilingual models. Zero-shot transfer is enabled by pairing the language adapter in the target language with an appropriate task adapter in a source language. If our target languages are known apriori, we explore how zero-shot transfer can be further improved within the adapter framework by utilizing unlabeled text during task-specific finetuning. We construct language-specific subspaces using standard linear algebra constructs and selectively project source-language representations into the target language subspace during task-specific finetuning using two schemes. Our experiments on three cross-lingual tasks, Named Entity Recognition (NER), Question Answering (QA) and Natural Language Inference (NLI) yield consistent benefits compared to adapter baselines over a wide variety of target languages with up to 11% relative improvement in NER, 2% relative improvement in QA and 5% relative improvement in NLI.

Cite

CITATION STYLE

APA

Deb, U., Parab, R., & Jyothi, P. (2023). Zero-shot Cross-lingual Transfer With Learned Projections Using Unlabeled Target-Language Data. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 449–457). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-short.39

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free