Investigating Post-pretraining Representation Alignment for Cross-Lingual Question Answering

2Citations
Citations of this article
50Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Human knowledge is collectively encoded in the roughly 6500 languages spoken around the world, but it is not distributed equally across languages. Hence, for information-seeking question answering (QA) systems to adequately serve speakers of all languages, they need to operate cross-lingually. In this work we investigate the capabilities of multilingually pretrained language models on cross-lingual QA. We find that explicitly aligning the representations across languages with a post-hoc finetuning step generally leads to improved performance. We additionally investigate the effect of data size as well as the language choice in this fine-tuning step, also releasing a dataset for evaluating cross-lingual QA systems.

Cite

CITATION STYLE

APA

Faisal, F., & Anastasopoulos, A. (2021). Investigating Post-pretraining Representation Alignment for Cross-Lingual Question Answering. In Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021 (pp. 133–148). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.mrqa-1.14

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free