Abstract
Human knowledge is collectively encoded in the roughly 6500 languages spoken around the world, but it is not distributed equally across languages. Hence, for information-seeking question answering (QA) systems to adequately serve speakers of all languages, they need to operate cross-lingually. In this work we investigate the capabilities of multilingually pretrained language models on cross-lingual QA. We find that explicitly aligning the representations across languages with a post-hoc finetuning step generally leads to improved performance. We additionally investigate the effect of data size as well as the language choice in this fine-tuning step, also releasing a dataset for evaluating cross-lingual QA systems.
Cite
CITATION STYLE
Faisal, F., & Anastasopoulos, A. (2021). Investigating Post-pretraining Representation Alignment for Cross-Lingual Question Answering. In Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021 (pp. 133–148). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.mrqa-1.14
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.