Cross-lingual training of neural models for document ranking

19Citations
Citations of this article
69Readers
Mendeley users who have this article in their library.

Abstract

We tackle the challenge of cross-lingual training of neural document ranking models for mono-lingual retrieval, specifically leveraging relevance judgments in English to improve search in non-English languages. Our work successfully applies multi-lingual BERT (mBERT) to document ranking and additionally compares against a number of alternatives: translating the training data, translating documents, multi-stage hybrids, and ensembles. Experiments on test collections in six different languages from diverse language families reveal many interesting findings: model-based relevance transfer using mBERT can significantly improve search quality in (non-English) mono-lingual retrieval, but other “low resource” approaches are competitive as well.

Cite

CITATION STYLE

APA

Shi, P., Bai, H., & Lin, J. (2020). Cross-lingual training of neural models for document ranking. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 2768–2773). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.249

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free