Abstract
We tackle the challenge of cross-lingual training of neural document ranking models for mono-lingual retrieval, specifically leveraging relevance judgments in English to improve search in non-English languages. Our work successfully applies multi-lingual BERT (mBERT) to document ranking and additionally compares against a number of alternatives: translating the training data, translating documents, multi-stage hybrids, and ensembles. Experiments on test collections in six different languages from diverse language families reveal many interesting findings: model-based relevance transfer using mBERT can significantly improve search quality in (non-English) mono-lingual retrieval, but other “low resource” approaches are competitive as well.
Cite
CITATION STYLE
Shi, P., Bai, H., & Lin, J. (2020). Cross-lingual training of neural models for document ranking. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 2768–2773). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.249
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.