MatRank: Text Re-ranking by Latent Preference Matrix

1Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

Text ranking plays a key role in providing content that best answers user queries. It is usually divided into two sub-tasks to perform efficient information retrieval given a query: text retrieval and text re-ranking. Recent research on pretrained language models (PLM) has demonstrated efficiency and gain on both sub-tasks. However, while existing methods have benefited from pre-trained language models and achieved high recall rates on passage retrieval, the ranking performance still demands further improvement. In this paper, we propose MatRank, which learns to re-rank the text retrieved for a given query by learning to predict the most relevant passage based on a latent preference matrix. Specifically, MatRank uses a PLM to generate an asymmetric latent matrix of relative preference scores between all pairs of retrieved passages. Then, the latent matrix is aggregated row-wise and column-wise to obtain global preferences and predictions of the most relevant passage in two of these directions, respectively. We conduct extensive experiments on MS MACRO, WikiAQ, and SemEval datasets. Experimental results show that MatRank has achieved new state-of-the-art results on these datasets, outperforming all prior methods on ranking performance metrics.

Cite

CITATION STYLE

APA

Luo, J., Yang, J., Guo, W., Li, C., Niu, D., & Xu, Y. (2022). MatRank: Text Re-ranking by Latent Preference Matrix. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 2011–2023). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.146

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free