DML: Dynamic Multi-Granularity Learning for BERT-Based Document Reranking

7Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, pre-trained language models have been successfully applied to the task of text retrieval and ranking. However, in real scenes, users' click behavior is usually affected by selection, position, or exposure bias, which may lead to insufficient positive annotations and introduce additional noise. And for different candidate documents of the same query, the previous optimization objectives usually use a single granularity and static loss weights. It makes the performance of ranking models more susceptible to the bias issue mentioned above. Thus, in this paper, we focus on BERT-based document reranking and propose Dynamic Multi-Granularity Learning (DML). By introducing Gaussian distribution into traditional loss functions, the weights of different documents can change dynamically according to the prediction probability to avoid the impact of unlabeled positive documents. Besides, both document-granularity and instance-granularity are considered to balance relative relations and absolute scores of candidate documents. Extensive experiments show that DML significantly outperforms previous state-of-the-art models on the MS MARCO document ranking dataset.

Cite

CITATION STYLE

APA

Zhang, X., & Yang, Q. (2021). DML: Dynamic Multi-Granularity Learning for BERT-Based Document Reranking. In International Conference on Information and Knowledge Management, Proceedings (pp. 3642–3646). Association for Computing Machinery. https://doi.org/10.1145/3459637.3482090

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free