QuadrupletBERT: An Efficient Model For Embedding-Based Large-Scale Retrieval

7Citations
Citations of this article
59Readers
Mendeley users who have this article in their library.

Abstract

The embedding-based large-scale query-document retrieval problem is a hot topic in the information retrieval (IR) field. Considering that pre-trained language models like BERT have achieved great success in a wide variety of NLP tasks, we present a QuadrupletBERT model for effective and efficient retrieval in this paper. Unlike most existing BERT-style retrieval models, which only focus on the ranking phase in retrieval systems, our model makes considerable improvements to the retrieval phase and leverages the distances between simple negative and hard negative instances to obtaining better embeddings. Experimental results demonstrate that our QuadrupletBERT achieves state-of-the-art results in embedding-based large-scale retrieval tasks.

References Powered by Scopus

A statistical interpretation of term specificity and its application in retrieval

2988Citations
N/AReaders
Get full text

The probabilistic relevance framework: BM25 and beyond

2207Citations
N/AReaders
Get full text

Embedding-based Retrieval in Facebook Search

217Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Semantic matching based legal information retrieval system for COVID-19 pandemic

13Citations
N/AReaders
Get full text

Distilling Knowledge from BERT into Simple Fully Connected Neural Networks for Efficient Vertical Retrieval

5Citations
N/AReaders
Get full text

Improving Embedding-based Large-scale Retrieval via Label Enhancement

5Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Liu, P., Wang, S., Wang, X., Ye, W., & Zhang, S. (2021). QuadrupletBERT: An Efficient Model For Embedding-Based Large-Scale Retrieval. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 3734–3739). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.292

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 14

67%

Researcher 4

19%

Lecturer / Post doc 2

10%

Professor / Associate Prof. 1

5%

Readers' Discipline

Tooltip

Computer Science 18

75%

Linguistics 4

17%

Neuroscience 1

4%

Social Sciences 1

4%

Save time finding and organizing research with Mendeley

Sign up for free