Neural Ranking with Weak Supervision for Open-Domain Question Answering: A Survey

2Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

Neural ranking (NR) has become a key component for open-domain question-answering in order to access external knowledge. However, training a good NR model requires substantial amounts of relevance annotations, which is very costly to scale. To address this, a growing body of research works have been proposed to reduce the annotation cost by training the NR model with weak supervision (WS) instead. These works differ in what resources they require and employ a diverse set of WS signals to train the model. Understanding such differences is crucial for choosing the right WS technique. To facilitate this understanding, we provide a structured overview of standard WS signals used for training a NR model. Based on their required resources, we divide them into three main categories: (1) only documents are needed; (2) documents and questions are needed; and (3) documents and question-answer pairs are needed. For every WS signal, we review its general idea and choices. Promising directions are outlined for future research.

Cite

CITATION STYLE

APA

Shen, X., Vakulenko, S., del Tredici, M., Barlacchi, G., Byrne, B., & de Gispert, A. (2023). Neural Ranking with Weak Supervision for Open-Domain Question Answering: A Survey. In EACL 2023 - 17th Conference of the European Chapter of the Association for Computational Linguistics, Findings of EACL 2023 (pp. 1691–1705). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-eacl.129

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free