Reqa: An evaluation for end-to-end answer retrieval models

38Citations
Citations of this article
123Readers
Mendeley users who have this article in their library.

Abstract

Popular QA benchmarks like SQuAD have driven progress on the task of identifying answer spans within a specific passage, with models now surpassing human performance. However, retrieving relevant answers from a huge corpus of documents is still a challenging problem, and places different requirements on the model architecture. There is growing interest in developing scalable answer retrieval models trained end-to-end, bypassing the typical document retrieval step. In this paper, we introduce Retrieval Question- Answering (ReQA), a benchmark for evaluating large-scale sentence-level answer retrieval models. We establish baselines using both neural encoding models as well as classical information retrieval techniques. We release our evaluation code to encourage further work on this challenging task.

Cite

CITATION STYLE

APA

Ahmad, A., Constant, N., Yang, Y., & Cer, D. (2019). Reqa: An evaluation for end-to-end answer retrieval models. In MRQA@EMNLP 2019 - Proceedings of the 2nd Workshop on Machine Reading for Question Answering (pp. 137–146). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d19-5819

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free