Mrqa 2019 shared task: Evaluating generalization in reading comprehension

228Citations
Citations of this article
157Readers
Mendeley users who have this article in their library.

Abstract

We present the results of the Machine Reading for Question Answering (MRQA) 2019 shared task on evaluating the generalization capabilities of reading comprehension systems. 1 In this task we adapted and unified 18 distinct question answering datasets into the same format. Among them six datasets were made available for training six datasets were made available for development and the final six were hidden for final evaluation. Ten teams submitted systems which explored various ideas including data sampling multi-task learning adversarial training and ensembling. The best system achieved an average F1 score of 72.5 on the 12 held-out datasets 10.7 absolute points higher than our initial baseline based on BERT.

Cite

CITATION STYLE

APA

Fisch, A., Talmor, A., Jia, R., Seo, M., Choi, E., & Chen, D. (2019). Mrqa 2019 shared task: Evaluating generalization in reading comprehension. In MRQA@EMNLP 2019 - Proceedings of the 2nd Workshop on Machine Reading for Question Answering (pp. 1–13). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d19-5801

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free