DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering

21Citations
Citations of this article
66Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Disfluencies is an under-studied topic in NLP, even though it is ubiquitous in human conversation. This is largely due to the lack of datasets containing disfluencies. In this paper, we present a new challenge question answering dataset, DISFL-QA, a derivative of SQUAD, where humans introduce contextual disfluencies in previously fluent questions. DISFL-QA contains a variety of challenging disfluencies that require a more comprehensive understanding of the text than what was necessary in prior datasets. Experiments show that the performance of existing state-of-the-art question answering models degrades significantly when tested on DISFL-QA in a zero-shot setting. We show data augmentation methods partially recover the loss in performance and also demonstrate the efficacy of using gold data for fine-tuning. We argue that we need large-scale disfluency datasets in order for NLP models to be robust to them. The dataset is publicly available at: https://github.com/google-research-datasets/disfl-qa.

Cite

CITATION STYLE

APA

Gupta, A., Xu, J., Upadhyay, S., Yang, D., & Faruqui, M. (2021). DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 3309–3319). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.293

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free