DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering

19Citations
Citations of this article
65Readers
Mendeley users who have this article in their library.

Abstract

Disfluencies is an under-studied topic in NLP, even though it is ubiquitous in human conversation. This is largely due to the lack of datasets containing disfluencies. In this paper, we present a new challenge question answering dataset, DISFL-QA, a derivative of SQUAD, where humans introduce contextual disfluencies in previously fluent questions. DISFL-QA contains a variety of challenging disfluencies that require a more comprehensive understanding of the text than what was necessary in prior datasets. Experiments show that the performance of existing state-of-the-art question answering models degrades significantly when tested on DISFL-QA in a zero-shot setting. We show data augmentation methods partially recover the loss in performance and also demonstrate the efficacy of using gold data for fine-tuning. We argue that we need large-scale disfluency datasets in order for NLP models to be robust to them. The dataset is publicly available at: https://github.com/google-research-datasets/disfl-qa.

References Powered by Scopus

SQuad: 100,000+ questions for machine comprehension of text

3973Citations
N/AReaders
Get full text

SWITCHBOARD: Telephone speech corpus for research and development

1519Citations
N/AReaders
Get full text

Know what you don’t know: Unanswerable questions for SQuAD

1350Citations
N/AReaders
Get full text

Cited by Powered by Scopus

QA Dataset Explosion: A Taxonomy of NLP Resources for Question Answering and Reading Comprehension

90Citations
N/AReaders
Get full text

A Mixed-Methods Approach to Understanding User Trust after Voice Assistant Failures

9Citations
N/AReaders
Get full text

Streaming Joint Speech Recognition and Disfluency Detection

4Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Gupta, A., Xu, J., Upadhyay, S., Yang, D., & Faruqui, M. (2021). DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 3309–3319). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.293

Readers over time

‘21‘22‘23‘24‘2507142128

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 13

57%

Researcher 7

30%

Professor / Associate Prof. 2

9%

Lecturer / Post doc 1

4%

Readers' Discipline

Tooltip

Computer Science 25

78%

Linguistics 5

16%

Philosophy 1

3%

Neuroscience 1

3%

Article Metrics

Tooltip
Mentions
Blog Mentions: 2

Save time finding and organizing research with Mendeley

Sign up for free
0