Disfluencies is an under-studied topic in NLP, even though it is ubiquitous in human conversation. This is largely due to the lack of datasets containing disfluencies. In this paper, we present a new challenge question answering dataset, DISFL-QA, a derivative of SQUAD, where humans introduce contextual disfluencies in previously fluent questions. DISFL-QA contains a variety of challenging disfluencies that require a more comprehensive understanding of the text than what was necessary in prior datasets. Experiments show that the performance of existing state-of-the-art question answering models degrades significantly when tested on DISFL-QA in a zero-shot setting. We show data augmentation methods partially recover the loss in performance and also demonstrate the efficacy of using gold data for fine-tuning. We argue that we need large-scale disfluency datasets in order for NLP models to be robust to them. The dataset is publicly available at: https://github.com/google-research-datasets/disfl-qa.
Mendeley helps you to discover research relevant for your work.
CITATION STYLE
Gupta, A., Xu, J., Upadhyay, S., Yang, D., & Faruqui, M. (2021). DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 3309–3319). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.293