FigurativeQA: A Test Benchmark for Figurativeness Comprehension for Question Answering

1Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Figurative language is widespread in human language (Lakoff and Johnson, 2008), posing potential challenges in NLP applications. In this paper, we investigate the effect of figurative language on the task of question answering (QA). We construct FigurativeQA, a test set of 400 yes-no questions with figurative and non-figurative contexts, extracted from product reviews and restaurant reviews. We demonstrate that a state-of-the-art RoBERTa QA model has considerably lower performance in question answering when the contexts are figurative rather than literal, indicating a gap in current models. We propose a general method for improving the performance of QA models by converting the figurative contexts into non-figurative by prompting GPT-3, and demonstrate its effectiveness. Our results indicate a need for building QA models infused with figurative language understanding capabilities.

Cite

CITATION STYLE

APA

Rakshit, G., & Flanigan, J. (2022). FigurativeQA: A Test Benchmark for Figurativeness Comprehension for Question Answering. In FLP 2022 - 3rd Workshop on Figurative Language Processing, Proceedings of the Workshop (pp. 160–166). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.flp-1.23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free