Abstract
Figurative language is widespread in human language (Lakoff and Johnson, 2008), posing potential challenges in NLP applications. In this paper, we investigate the effect of figurative language on the task of question answering (QA). We construct FigurativeQA, a test set of 400 yes-no questions with figurative and non-figurative contexts, extracted from product reviews and restaurant reviews. We demonstrate that a state-of-the-art RoBERTa QA model has considerably lower performance in question answering when the contexts are figurative rather than literal, indicating a gap in current models. We propose a general method for improving the performance of QA models by converting the figurative contexts into non-figurative by prompting GPT-3, and demonstrate its effectiveness. Our results indicate a need for building QA models infused with figurative language understanding capabilities.
Cite
CITATION STYLE
Rakshit, G., & Flanigan, J. (2022). FigurativeQA: A Test Benchmark for Figurativeness Comprehension for Question Answering. In FLP 2022 - 3rd Workshop on Figurative Language Processing, Proceedings of the Workshop (pp. 160–166). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.flp-1.23
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.