Advances in NLP have yielded impressive results for the task of machine reading comprehension (MRC), with approaches having been reported to achieve performance comparable to that of humans. In this paper, we investigate whether stateof- the-art MRC models are able to correctly process Semantics Altering Modifications (SAM): linguistically-motivated phenomena that alter the semantics of a sentence while preserving most of its lexical surface form. We present a method to automatically generate and align challenge sets featuring original and altered examples. We further propose a novel evaluation methodology to correctly assess the capability of MRC systems to process these examples independent of the data they were optimised on, by discounting for effects introduced by domain shift. In a large-scale empirical study, we apply the methodology in order to evaluate extractive MRC models with regard to their capability to correctly process SAM-enriched data. We comprehensively cover 12 different state-of-the-art neural architecture configurations and four training datasets and find that - despite their well-known remarkable performance - optimised models consistently struggle to correctly process semantically altered data.
CITATION STYLE
Schlegel, V., Nenadic, G., & Batista-Navarro, R. (2021). Semantics Altering Modifications for Evaluating Comprehension in Machine Reading. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 15, pp. 13762–13770). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i15.17622
Mendeley helps you to discover research relevant for your work.