Probing neural network comprehension of natural language arguments

228Citations
Citations of this article
563Readers
Mendeley users who have this article in their library.

Abstract

We are surprised to find that BERT's peak performance of 77% on the Argument Reasoning Comprehension Task reaches just three points below the average untrained human baseline. However, we show that this result is entirely accounted for by exploitation of spurious statistical cues in the dataset. We analyze the nature of these cues and demonstrate that a range of models all exploit them. This analysis informs the construction of an adversarial dataset on which all models achieve random accuracy. Our adversarial dataset provides a more robust assessment of argument comprehension and should be adopted as the standard in future work.

Cite

CITATION STYLE

APA

Niven, T., & Kao, H. Y. (2020). Probing neural network comprehension of natural language arguments. In ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (pp. 4658–4664). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p19-1459

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free