Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks?

48Citations
Citations of this article
137Readers
Mendeley users who have this article in their library.

Abstract

Do state-of-the-art natural language understanding models care about word order? Not always! We found 75% to 90% of the correct predictions of BERT-based classifiers, trained on many GLUE tasks, remain constant after input words are randomly shuffled. Although BERT embeddings are famously contextual, the contribution of each individual word to classification is almost unchanged even after its surrounding words are shuffled. BERT-based models exploit superficial cues (e.g. the sentiment of keywords in sentiment analysis; or the word-wise similarity between sequence-pair inputs in natural language inference) to make correct decisions when tokens are randomly shuffled. Encouraging models to capture word order information improves the performance on most GLUE tasks and SQuAD 2.0. Our work suggests that many GLUE tasks are not challenging machines to understand the meaning of a sentence.

References Powered by Scopus

Long Short-Term Memory

77867Citations
N/AReaders
Get full text

"Why should i trust you?" Explaining the predictions of any classifier

12161Citations
N/AReaders
Get full text

Finding structure in time

8248Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Structural Persistence in Language Models: Priming as a Window into Abstract Language Representations

25Citations
N/AReaders
Get full text

Neural reality of argument structure constructions

19Citations
N/AReaders
Get full text

Why is Winoground Hard? Investigating Failures in Visuolinguistic Compositionality

17Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Pham, T. M., Bui, T., Mai, L., & Nguyen, A. (2021). Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks? In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 1145–1160). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.98

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 43

64%

Researcher 16

24%

Professor / Associate Prof. 4

6%

Lecturer / Post doc 4

6%

Readers' Discipline

Tooltip

Computer Science 58

77%

Linguistics 12

16%

Neuroscience 3

4%

Engineering 2

3%

Save time finding and organizing research with Mendeley

Sign up for free