What Makes My Model Perplexed? A Linguistic Investigation on Neural Language Models Perplexity

9Citations
Citations of this article
65Readers
Mendeley users who have this article in their library.

Abstract

This paper presents an investigation aimed at studying how the linguistic structure of a sentence affects the perplexity of two of the most popular Neural Language Models (NLMs), BERT and GPT-2. We first compare the sentence–level likelihood computed with BERT and the GPT-2’s perplexity showing that the two metrics are correlated. In addition, we exploit linguistic features capturing a wide set of morpho-syntactic and syntactic phenomena showing how they contribute to predict the perplexity of the two NLMs.

References Powered by Scopus

Aligning books and movies: Towards story-like visual explanations by watching movies and reading books

1682Citations
N/AReaders
Get full text

Data from eye-tracking corpora as evidence for theories of syntactic processing complexity

422Citations
N/AReaders
Get full text

Blimp: The benchmark of linguistic minimal pairs for english

265Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Adversarial Text Rewriting for Text-aware Recommender Systems

1Citations
N/AReaders
Get full text

Leveraging Deep Embedding Models for Arabic Text Summaries Evaluation

0Citations
N/AReaders
Get full text

Exploring Players’ Perspectives: A Comprehensive Topic Modeling Case Study on Elden Ring

0Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Miaschi, A., Brunato, D., Dell’Orletta, F., & Venturi, G. (2021). What Makes My Model Perplexed? A Linguistic Investigation on Neural Language Models Perplexity. In Deep Learning Inside Out: 2nd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, DeeLIO 2021 - Proceedings, co-located with the Annual Conference of the North American Chapter of the Association for Computational Linguistics, NAACL-HLT 2021 (pp. 40–47). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.deelio-1.5

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 15

63%

Researcher 5

21%

Professor / Associate Prof. 2

8%

Lecturer / Post doc 2

8%

Readers' Discipline

Tooltip

Computer Science 19

68%

Linguistics 6

21%

Business, Management and Accounting 2

7%

Neuroscience 1

4%

Article Metrics

Tooltip
Mentions
References: 1

Save time finding and organizing research with Mendeley

Sign up for free