Challenges in including extra-linguistic context in pre-trained language models

0Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

To successfully account for language, computational models need to take into account both the linguistic context (the content of the utterances) and the extra-linguistic context (for instance, the participants in a dialogue). We focus on a referential task that asks models to link entity mentions in a TV show to the corresponding characters, and design an architecture that attempts to account for both kinds of context. In particular, our architecture combines a previously proposed specialized module (an “entity library”) for character representation with transfer learning from a pre-trained language model. We find that, although the model does improve linguistic contextualization, it fails to successfully integrate extra-linguistic information about the participants in the dialogue. Our work shows that it is very challenging to incorporate extra-linguistic information into pretrained language models.

Cite

CITATION STYLE

APA

Sorodoc, I. T., Aina, L., & Boleda, G. (2022). Challenges in including extra-linguistic context in pre-trained language models. In Insights 2022 - 3rd Workshop on Insights from Negative Results in NLP, Proceedings of the Workshop (pp. 134–138). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.insights-1.18

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free