Abstract
To successfully account for language, computational models need to take into account both the linguistic context (the content of the utterances) and the extra-linguistic context (for instance, the participants in a dialogue). We focus on a referential task that asks models to link entity mentions in a TV show to the corresponding characters, and design an architecture that attempts to account for both kinds of context. In particular, our architecture combines a previously proposed specialized module (an “entity library”) for character representation with transfer learning from a pre-trained language model. We find that, although the model does improve linguistic contextualization, it fails to successfully integrate extra-linguistic information about the participants in the dialogue. Our work shows that it is very challenging to incorporate extra-linguistic information into pretrained language models.
Cite
CITATION STYLE
Sorodoc, I. T., Aina, L., & Boleda, G. (2022). Challenges in including extra-linguistic context in pre-trained language models. In Insights 2022 - 3rd Workshop on Insights from Negative Results in NLP, Proceedings of the Workshop (pp. 134–138). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.insights-1.18
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.