Language Modeling with Latent Situations

2Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

Language models (LMs) often generate incoherent outputs: they refer to events and entity states that are incompatible with the state of the world described in inputs. We introduce SITUATIONSUPERVISION, a family of approaches for improving coherence in LMs by training them to construct and condition on explicit representations of entities and their states. SITUATIONSUPERVISION has two components: an auxiliary situation modeling task that trains models to predict entity state representations in context, and a latent state inference procedure that imputes these states from partially annotated training data. SITUATIONSUPERVISION can be applied via fine-tuning (by supervising LMs to encode state variables in their hidden representations) and prompting (by inducing LMs to interleave textual descriptions of entity states with output text). In both cases, it requires only a small number of state annotations to produce substantial coherence improvements (up to an 16% reduction in errors), showing that standard LMs can be efficiently adapted to explicitly model language and aspects of its meaning.

Cite

CITATION STYLE

APA

Li, B. Z., Nye, M., & Andreas, J. (2023). Language Modeling with Latent Situations. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 12556–12571). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.795

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free