Entailment Semantics Can Be Extracted from an Ideal Language Model

5Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Language models are often trained on text alone, without additional grounding. There is debate as to how much of natural language semantics can be inferred from such a procedure. We prove that entailment judgments between sentences can be extracted from an ideal language model that has perfectly learned its target distribution, assuming the training sentences are generated by Gricean agents, i.e., agents who follow fundamental principles of communication from the linguistic theory of pragmatics. We also show entailment judgments can be decoded from the predictions of a language model trained on such Gricean data. Our results reveal a pathway for understanding the semantic information encoded in unlabeled linguistic data and a potential framework for extracting semantics from language models.

Cite

CITATION STYLE

APA

Merrill, W., Warstadt, A., & Linzen, T. (2022). Entailment Semantics Can Be Extracted from an Ideal Language Model. In CoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 176–193). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.conll-1.13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free