Black-box language model explanation by context length probing

3Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The increasingly widespread adoption of large language models has highlighted the need for improving their explainability. We present context length probing, a novel explanation technique for causal language models, based on tracking the predictions of a model as a function of the length of available context, and allowing to assign differential importance scores to different contexts. The technique is model-agnostic and does not rely on access to model internals beyond computing token-level probabilities. We apply context length probing to large pre-trained language models and offer some initial analyses and insights, including the potential for studying long-range dependencies. The source code1 and an interactive demo2 of the method are available.

Cite

CITATION STYLE

APA

Cífka, O., & Liutkus, A. (2023). Black-box language model explanation by context length probing. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 1067–1079). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-short.92

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free