Explaining data patterns in natural language with language models

2Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Large language models (LLMs) have displayed an impressive ability to harness natural language to perform complex tasks. We explore whether we can leverage this ability to find and explain patterns in data. Specifically, given a pre-trained LLM and data examples, we apply interpretable autoprompting (iPrompt) to generate a natural language string explaining the data. iPrompt iteratively generates explanations with an LLM and reranks them based on their performance when used as a prompt. Experiments on a wide range of datasets, from synthetic mathematics to natural language understanding, show that iPrompt can yield meaningful insights by accurately finding dataset explanations that are human-interpretable. Moreover, iPrompt is reasonably efficient, as it does not require access to model gradients and works with relatively small models (e.g. 6 billion parameters rather than ≥100 billion). Finally, experiments with scientific datasets show the potential for iPrompt to aid in scientific discovery.1

Cite

CITATION STYLE

APA

Singh, C., Morris, J. X., Aneja, J., Rush, A. M., & Gao, J. (2023). Explaining data patterns in natural language with language models. In BlackboxNLP 2023 - Analyzing and Interpreting Neural Networks for NLP, Proceedings of the 6th Workshop (pp. 31–55). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.blackboxnlp-1.3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free