Flatness-Aware Prompt Selection Improves Accuracy and Sample Efficiency

0Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

With the growing capabilities of large language models, prompting them has become the dominant way to access them. This has motivated the development of strategies for automatically selecting effective language prompts. In this paper, we introduce PFLAT (prompt flatness), a new metric to quantify the expected utility of a language prompt. This metric is inspired by flatness regularization in statistical learning that quantifies the robustness of the model towards its parameter perturbations. We provide theoretical foundations for this metric and its relationship with other prompt selection metrics, providing a comprehensive understanding of existing methods. Empirically, we show that combining PFLAT with existing metrics improves both performance and sample efficiency. Our metric outperforms the previous prompt selection metrics with an average increase of 10% in Pearson correlation across 6 classification benchmarks, and the prompt selected by our metric gains 5% higher accuracy than previous metrics across the benchmarks.

Cite

CITATION STYLE

APA

Shen, L., Tan, W., Zheng, B., & Khashabi, D. (2023). Flatness-Aware Prompt Selection Improves Accuracy and Sample Efficiency. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 7795–7817). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.523

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free