In-Context Demonstration Selection with Cross Entropy Difference

4Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

Large language models (LLMs) can use in-context demonstrations to improve performance on zero-shot tasks. However, selecting the best in-context examples is challenging because model performance can vary widely depending on the selected examples. We present a cross-entropy difference (CED) method for selecting in-context demonstrations. Our method is based on the observation that the effectiveness of in-context demonstrations negatively correlates with the perplexity of the test example by a language model that was finetuned on that demonstration. We utilize parameter efficient finetuning to train small models on training data that are used for computing the cross-entropy difference between a test example and every candidate in-context demonstration. This metric is used to rank and select in-context demonstrations independently for each test input. We evaluate our method on a mix-domain dataset that combines 8 benchmarks, representing 4 text generation tasks, showing that CED for in-context demonstration selection can improve performance for a variety of LLMs over baseline selection methods.

Cite

CITATION STYLE

APA

Iter, D., Pryzant, R., Xu, R., Wang, S., Liu, Y., Xu, Y., & Zhu, C. (2023). In-Context Demonstration Selection with Cross Entropy Difference. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 1150–1162). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.81

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free