Retrieval-Augmented Few-shot Text Classification

8Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Retrieval-augmented methods are successful in the standard scenario where the retrieval space is sufficient; whereas in the few-shot scenario with limited retrieval space, this paper shows it is non-trivial to put them into practice. First, it is impossible to retrieve semantically similar examples by using an off-the-shelf metric and it is crucial to learn a task-specific retrieval metric; Second, our preliminary experiments demonstrate that it is difficult to optimize a plausible metric by minimizing the standard cross-entropy loss. The in-depth analyses quantitatively show minimizing cross-entropy loss suffers from the weak supervision signals and the severe gradient vanishing issue during the optimization. To address these issues, we introduce two novel training objectives, namely EML and R-L, which provide more task-specific guidance to the retrieval metric by the EM algorithm and a ranking-based loss, respectively. Extensive experiments on 10 datasets prove the superiority of the proposed retrieval augmented methods on the performance.

Cite

CITATION STYLE

APA

Yu, G., Liu, L., Jiang, H., Shi, S., & Ao, X. (2023). Retrieval-Augmented Few-shot Text Classification. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 6721–6735). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.447

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free