TART: Improved Few-shot Text Classification Using Task-Adaptive Reference Transformation

16Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Meta-learning has emerged as a trending technique to tackle few-shot text classification and achieve state-of-the-art performance. However, the performance of existing approaches heavily depends on the inter-class variance of the support set. As a result, it can perform well on tasks when the semantics of sampled classes are distinct while failing to differentiate classes with similar semantics. In this paper, we propose a novel Task-Adaptive Reference Transformation (TART) network, aiming to enhance the generalization by transforming the class prototypes to per-class fixed reference points in task-adaptive metric spaces. To further maximize divergence between transformed prototypes in task-adaptive metric spaces, TART introduces a discriminative reference regularization among transformed prototypes. Extensive experiments are conducted on four benchmark datasets and our method demonstrates clear superiority over the state-of-the-art models in all the datasets. In particular, our model surpasses the state-of-the-art method by 7.4% and 5.4% in 1-shot and 5-shot classification on the 20 Newsgroups dataset, respectively. Our code is available at https://github.com/slei109/TART.

Cite

CITATION STYLE

APA

Lei, S., Zhang, X., He, J., Chen, F., & Lu, C. T. (2023). TART: Improved Few-shot Text Classification Using Task-Adaptive Reference Transformation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 11014–11026). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.617

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free