Coarse-to-fine Few-shot Learning for Named Entity Recognition

11Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Recently, Few-shot Named Entity Recognition has received wide attention with the growing need for NER models to learn new classes with minimized annotation costs. However, one common yet understudied situation is to transfer a model trained with coarse-grained classes to recognize fine-grained classes, such as separating a product category into sub-classes. We find that existing few-shot NER solutions are not suitable for such a situation since they do not consider the sub-class discrimination during coarse training and various granularity of new classes during few-shot learning. In this work, we introduce the Coarse-to-fine Few-shot NER (C2FNER) task and propose an effective solution. Specifically, during coarse training, we propose a cluster-based prototype margin loss to learn group-wise discriminative representations, so as to benefit fine-grained learning. Targeting various granularity of new classes, we separate the coarse classes into extra-fine clusters and propose a novel prototype retrieval and bootstrapping algorithm to retrieve representative clusters for each fine class. We then adopt a mixture prototype loss to efficiently learn the representations of fine classes. We conduct experiments on both in-domain and cross-domain C2FNER settings with various target granularity, and the proposed method shows superior performance over the baseline methods.

Cite

CITATION STYLE

APA

Ma, R., Zhang, L., Chen, X., Zhou, X., Wang, J., Gui, T., … Chen, Y. (2023). Coarse-to-fine Few-shot Learning for Named Entity Recognition. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 4115–4129). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.253

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free