Noise-Robust Training with Dynamic Loss and Contrastive Learning for Distantly-Supervised Named Entity Recognition

6Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Distantly-supervised named entity recognition (NER) aims at training networks with distantly-labeled data, which is automatically obtained by matching entity mentions in the raw text with entity types in a knowledge base. Distant supervision may induce incomplete and noisy labels, so recent state-of-the-art methods employ sample selection mechanism to separate clean data from noisy data based on the model's prediction scores. However, they ignore the noise distribution change caused by data selection, and they simply excludes noisy data during training, resulting in information loss. We propose to (1) use a dynamic loss function to better adapt to the changing noise during the training process, and (2) incorporate token level contrastive learning to fully utilize the noisy data as well as facilitate feature learning without relying on labels. We conduct extensive experiments on multiple datasets and our method obtains 4.3%, 1.5%, 0.9% F1 score improvements over the current state-of-the-art on Wikigold, CoNLL03 and OntoNotes5.0.

Cite

CITATION STYLE

APA

Ma, Z., Du, J., & Zhou, S. (2023). Noise-Robust Training with Dynamic Loss and Contrastive Learning for Distantly-Supervised Named Entity Recognition. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 10119–10128). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.643

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free