Identifying Incorrect Labels in the CoNLL-2003 Corpus

21Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

Abstract

The CoNLL-2003 corpus for English-language named entity recognition (NER) is one of the most influential corpora for NER model research. A large number of publications, including many landmark works, have used this corpus as a source of ground truth for NER tasks. In this paper, we examine this corpus and identify over 1300 incorrect labels (out of 35089 in the corpus). In particular, the number of incorrect labels in the test fold is comparable to the number of errors that state-of-the-art models make when running inference over this corpus. We describe the process by which we identified these incorrect labels, using novel variants of techniques from semi-supervised learning. We also summarize the types of errors that we found, and we revisit several recent results in NER in light of the corrected data. Finally, we show experimentally that our corrections to the corpus have a positive impact on three state-ofthe-art models.

Cite

CITATION STYLE

APA

Reiss, F., Xu, H., Cutler, B., Muthuraman, K., & Eichenberger, Z. (2020). Identifying Incorrect Labels in the CoNLL-2003 Corpus. In CoNLL 2020 - 24th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 215–226). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.conll-1.16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free