Abstract
We discuss two named-entity recognition models which use characters and character n-grams either exclusively or as an important part of their data representation. The first model is a character-level HMM with minimal context information, and the second model is a maximum-entropy conditional markov model with substantially richer context features. Our best model achieves an overall F1 of 86.07% on the English test data (92.31% on the development data). This number represents a 25% error reduction over the same model without word-internal (substring) features.
Cite
CITATION STYLE
Klein, D., Smarr, J., Nguyen, H., & Manning, C. D. (2003). Named Entity Recognition with Character-Level Models. In Proceedings of the 7th Conference on Natural Language Learning, CoNLL 2003 at HLT-NAACL 2003 (pp. 180–183). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1119176.1119204
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.