Abstract
Neural named entity recognition (NER) models may easily encounter the over-confidence issue, which degrades the performance and calibration. Inspired by label smoothing and driven by the ambiguity of boundary annotation in NER engineering, we propose boundary smoothing as a regularization technique for span-based neural NER models. It re-assigns entity probabilities from annotated spans to the surrounding ones. Built on a simple but strong baseline, our model achieves results better than or competitive with previous state-of-the-art systems on eight well-known NER benchmarks. Further empirical analysis suggests that boundary smoothing effectively mitigates over-confidence, improves model calibration, and brings flatter neural minima and more smoothed loss landscapes.
Cite
CITATION STYLE
Zhu, E., & Li, J. (2022). Boundary Smoothing for Named Entity Recognition. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 7096–7108). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.490
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.