In recent years, transformer-based coreference resolution systems have achieved remarkable improvements on the CoNLL dataset. However, how coreference resolvers can benefit from discourse coherence is still an open question. In this paper, we propose to incorporate centering transitions derived from centering theory in the form of a graph into a neural coreference model. Our method improves the performance over the SOTA baselines, especially on pronoun resolution in long documents, formal well-structured text, and clusters with scattered mentions.
CITATION STYLE
Chai, H., & Strube, M. (2022). Incorporating Centering Theory into Neural Coreference Resolution. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2996–3002). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.218
Mendeley helps you to discover research relevant for your work.