Discourse coherence is strongly associated with text quality, making it important to natural language generation and understanding. Yet existing models of coherence focus on measuring individual aspects of coherence (lexical overlap, rhetorical structure, entity centering) in narrow domains. In this paper, we describe domain-independent neural models of discourse coherence that are capable of measuring multiple aspects of coherence in existing sentences and can maintain coherence while generating new sentences. We study both discriminative models that learn to distinguish coherent from incoherent discourse, and generative models that produce coherent text, including a novel neural latent-variable Markovian generative model that captures the latent discourse dependencies between sentences in a text. Our work achieves state-of-the-art performance on multiple coherence evaluations, and marks an initial step in generating coherent texts given discourse contexts.
CITATION STYLE
Li, J., & Jurafsky, D. (2017). Neural net models of open-domain discourse coherence. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 198–209). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1019
Mendeley helps you to discover research relevant for your work.