Neural net models of open-domain discourse coherence

77Citations
Citations of this article
225Readers
Mendeley users who have this article in their library.

Abstract

Discourse coherence is strongly associated with text quality, making it important to natural language generation and understanding. Yet existing models of coherence focus on measuring individual aspects of coherence (lexical overlap, rhetorical structure, entity centering) in narrow domains. In this paper, we describe domain-independent neural models of discourse coherence that are capable of measuring multiple aspects of coherence in existing sentences and can maintain coherence while generating new sentences. We study both discriminative models that learn to distinguish coherent from incoherent discourse, and generative models that produce coherent text, including a novel neural latent-variable Markovian generative model that captures the latent discourse dependencies between sentences in a text. Our work achieves state-of-the-art performance on multiple coherence evaluations, and marks an initial step in generating coherent texts given discourse contexts.

Cite

CITATION STYLE

APA

Li, J., & Jurafsky, D. (2017). Neural net models of open-domain discourse coherence. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 198–209). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1019

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free