DISCODVT: Generating Long Text with Discourse-Aware Discrete Variational Transformer

19Citations
Citations of this article
72Readers
Mendeley users who have this article in their library.

Abstract

Despite the recent advances in applying pretrained language models to generate high-quality texts, generating long passages that maintain long-range coherence is yet challenging for these models. In this paper, we propose DISCODVT, a discourse-aware discrete variational Transformer to tackle the incoherence issue. DISCODVT learns a discrete variable sequence that summarizes the global structure of the text and then applies it to guide the generation process at each decoding step. To further embed discourse-aware information into the discrete latent representations, we introduce an auxiliary objective to model the discourse relations within the text. We conduct extensive experiments on two open story generation datasets and demonstrate that the latent codes learn meaningful correspondence to the discourse structures that guide the model to generate long texts with better long-range coherence.

Cite

CITATION STYLE

APA

Ji, H., & Huang, M. (2021). DISCODVT: Generating Long Text with Discourse-Aware Discrete Variational Transformer. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 4208–4224). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.347

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free