DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation

43Citations
Citations of this article
73Readers
Mendeley users who have this article in their library.

Abstract

Dialog response generation in open domain is an important research topic where the main challenge is to generate relevant and diverse responses. In this paper, we propose a new dialog pre-training framework called DialogVED, which introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses. With the help of a large dialog corpus (Reddit), we pre-train the model using the following 4 tasks, used in training language models (LMs) and Variational Autoencoders (VAEs) literature: 1) masked language model; 2) response generation; 3) bag-of-words prediction; and 4) KL divergence reduction. We also add additional parameters to model the turn structure in dialogs to improve the performance of the pre-trained model. We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation. Experimental results show that our model achieves the new state-of-the-art results on all these datasets.

Cite

CITATION STYLE

APA

Chen, W., Gong, Y., Wang, S., Yao, B., Qi, W., Wei, Z., … Duan, N. (2022). DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 4852–4864). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.333

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free