Abstract
Dialog response generation in open domain is an important research topic where the main challenge is to generate relevant and diverse responses. In this paper, we propose a new dialog pre-training framework called DialogVED, which introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses. With the help of a large dialog corpus (Reddit), we pre-train the model using the following 4 tasks, used in training language models (LMs) and Variational Autoencoders (VAEs) literature: 1) masked language model; 2) response generation; 3) bag-of-words prediction; and 4) KL divergence reduction. We also add additional parameters to model the turn structure in dialogs to improve the performance of the pre-trained model. We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation. Experimental results show that our model achieves the new state-of-the-art results on all these datasets.
Cite
CITATION STYLE
Chen, W., Gong, Y., Wang, S., Yao, B., Qi, W., Wei, Z., … Duan, N. (2022). DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 4852–4864). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.333
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.