Improving variational encoder-decoders in dialogue generation

73Citations
Citations of this article
150Readers
Mendeley users who have this article in their library.

Abstract

Variational encoder-decoders (VEDs) have shown promising results in dialogue generation. However, the latent variable distributions are usually approximated by a much simpler model than the powerful RNN structure used for encoding and decoding, yielding the KL-vanishing problem and inconsistent training objective. In this paper, we separate the training step into two phases: The first phase learns to autoencode discrete texts into continuous embeddings, from which the second phase learns to generalize latent representations by reconstructing the encoded embedding. In this case, latent variables are sampled by transforming Gaussian noise through multi-layer perceptrons and are trained with a separate VED model, which has the potential of realizing a much more flexible distribution. We compare our model with current popular models and the experiment demonstrates substantial improvement in both metric-based and human evaluations.

Cite

CITATION STYLE

APA

Shen, X., Su, H., Niu, S., & Demberg, V. (2018). Improving variational encoder-decoders in dialogue generation. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 5456–5463). AAAI press. https://doi.org/10.1609/aaai.v32i1.11960

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free