A stochastic decoder for neural machine translation

19Citations
Citations of this article
196Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The process of translation is ambiguous, in that there are typically many valid translations for a given sentence. This gives rise to significant variation in parallel corpora, however, most current models of machine translation do not account for this variation, instead treating the problem as a deterministic process. To this end, we present a deep generative model of machine translation which incorporates a chain of latent variables, in order to account for local lexical and syntactic variation in parallel corpora. We provide an in-depth analysis of the pitfalls encountered in variational inference for training deep generative models. Experiments on several different language pairs demonstrate that the model consistently improves over strong baselines.

Cite

CITATION STYLE

APA

Schulz, P., Aziz, W., & Cohn, T. (2018). A stochastic decoder for neural machine translation. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 1243–1252). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-1115

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free