Abstract
The process of translation is ambiguous, in that there are typically many valid translations for a given sentence. This gives rise to significant variation in parallel corpora, however, most current models of machine translation do not account for this variation, instead treating the problem as a deterministic process. To this end, we present a deep generative model of machine translation which incorporates a chain of latent variables, in order to account for local lexical and syntactic variation in parallel corpora. We provide an in-depth analysis of the pitfalls encountered in variational inference for training deep generative models. Experiments on several different language pairs demonstrate that the model consistently improves over strong baselines.
Cite
CITATION STYLE
Schulz, P., Aziz, W., & Cohn, T. (2018). A stochastic decoder for neural machine translation. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 1243–1252). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-1115
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.