Learning continuous phrase representations for translation modeling

89Citations
Citations of this article
159Readers
Mendeley users who have this article in their library.

Abstract

This paper tackles the sparsity problem in estimating phrase translation probabilities by learning continuous phrase representations, whose distributed nature enables the sharing of related phrases in their representations. A pair of source and target phrases are projected into continuous-valued vector representations in a low-dimensional latent space, where their translation score is computed by the distance between the pair in this new space. The projection is performed by a neural network whose weights are learned on parallel training data. Experimental evaluation has been performed on two WMT translation tasks. Our best result improves the performance of a state-of-the-art phrase-based statistical machine translation system trained on WMT 2012 French-English data by up to 1.3 BLEU points.

Cite

CITATION STYLE

APA

Gao, J., He, X., Yih, W. T., & Deng, L. (2014). Learning continuous phrase representations for translation modeling. In 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014 - Proceedings of the Conference (Vol. 1, pp. 699–709). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/p14-1066

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free