Abstract
We propose a method for simultaneously translating from a single source language to multiple target languages T1, T2, etc. The motivation behind this method is that if we only have a weak language model for T1 and translations in T1 and T2 are associated, we can use the information from a strong language model over T2 to disambiguate the translations in T1, providing better translation results. As a specific framework to realize multi-target translation, we expand the formalism of synchronous context-free grammars to handle multiple targets, and describe methods for rule extraction, scoring, pruning, and search with these models. Experiments find that multi-target translation with a strong language model in a similar second target language can provide gains of up to 0.8-1.5 BLEU points.
Cite
CITATION STYLE
Neubig, G., Arthur, P., & Duh, K. (2015). Multi-target machine translation with multi-synchronous context-free grammars. In NAACL HLT 2015 - 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 293–302). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/n15-1033
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.