Guiding Neural Machine Translation with Semantic Kernels

1Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

Machine Translation task has made great progress with the help of auto-regressive decoding paradigm and Transformer architecture. In this paradigm, though the encoder can obtain global source representations, the decoder can only use translation history to determine the current word. Previous promising works attempted to address this issue by applying a draft or a fixed-length semantic embedding as target-side global information. However, these methods either degrade model efficiency or show limitations in expressing semantics. Motivated by Functional Equivalence Theory, we extract several semantic kernels from a source sentence, each of which can express one semantic segment of the original sentence. Together, these semantic kernels can capture global semantic information, and we project them into target embedding space to guide target sentence generation. We further force our model to use semantic kernels at each decoding step through an adaptive mask algorithm. Empirical studies on various machine translation benchmarks show that our approach gains approximately an improvement of 1 BLEU score on most benchmarks over the Transformer baseline and about 1.7 times faster than previous works on average at inference time.

Cite

CITATION STYLE

APA

Guo, P., Hu, Y., Wei, X., Ren, Y., Li, Y., Xing, L., & Xie, Y. (2022). Guiding Neural Machine Translation with Semantic Kernels. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 7345–7356). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.541

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free