Neural Machine Translation with Bilingual History Involved Attention

0Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The using of attention in neural machine translation (NMT) has greatly improved translation performance, but NMT models usually calculate attention vectors independently at different time steps and consequently suffer from over-translation and under-translation. To mitigate the problem, in this paper we propose a method to consider the translated source and target information up to now related to each source word when calculating attentions. The main idea is to keep track of the translated source and target information assigned to each source word at each time step and then accumulate these information to get the completion degree for each source word. In this way, in the later calculation of the attention, the model can adjust the attention weights to give a reasonable final completion degree for each source word. Experimental results show that our method can outperform the strong baseline systems significantly both on the Chinese-English and English-German translation tasks and produce better alignment on the human aligned data set.

Cite

CITATION STYLE

APA

Xue, H., Feng, Y., You, D., Zhang, W., & Li, J. (2019). Neural Machine Translation with Bilingual History Involved Attention. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11839 LNAI, pp. 265–275). Springer. https://doi.org/10.1007/978-3-030-32236-6_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free