Chunk-based decoder for neural machine translation

12Citations
Citations of this article
109Readers
Mendeley users who have this article in their library.

Abstract

Chunks (or phrases) once played a pivotal role in machine translation. By using a chunk rather than a word as the basic translation unit, local (intra-chunk) and global (inter-chunk) word orders and dependencies can be easily modeled. The chunk structure, despite its importance, has not been considered in the decoders used for neural machine translation (NMT). In this paper, we propose chunk-based decoders for NMT, each of which consists of a chunk-level decoder and a word-level decoder. The chunk-level decoder models global dependencies while the word-level decoder decides the local word order in a chunk. To output a target sentence, the chunk-level decoder generates a chunk representation containing global information, which the word-level decoder then uses as a basis to predict the words inside the chunk. Experimental results show that our proposed decoders can significantly improve translation performance in a WAT'16 English-to-Japanese translation task.

Cite

CITATION STYLE

APA

Ishiwatar, S., Yao, J., Liu, S., Li, M., Zhou, M., Yoshinaga, N., … Jia, W. (2017). Chunk-based decoder for neural machine translation. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 1901–1912). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-1174

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free