Understanding neural machine translation by simplification: The Case of Encoder-free Models

12Citations
Citations of this article
91Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we try to understand neural machine translation (NMT) via simplifying NMT architectures and training encoder-free NMT models. In an encoder-free model, the sums of word embeddings and positional embeddings represent the source. The decoder is a standard Transformer or recurrent neural network that directly attends to embeddings via attention mechanisms. Experimental results show (1) that the attention mechanism in encoder-free models acts as a strong feature extractor, (2) that the word embeddings in encoder-free models are competitive to those in conventional models, (3) that non-contextualized source representations lead to a big performance drop, and (4) that encoder-free models have different effects on alignment quality for German→English and Chinese→English.

Cite

CITATION STYLE

APA

Tang, G., Sennrich, R., & Nivre, J. (2019). Understanding neural machine translation by simplification: The Case of Encoder-free Models. In International Conference Recent Advances in Natural Language Processing, RANLP (Vol. 2019-September, pp. 1186–1193). Incoma Ltd. https://doi.org/10.26615/978-954-452-056-4_136

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free