Abstract
In this paper, we try to understand neural machine translation (NMT) via simplifying NMT architectures and training encoder-free NMT models. In an encoder-free model, the sums of word embeddings and positional embeddings represent the source. The decoder is a standard Transformer or recurrent neural network that directly attends to embeddings via attention mechanisms. Experimental results show (1) that the attention mechanism in encoder-free models acts as a strong feature extractor, (2) that the word embeddings in encoder-free models are competitive to those in conventional models, (3) that non-contextualized source representations lead to a big performance drop, and (4) that encoder-free models have different effects on alignment quality for German→English and Chinese→English.
Cite
CITATION STYLE
Tang, G., Sennrich, R., & Nivre, J. (2019). Understanding neural machine translation by simplification: The Case of Encoder-free Models. In International Conference Recent Advances in Natural Language Processing, RANLP (Vol. 2019-September, pp. 1186–1193). Incoma Ltd. https://doi.org/10.26615/978-954-452-056-4_136
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.