Convolutional over Recurrent Encoder for Neural Machine Translation

  • Dakwale P
  • Monz C
N/ACitations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Neural machine translation is a recently proposed approach which has shown competitive results to traditional MT approaches. Standard neural MT is an end-to-end neural network where the source sentence is encoded by a recurrent neural network (RNN) called encoder and the target words are predicted using another RNN known as decoder. Recently, various models have been proposed which replace the RNN encoder with a convolutional neural network (CNN). In this paper, we propose to augment the standard RNN encoder in NMT with additional convolutional layers in order to capture wider context in the encoder output. Experiments on English to German translation demonstrate that our approach can achieve significant improvements over a standard RNN-based baseline.

Cite

CITATION STYLE

APA

Dakwale, P., & Monz, C. (2017). Convolutional over Recurrent Encoder for Neural Machine Translation. The Prague Bulletin of Mathematical Linguistics, 108(1), 37–48. https://doi.org/10.1515/pralin-2017-0007

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free