Exponentially decaying bag-of-words input features for feed-forward neural network in statistical machine translation

2Citations
Citations of this article
95Readers
Mendeley users who have this article in their library.

Abstract

Recently, neural network models have achieved consistent improvements in statistical machine translation. However, most networks only use one-hot encoded input vectors of words as their input. In this work, we investigated the exponentially decaying bag-of-words input features for feed-forward neural network translation models and proposed to train the decay rates along with other weight parameters. This novel bag-of-words model improved our phrase-based state-of-the-art system, which already includes a neural network translation model, by up to 0.5% BLEU and 0.6% TER on three different translation tasks and even achieved a similar performance to the bidirectional LSTM translation model.

Cite

CITATION STYLE

APA

Peter, J. T., Wang, W., & Ney, H. (2016). Exponentially decaying bag-of-words input features for feed-forward neural network in statistical machine translation. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Short Papers (pp. 293–298). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p16-2048

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free