Abstract
Recently, neural network models have achieved consistent improvements in statistical machine translation. However, most networks only use one-hot encoded input vectors of words as their input. In this work, we investigated the exponentially decaying bag-of-words input features for feed-forward neural network translation models and proposed to train the decay rates along with other weight parameters. This novel bag-of-words model improved our phrase-based state-of-the-art system, which already includes a neural network translation model, by up to 0.5% BLEU and 0.6% TER on three different translation tasks and even achieved a similar performance to the bidirectional LSTM translation model.
Cite
CITATION STYLE
Peter, J. T., Wang, W., & Ney, H. (2016). Exponentially decaying bag-of-words input features for feed-forward neural network in statistical machine translation. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Short Papers (pp. 293–298). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p16-2048
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.