Abstract
In this paper, we propose a new method for calculating the output layer in neural machine translation systems. The method is based on predicting a binary code for each word, and can reduce computation time/memory requirements of the output time/memory requirements of the output layer to be logarithmic in vocabulary size in the best case. In addition, we also introduce two advanced approaches to improve robustness of the proposed model: using error-correcting codes and combining soft-max and binary codes. Experiments show the proposed model achieves translation accuracies that approach the softmax, while reducing memory usage on the order of one tenths, and also improving decoding speed on CPUs by x5 to x10.
Cite
CITATION STYLE
Oda, Y., Arthur, P., Neubig, G., Yoshino, K., & Nakamura, S. (2018). Neural Machine Translation Models using Binarized Prediction and Error Correction. Journal of Natural Language Processing, 25(2), 167–199. https://doi.org/10.5715/jnlp.25.167
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.