Transformer with Syntactic Position Encoding for Machine Translation

2Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It has been widely recognized that syntax information can help end-to-end neural machine translation (NMT) systems to achieve better translation. In order to integrate dependency information into Transformer based NMT, existing approaches either exploit words' local head-dependent relations, ignoring their non-local neighbors carrying important context; or approximate two words' syntactic relation by their relative distance on the dependency tree, sacrificing exactness. To address these issues, we propose global positional encoding for dependency tree, a new scheme that facilitates syntactic relation modeling between any two words with keeping exactness and without immediate neighbor constraint. Experiment results on NC11 German?English, English?German and WMT English?German datasets show that our approach is more effective than the above two strategies. In addition, our experiments quantitatively show that compared with higher layers, lower layers of the model are more proper places to incorporate syntax information in terms of each layer's preference to the syntactic pattern and the final performance.

Cite

CITATION STYLE

APA

Xie, Y., Wang, W., Du, M., & He, Q. (2021). Transformer with Syntactic Position Encoding for Machine Translation. In International Conference Recent Advances in Natural Language Processing, RANLP (pp. 1536–1544). Incoma Ltd. https://doi.org/10.26615/978-954-452-072-4_172

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free