Saliency-driven word alignment interpretation for neural machine translation

43Citations
Citations of this article
118Readers
Mendeley users who have this article in their library.

Abstract

Despite their original goal to jointly learn to align and translate, Neural Machine Translation (NMT) models, especially Transformer, are often perceived as not learning interpretable word alignments. In this paper, we show that NMT models do learn interpretable word alignments, which could only be revealed with proper interpretation methods. We propose a series of such methods that are model-agnostic, are able to be applied either offline or online, and do not require parameter update or architectural change. We show that under the force decoding setup, the alignments induced by our interpretation method are of better quality than fast-align for some systems, and when performing free decoding, they agree well with the alignments induced by automatic alignment tools.

Cite

CITATION STYLE

APA

Ding, S., Xu, H., & Koehn, P. (2019). Saliency-driven word alignment interpretation for neural machine translation. In WMT 2019 - 4th Conference on Machine Translation, Proceedings of the Conference (Vol. 1, pp. 1–12). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w19-5201

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free