Language Model Augmented Monotonic Attention for Simultaneous Translation

9Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.

Abstract

The state-of-the-art adaptive policies for simultaneous neural machine translation (SNMT) use monotonic attention to perform read/write decisions based on the partial source and target sequences. The lack of sufficient information might cause the monotonic attention to take poor read/write decisions, which in turn negatively affects the performance of the SNMT model. On the other hand, human translators make better read/write decisions since they can anticipate the immediate future words using linguistic information and domain knowledge. In this work, we propose a framework to aid monotonic attention with an external language model to improve its decisions. Experiments on MuST-C English-German and English-French speech-to-text translation tasks show the future information from language model improves the state-of-the-art monotonic multi-head attention model further.

Cite

CITATION STYLE

APA

Indurthi, S., Zaidi, M. A., Lee, B., Lakumarapu, N. K., & Kim, S. (2022). Language Model Augmented Monotonic Attention for Simultaneous Translation. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 38–45). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free