Abstract
The state-of-the-art adaptive policies for simultaneous neural machine translation (SNMT) use monotonic attention to perform read/write decisions based on the partial source and target sequences. The lack of sufficient information might cause the monotonic attention to take poor read/write decisions, which in turn negatively affects the performance of the SNMT model. On the other hand, human translators make better read/write decisions since they can anticipate the immediate future words using linguistic information and domain knowledge. In this work, we propose a framework to aid monotonic attention with an external language model to improve its decisions. Experiments on MuST-C English-German and English-French speech-to-text translation tasks show the future information from language model improves the state-of-the-art monotonic multi-head attention model further.
Cite
CITATION STYLE
Indurthi, S., Zaidi, M. A., Lee, B., Lakumarapu, N. K., & Kim, S. (2022). Language Model Augmented Monotonic Attention for Simultaneous Translation. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 38–45). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.3
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.