Abstract
We introduce our system that is submitted to the News Commentary task (Japanese↔Russian) of the 6th Workshop on Asian Translation. The goal of this shared task is to study extremely low resource situations for distant language pairs. It is known that using parallel corpora of different language pair as training data is effective for multilingual neural machine translation model in extremely low resource scenarios. Therefore, to improve the translation quality of Japanese↔Russian language pair, our method leverages other in-domain Japanese-English and English-Russian parallel corpora as additional training data for our multilingual NMT model.
Cite
CITATION STYLE
Imankulova, A., Kaneko, M., & Komachi, M. (2021). Japanese-Russian TMU neural machine translation system using multilingual model for WAT 2019. In WAT@EMNLP-IJCNLP 2019 - 6th Workshop on Asian Translation, Proceedings (pp. 165–170). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d19-5221
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.