This study describes a segment-level metric for automatic machine translation evaluation (MTE). Although various MTE metrics have been proposed, most MTE metrics, including the current de facto standard BLEU, can handle only limited information for segment-level MTE. Therefore, we propose an MTE metric using pre-trained sentence embeddings in order to evaluate MT translation considering global information. In our proposed method, we obtain sentence embeddings of MT translation and reference translation using a sentence encoder pre-trained on a large corpus. Then, we estimate the translation quality by a regression model based on sentence embeddings of MT translation and reference translation as input. Our metric achieved state-of-the-art performance in segment-level metrics tasks for all to-English language pairs on the WMT dataset with human evaluation score.
CITATION STYLE
Shimanaka, H., Kajiwara, T., & Komachi, M. (2019). Metric for Automatic Machine Translation Evaluation based on Pre-trained Sentence Embeddings. Journal of Natural Language Processing, 26(3), 613–634. https://doi.org/10.5715/jnlp.26.613
Mendeley helps you to discover research relevant for your work.