Metric for Automatic Machine Translation Evaluation based on Pre-trained Sentence Embeddings

  • Shimanaka H
  • Kajiwara T
  • Komachi M
N/ACitations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

This study describes a segment-level metric for automatic machine translation evaluation (MTE). Although various MTE metrics have been proposed, most MTE metrics, including the current de facto standard BLEU, can handle only limited information for segment-level MTE. Therefore, we propose an MTE metric using pre-trained sentence embeddings in order to evaluate MT translation considering global information. In our proposed method, we obtain sentence embeddings of MT translation and reference translation using a sentence encoder pre-trained on a large corpus. Then, we estimate the translation quality by a regression model based on sentence embeddings of MT translation and reference translation as input. Our metric achieved state-of-the-art performance in segment-level metrics tasks for all to-English language pairs on the WMT dataset with human evaluation score.

Cite

CITATION STYLE

APA

Shimanaka, H., Kajiwara, T., & Komachi, M. (2019). Metric for Automatic Machine Translation Evaluation based on Pre-trained Sentence Embeddings. Journal of Natural Language Processing, 26(3), 613–634. https://doi.org/10.5715/jnlp.26.613

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free