Results of the WMT16 Metrics Shared Task

86Citations
Citations of this article
171Readers
Mendeley users who have this article in their library.

Abstract

This paper presents the results of the WMT16 Metrics Shared Task. We asked participants of this task to score the outputs of the MT systems involved in the WMT16 Shared Translation Task. We collected scores of 16 metrics from 9 research groups. In addition to that, we computed scores of 9 standard metrics (BLEU, SentBLEU, NIST, WER, PER, TER and CDER) as baselines. The collected scores were evaluated in terms of system-level correlation (how well each metric's scores correlate with WMT16 official manual ranking of systems) and in terms of segment level correlation (how often a metric agrees with humans in comparing two translations of a particular sentence). This year there are several additions to the setup: large number of language pairs (18 in total), datasets from different domains (news, IT and medical), and different kinds of judgments: relative ranking (RR), direct assessment (DA) and HUME manual semantic judgments. Finally, generation of large number of hybrid systems was trialed for provision of more conclusive system-level metric rankings.

Cite

CITATION STYLE

APA

Bojar, O., Graham, Y., Kamran, A., & Stanojević, M. (2016). Results of the WMT16 Metrics Shared Task. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 199–231). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w17-4755

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free