Stabilizing Minimum Error Rate Training

26Citations
Citations of this article
84Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The most commonly used method for training feature weights in statistical machine translation (SMT) systems is Och's minimum error rate training (MERT) procedure. A well-known problem with Och's procedure is that it tends to be sensitive to small changes in the system, particularly when the number of features is large. In this paper, we quantify the stability of Och's procedure by supplying different random seeds to a core component of the procedure (Powell's algorithm). We show that for systems with many features, there is extensive variation in outcomes, both on the development data and on the test data. We analyze the causes of this variation and propose modifications to the MERT procedure that improve stability while helping performance on test data.

Cite

CITATION STYLE

APA

Foster, G., & Kuhn, R. (2009). Stabilizing Minimum Error Rate Training. In EACL 2009 - 4th Workshop on Statistical Machine Translation, Proceedings of theWorkshop (pp. 242–249). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1626431.1626478

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free