Phrase-based machine translation is state-of-the-art for automatic grammatical error correction

70Citations
Citations of this article
160Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this work, we study parameter tuning towards the M2 metric, the standard metric for automatic grammar error correction (GEC) tasks. After implementing M2 as a scorer in the Moses tuning framework, we investigate interactions of dense and sparse features, different optimizers, and tuning strategies for the CoNLL-2014 shared task. We notice erratic behavior when optimizing sparse feature weights with M2 and offer partial solutions. We find that a bare-bones phrase-based SMT setup with task-specific parameter-tuning outperforms all previously published results for the CoNLL-2014 test set by a large margin (46.37% M2 over previously 41.75%, by an SMT system with neural features) while being trained on the same, publicly available data. Our newly introduced dense and sparse features widen that gap, and we improve the state-of-the-art to 49.49% M2.

Cite

CITATION STYLE

APA

Junczys-Dowmunt, M., & Grundkiewicz, R. (2016). Phrase-based machine translation is state-of-the-art for automatic grammatical error correction. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1546–1556). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1161

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free