Abstract
In this paper we describe the PKU system for the CoNLL-2014 grammar error correction shared task. We propose a unified framework for correcting all types of errors. We use unlabeled news texts instead of large amount of human annotated texts as training data. Based on these data, a tri-gram language model is used to correct the replacement errors while two extra classification models are trained to correct errors related to determiners and prepositions. Our system achieves 25.32% in f0.5 on the original test data and 29.10% on the revised test data.
Cite
CITATION STYLE
Zhang, L., & Wang, H. (2014). A unified framework for grammar error correction. In CoNLL 2014 - 18th Conference on Computational Natural Language Learning, Proceedings of the Shared Task (pp. 96–102). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/w14-1713
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.