Recent progress in the task of Grammatical Error Correction (GEC) has been driven by addressing data sparsity, both through new methods for generating large and noisy pre-training data and through the publication of small and higher-quality finetuning data in the BEA-2019 shared task. Building upon recent work in Neural Machine Translation (NMT), we make use of both kinds of data by deriving example-level scores on our large pretraining data based on a smaller, higher-quality dataset. In this work, we perform an empirical study to discover how to best incorporate delta-log-perplexity, a type of example scoring, into a training schedule for GEC. In doing so, we perform experiments that shed light on the function and applicability of delta-log-perplexity. Models trained on scored data achieve state-of-the-art results on common GEC test sets.
CITATION STYLE
Lichtarge, J., Alberti, C., & Kumar, S. (2020). Data weighted training strategies for grammatical error correction. Transactions of the Association for Computational Linguistics, 8, 634–646. https://doi.org/10.1162/tacl_a_00336
Mendeley helps you to discover research relevant for your work.