A Closer Look at k-Nearest Neighbors Grammatical Error Correction

1Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.

Abstract

In various natural language processing tasks, such as named entity recognition and machine translation, example-based approaches have been used to improve performance by leveraging existing knowledge. However, the effectiveness of this approach for Grammatical Error Correction (GEC) is unclear. In this work, we explore how an example-based approach affects the accuracy and interpretability of the output of GEC systems and the trade-offs involved. The approach we investigate has shown great promise in machine translation by using the k nearest translation examples to improve the results of a pretrained Transformer model. We find that using this technique increases precision by reducing the number of false positives, but recall suffers as the model becomes more conservative overall. Increasing the number of example sentences in the datastore does lead to better performing systems, but with diminishing returns and a high decoding cost. Synthetic data can be used as examples, but the effectiveness varies depending on the base model. Finally, we find that finetuning on a set of data may be more effective than using that data during decoding as examples.

Cite

CITATION STYLE

APA

Vasselli, J., & Watanabe, T. (2023). A Closer Look at k-Nearest Neighbors Grammatical Error Correction. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 220–231). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.bea-1.19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free