An Error-Guided Correction Model for Chinese Spelling Error Correction

8Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Although existing neural network approaches have achieved great success on Chinese spelling correction, there is still room to improve. The model is required to avoid over-correction and to distinguish a correct token from its phonological and visually similar ones. In this paper, we propose an error-guided correction model (EGCM) to improve Chinese spelling correction. By borrowing the powerful ability of BERT, we propose a novel zero-shot error detection method to do a preliminary detection, which guides our model to attend more on the probably wrong tokens in encoding and to avoid modifying the correct tokens in generating. Furthermore, we introduce a new loss function to integrate the error confusion set, which enables our model to distinguish easily misused tokens. Moreover, our model supports highly parallel decoding to meet real application requirements. Experiments are conducted on widely used benchmarks. Our model achieves superior performance against state-of-the-art approaches by a remarkable margin, on both the correction quality and computation speed.

Cite

CITATION STYLE

APA

Sun, R., Wu, X., & Wu, Y. (2022). An Error-Guided Correction Model for Chinese Spelling Error Correction. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 3829–3839). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.319

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free