CDGP: Automatic Cloze Distractor Generation based on Pre-trained Language Model

21Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Manually designing cloze test consumes enormous time and efforts. The major challenge lies in wrong option (distractor) selection. Having carefully-design distractors improves the effectiveness of learner ability assessment. As a result, the idea of automatically generating cloze distractor is motivated. In this paper, we investigate cloze distractor generation by exploring the employment of pre-trained language models (PLMs) as an alternative for candidate distractor generation. Experiments show that the PLM-enhanced model brings a substantial performance improvement. Our best performing model advances the state-of-the-art result from 14.94 to 34.17 (NDCG@10 score). Our code and dataset is available at https://github.com/AndyChiangSH/CDGP.

Cite

CITATION STYLE

APA

Chiang, S. H., Wang, S. C., & Fan, Y. C. (2022). CDGP: Automatic Cloze Distractor Generation based on Pre-trained Language Model. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 5864–5869). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.31

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free