แนวคิดหลัก
Using pre-trained language models for cloze distractor generation significantly improves performance.
บทคัดย่อ
Automatically generating cloze test distractors can enhance learner ability assessment by improving the effectiveness of the test. This paper explores using pre-trained language models (PLMs) to generate distractors, resulting in a substantial performance improvement. The CDGP framework incorporates training and ranking strategies to boost PLM-based distractor generation. Evaluation using benchmarking datasets shows significant outperformance compared to previous methods, advancing NDCG@10 score from 19.31 to 34.17, an improvement of up to 177%. The study also includes related work on cloze distractor generation methods and methodology details on candidate set generation and distractor selection.
สถิติ
Our best performing model advances the state-of-the-art result from 14.94 to 34.17 (NDCG@10 score).
The dataset consists of passages with cloze stems, answers, and distractors.
The CLOTH dataset statistics include average number of sentences and words per passage.
Different pre-trained language models used in experiments include BERT, SciBERT, RoBERTa, and BART.
Evaluation metrics include Precision (P@1), F1 score (F1@3, F1@10), Mean Reciprocal Rank (MRR@10), and Normalized Discounted Cumulative Gain (NDCG@10).
คำพูด
"Manually designing cloze test consumes enormous time and efforts."
"The major challenge lies in wrong option (distractor) selection."
"Our CDGP significantly outperforms the state-of-the-art result."