Generation and Evaluation of Multiple-choice Reading Comprehension Questions for Swedish
DOI:
https://doi.org/10.3384/nejlt.2000-1533.2024.4886Abstract
Multiple-choice questions (MCQs) provide a widely used means of assessing reading comprehension. The automatic generation of such MCQs is a challenging language-technological problem that also has interesting educational applications. This article presents several methods for automatically producing reading comprehension questions MCQs from Swedish text. Unlike previous approaches, we construct models to generate the whole MCQ in one go, rather than using a pipeline architecture. Furthermore, we propose a two-stage method for evaluating the quality of the generated MCQs, first evaluating on carefully designed single-sentence texts, and then on texts from the SFI national exams. An extensive evaluation of the MCQ-generating capabilities of 12 different models, using this two-stage scheme, reveals that GPT-based models surpass smaller models that have been fine-tuned using small-scale datasets on this specific problem.
Downloads
Published
Issue
Section
License
Copyright (c) 2024 Dmytro Kalpakchi, Johan Boye
This work is licensed under a Creative Commons Attribution 4.0 International License.