期刊名称:Research and Practice in Technology Enhanced Learning
电子版ISSN:1793-7078
出版年度:2017
卷号:12
期号:1
页码:1-21
DOI:10.1186/s41039-017-0051-y
语种:English
出版社:Springer
摘要:Abstract This paper describes details of the evaluation experiments for questions created by an automatic question generation system. Given a target word and one of its word senses, the system generates a multiple-choice English vocabulary question asking for the closest in meaning to the target word in the reading passage. Two kinds of evaluation were conducted considering two aspects: (1) measuring English learners’ proficiency and (2) their similarity to the human-made questions. The first evaluation is based on the responses from English learners obtained through administering the machine-generated and human-made questions to them, and the second is based on the subjective judgement by English teachers. Both evaluations showed that the machine-generated questions were able to achieve a comparable level with the human-made questions in both measuring English proficiency and similarity.
关键词:English vocabulary question;Automatic question generation;Multiple-choice question;Evaluation of question items;Neural test theory;Language learning