
agree with the statement “I understand the concept
of AIG and know what corresponding software can
be used for.”, indicates that this understanding of the
AIG process was also conveyed to the participants
through their work with the editor.
6 FUTURE WORK
Currently, the AIG Model Editor only generates tasks
and items through the alternation of all valid combi-
nations of generic and case-specific features (see Sec-
tion 2.3). Others proposed to use large language
models (Sayin and Gierl, 2024; Kıyak, 2023). This
approach allows to generate tasks and items whose
wording and content is not explicitly given as gener-
ation input. But like the AIG Model Editor, it still
requires describing the topic and the form of target
tasks and items, and evaluating the generation result.
In further development of the AIG Model Editor,
large language models could be used for the genera-
tion, similar to the other works. Besides, it is believed
that the integration of artificial intelligence into the
modeling process would further reduce the cognitive
effort required during modeling. Therefore, it is cur-
rently examined in which steps and how users could
best be supported by this new technology. At the mo-
ment, analyzing the subject area, creating the cogni-
tive model, and formulating item stems seem to be
possible candidates.
REFERENCES
Alrakhawi, H. A., Jamiat, N., & Abu-Naser, S. S. (2023).
Intelligent Tutoring Systems In Education: A System-
atic Review Of Usage, Tools, Effects And Evaluation.
Journal of Theoretical and Applied Information Tech-
nology. 101. 1205 - 1226.
Arsovic, B., & Stefanovic, N. (2020). E-learning based on
the adaptive learning model: case study in Serbia.
S
¯
adhan
¯
a, 45(1), 266.
Baum, H., Damnik, G., Gierl, M. & Braun, I. (2021).
A Shift in automatic Item Generation towards more
complex Tasks, INTED2021 Proceedings, pp. 3235-
3241.
Braun, I., Kapp, F., Hara, T., Kubica, T., & Schill, A.
(2018). AMCS (Auditorium Mobile Classroom Ser-
vice)–an ARS with Learning Questions, Push Notifi-
cations, and extensive Means of Evaluation. In CEUR
Workshop Proceedings (Vol. 2092).
Braun, I., Damnik, G., & Baum, H. (2022). Rethinking As-
sessment with Automatic Item Generation, Inted2022
Proceedings, pp. 1176-1180.
Brooke, J. (1996). SUS: a “quick and dirty” usability scale.
In P. Jordan, B. Thomas, & B. Weerdmeester (Eds.),
Usability Evaluation in Industry (pp. 189–194). Lon-
don, UK: Taylor & Francis.
Chi, M., & Boucher, N. (2023). Applying the ICAP Frame-
work to Improve Classroom Learning. In C. E. Over-
son, C. M. Hakala, L. L. Kordonowy, & V. A. Benassi
(Eds.), In their own words: What scholars and teach-
ers want you to know about why and how to apply
the science of learning in your academic setting (pp.
94-110). Society for the Teaching of Psychology.
Chi, M., & Wylie, R. (2014). The ICAP framework: Link-
ing cognitive engagement to active learning outcomes.
Educational psychologist, 49(4), 219-243.
Damnik, G., Gierl, M., Proske, A., K
¨
orndle, H., & Nar-
ciss, S. (2018). Automatic Item Generation as a
Means to Increase Interactivity and Adaptivity in
Digital Learning Resources [Automatische Erzeugung
von Aufgaben als Mittel zur Erh
¨
ohung von Interak-
tivit
¨
at und Adaptivit
¨
at in digitalen Lernressourcen].
In E-Learning Symposium 2018 (pp. 5-16). Univer-
sit
¨
atsverlag Potsdam.
Drasgow, F. (2016). Technology and testing: Improving ed-
ucational and psychological measurement. New York:
Routledge.
Embretson, S. E. (2002). Generating abstract reasoning
items with cognitive theory. In S. H. Irvine and P. C.
Kyllonen (Eds.), Item generation for test development
(pp. 219–250). Mahwah, NJ: Lawrence Erlbaum As-
sociates.
Embretson, S. E., & Yang, X. (2007). Automatic item gen-
eration and cognitive psychology. In C. R. Rao &
S. Sinharay (Eds.), Handbook of statistics: Psycho-
metrics, Volume 26 (pp. 747–768). Amsterdam, The
Netherlands: Elsevier.
Falc
˜
ao, F., Costa, P., & P
ˆ
ego, J. M. (2022). Feasibility assur-
ance: a review of automatic item generation in med-
ical assessment. Advances in Health Sciences Educa-
tion, 27(2), 405-425.
Gierl, M. J., & Lai, H. (2013a). Evaluating the quality of
medical multiple-choice items created with automated
processes. Medical education, 47(7), 726-733.
Gierl, M. J., & Lai, H. (2013b). Instructional topics in ed-
ucational measurement (ITEMS) module: Using au-
tomated processes to generate test items. Educational
Measurement: Issues and Practice, 32(3), 36-50.
Gierl, M. J., & Lai, H. (2013c). Using weak and strong the-
ory to create item models for automatic item genera-
tion: Some practical guidelines with examples. In M.
J. Gierl & T. Haladyna (Eds.), Automatic item genera-
tion: Theory and practice (pp. 26–39). New York, NY:
Routledge.
Gierl, M. J., & Lai, H. (2016). Automatic item generation.
In S. Lane, M. R. Raymond, & T.M. Haladyna (Eds.),
Handbook of test development (2nd ed., pp. 410–429).
New York, NY: Routledge.
Gierl, M. J., Lai, H., & Turner, S. (2012). Using auto-
matic item generation to create multiple-choice items
for assessments in medical education. Medical Educa-
tion,46, 757–765.
Kapp, F., Proske, A., Narciss, S., & K
¨
orndle, H. (2015).
Distributing vs. blocking learning questions in a web-
Overcoming Student Passivity with Automatic Item Generation
797