
6 CONCLUSIONS AND FUTURE
WORK
This paper presented the automatic item generation
concepts within the e-assessment system JACK. The
focus was not on new capabilities in item generation,
but on the software design. In particular, the con-
cept can be considered more sustainable than com-
mon standalone, single-purpose solutions, as it is in-
tegrated, reusable and extensible:
• The concept is integrated, because item genera-
tion happens directly within an e-assessment sys-
tem that is also responsible for delivering assess-
ment items to students and for automatic grading
and feedback generation. Thus, there is no need
to create an ecosystem around the tool as it might
be the case with other approaches.
• The concept supports reusability, because any of
its features is realized in a distinct component that
can be combined freely. The core concepts of item
variables and dynamic objects are generic for as-
sessment items and can be used regardless of the
item types of their parts.
• The concept supports extensibility, because each
component can be updated without interfering
with the remaining system. There is no need to
build a new system just for new types of depen-
dencies between item variables. Instead, evaluator
functions can be added as needed. There is also no
need to build a new system just for new types of
content elements. Instead, evaluator functions or
dynamic objects can be added as needed. There is
also no need for a new system just because of dif-
ferent representations. Instead, dynamic objects
or output conversions can be added as needed.
Besides a general extension of JACK’s item gen-
eration capabilities in terms of new evaluator func-
tions, future work particularly includes the inclusion
of a new type of dynamic objects for data structures in
computer science education. While visualizing data
structures in general is not a hard problem, it puts
more emphasis on the internal structure of item vari-
ables (e. g. it might become important in what order a
list contains elements).
A current research project tackles the automatic
generation of questions on program code submitted
by students in response to programming assignments.
While the research project is concerned with more
fundamental aspects of asking questions about pro-
gram code, the creation of practical demonstrators
may require to extend the current item generation ca-
pabilities of JACK with specific features for handling
program code.
REFERENCES
Baum, H., Damnik, G., Gierl, M., and Braun, I. (2021). A
shift in automatic item generation towards more com-
plex tasks. In INTED2021 Proceedings, 15th Interna-
tional Technology, Education and Development Con-
ference, pages 3235–3241. IATED.
Choi, J., Kim, H., and Pak, S. (2018). Evaluation of Auto-
matic Item Generation Utilities in Formative Assess-
ment Application for Korean High School Students.
Journal of Educational Issues, 4(1).
Choi, J. and Zhang, X. (2019). Computerized Item Mod-
eling Practices using Computer Adaptive Formative
Assessment Automatic Item Generation System: A
Tutorial. The Quantitative Methods for Psychology,
15(3):214–225.
Christ, P., Munkelt, T., and Haake, J. M. (2024). Gener-
alized Automatic Item Generation for Graphical Con-
ceptual Modeling Tasks. In Proceedings of the 16th
International Conference on Computer Supported Ed-
ucation, CSEDU 2024, Angers, France, May 2-4,
2024, Volume 1, pages 807–818. SCITEPRESS.
Circi, R., Hicks, J., and Sikali, E. (2023). Automatic item
generation: foundations and machine learning-based
approaches for assessments. Front. in Educ., 8.
Foulonneau, M. and Ras, E. (2013). Using Educational Do-
main Models for Automatic Item Generation Beyond
Factual Knowledge Assessment. In Proceedings of
EC-TEL 2013: Scaling up Learning for Sustained Im-
pact, pages 442–447. Springer Berlin Heidelberg.
Gierl, M. J. and Lai, H. (2012a). The Role of Item Models
in Automatic Item Generation. International Journal
of Testing, 12(3):273–298.
Gierl, M. J. and Lai, H. (2012b). Using Weak and Strong
Theory to Create Item Models for Automatic Item
Generation: Some Practical Guidelines with Exam-
ples. In Automatic Item Generation. Routledge.
Kosh, A. E., Simpson, M. A., Bickel, L., Kellogg, M., and
Sanford-Moore, E. (2019). A Cost–Benefit Analysis
of Automatic Item Generation. Educational Measure-
ment: Issues and Practice, 38(1):48–53.
Kıyak, Y. S. and Kononowicz, A. A. (2024). Case-based
MCQ generator: A custom ChatGPT based on pub-
lished prompts in the literature for automatic item gen-
eration. Medical Teacher, 46(8):1018–1020. PMID:
38340312.
Liu, B. (2009). SARAC: A Framework for Automatic Item
Generation. In 2009 Ninth IEEE International Confer-
ence on Advanced Learning Technologies, pages 556–
558.
Massing, T., Schwinning, N., Striewe, M., Hanck, C., and
Goedicke, M. (2018). E-Assessment Using Variable-
Content Exercises in Mathematical Statistics. Journal
of Statistics Education, 26(3):174–189.
Mortimer, T., Stroulia, E., and Yazdchi, M. V. (2012).
IGOR: A Web-Based Automatic Item Generation
Tool. In Automatic Item Generation. Routledge.
Striewe, M. (2016). An architecture for modular grading
and feedback generation for complex exercises. Sci-
ence of Computer Programming, 129:35–47.
Automatic Item Generation Integrated into the E-Assessment-System JACK
753