
7 CONCLUSIONS
In this study, we explored if and how often 199 stu-
dents from five community colleges in the United
States checked auto-grader feedback as they took the
same introductory Python programming course, to
see if their feedback-checking behavior seems related
to their scores. Our results clearly show the relation-
ship between the scores and students checking the
feedback. The more often a student checks auto-
grader feedback for a programming assignment, the
more likely they are to get a higher score for that
assignment. Furthermore, checking the auto-grader
feedback for a non-maximal, non-terminal submis-
sion is associated with a 4.69% higher probability of
getting a higher score in the subsequent submission
for the same task, than not checking it. Our findings
are based on logging student navigation to the web
pages hosting the individualized feedback, though we
do not know if the student indeed read the feedback
or how carefully they read it.
8 FUTURE WORK
Further analysis and categorization of our feedback,
using frameworks such as Narciss’ (Narciss, 2008),
will allow us to better understand what types of feed-
back are more useful than others. Keuning et al. also
called for such a comparison in their systematic liter-
ature review (Keuning et al., 2018).
Additionally, both researchers and instructors
could benefit from instrumenting the learning plat-
form for students to provide feedback on the feedback
they get, so that researchers can evaluate their useful-
ness and instructors can improve their courses. As we
enter an era where programming education is ubiq-
uitous for learners at all levels, and generative AI is
starting to generate course content and contextualized
feedback, it becomes all the more necessary that we
continue to demonstrate the usefulness (and usabil-
ity) of auto-grader feedback provided to learners, so
as to ensure that the next generation of programmers
is well-prepare to enter today’s technology workforce.
ACKNOWLEDGMENT
This material is based upon work supported by
the National Science Foundation under Grant No.
2111305.
Hosting of the educational platform, and Azure
credits for some student learning activities on the
cloud service provider, are sponsored by Microsoft.
Recruitment of the participating community col-
leges was accomplished in collaboration with the Na-
tional Institute for Staff and Organizational Develop-
ment (NISOD).
REFERENCES
An, M., Zhang, H., Savelka, J., Zhu, S., Bogart, C., and
Sakr, M. (2021). Are working habits different between
well-performing and at-risk students in online project-
based courses? In Proceedings of the 26th ACM Con-
ference on Innovation and Technology in Computer
Science Education V. 1, pages 324–330.
Bennedsen, J. and Caspersen, M. E. (2007). Failure
rates in introductory programming. SIGCSE Bull.,
39(2):32–36.
Bogart, C., An, M., Keylor, E., Singh, P., Savelka, J., and
Sakr, M. (2024). What factors influence persistence
in project-based programming courses at community
colleges? In Proceedings of the 55th ACM Techni-
cal Symposium on Computer Science Education V. 1,
pages 116–122.
Gabbay, H. and Cohen, A. (2022). Exploring the connec-
tions between the use of an automated feedback sys-
tem and learning behavior in a mooc for program-
ming. In Educating for a New Future: Making
Sense of Technology-Enhanced Learning Adoption:
17th European Conference on Technology Enhanced
Learning, EC-TEL 2022, Toulouse, France, Septem-
ber 12–16, 2022, Proceedings, page 116–130, Berlin,
Heidelberg. Springer-Verlag.
Goldstein, S. C., Zhang, H., Sakr, M., An, H., and Dashti,
C. (2019). Understanding how work habits influence
student performance. In Proceedings of the 2019 ACM
Conference on Innovation and Technology in Com-
puter Science Education, pages 154–160.
Keuning, H., Jeuring, J., and Heeren, B. (2018). A system-
atic literature review of automated feedback genera-
tion for programming exercises. ACM Trans. Comput.
Educ., 19(1).
Kokotsaki, D., Menzies, V., and Wiggins, A. (2016).
Project-based learning: A review of the literature. Im-
proving Schools, 19.
Kumar, A. N. (2005). Generation of problems, answers,
grade, and feedback—case study of a fully automated
tutor. J. Educ. Resour. Comput., 5(3):3–es.
Kurniawan, O., Poskitt, C. M., Al Hoque, I., Lee, N. T. S.,
J
´
egourel, C., and Sockalingam, N. (2023). How help-
ful do novice programmers find the feedback of an
automated repair tool? In 2023 IEEE International
Conference on Teaching, Assessment and Learning for
Engineering (TALE), pages 1–6.
Messer, M., Brown, N. C. C., K
¨
olling, M., and Shi, M.
(2024). Automated grading and feedback tools for
programming education: A systematic review. ACM
Trans. Comput. Educ., 24(1).
Mitra, J. (2023). Studying the impact of auto-graders giving
immediate feedback in programming assignments. In
CSEDU 2025 - 17th International Conference on Computer Supported Education
362