
These non-significant correlations suggest that
others unexplored variables may be playing an
important role in these relationships. Possible
limitations, such as sample size or lack of control for
confounding variables, may influence the strength of
the observed associations.
One hypothesis for this lack of correlation is the
possibility that problems related to complexity and
refactoring are too complex for beginner students to
solve on their own, or even that the suggestions
presented for these code smells are not enough for
students to be able to refactor their codes.
6 CONCLUSIONS
In this study, we developed a qualitative feedback
system and investigated its ability to influence and
improve the quality of beginning students’ code,
analyzing the correlation between resubmissions to
the system and some classes of code smells. Our
results provide positive insights into the use of the
system to improve the quality of student codes,
showing great promise in the task of refactoring
variable names.
During the study, we observed a significant
correlation between the number of resubmissions to
the system and the number of code smells related to
the nomenclature of variable names in the students’
code.
Despite the promising results obtained in this
study, it is important to recognize some limitations
that may influence our conclusions. Firstly, the
sample used in this study was restricted to a single
class of algorithms, which may limit the
generalization of the results to other populations.
Furthermore, due to the nature of longitudinal design
adopted, we can only capture participant behavior
during this specific period, without the ability to
observe long-term changes or usage patterns. This
may limit our understanding of tool usage trends over
time and its long-term sustainability. Finally, it is
important to note that external factors, such as
changes in participants’ individual circumstances or
unforeseen events, may have influenced the use of the
tool throughout the study period.
Ultimately, our study highlights the importance of
automated qualitative feedback related to code
quality in online judge environments as a practical
intervention to promote evaluation of novice
students’ codes beyond the functional. We hope that
this study inspires other researchers to contribute
even more to this still little explored field.
ACKNOWLEDGMENTS
We would like to acknowledge CAPES for the
financial support.
REFERENCES
Araujo, E., Serey, D., & Figueiredo, J. (2016, October).
Qualitative aspects of students' programs: Can we make
them measurable?. In 2016 IEEE Frontiers in
Education Conference (FIE) (pp. 1-8). IEEE.
Birillo, A., Vlasov, I., Burylov, A., Selishchev, V.,
Goncharov, A., Tikhomirova, E., ... & Bryksin, T.
(2022, February). Hyperstyle: A tool for assessing the
code quality of solutions to programming assignments.
In Proceedings of the 53rd ACM Technical Symposium
on Computer Science Education-Volume 1 (pp. 307-
313).
Chren, S., Macák, M., Rossi, B., & Buhnova, B. (2022,
June). Evaluating code improvements in software
quality course projects. In Proceedings of the 26th
International Conference on Evaluation and
Assessment in Software Engineering (pp. 160-169).
Fowler, M. (1999). Refactoring: improving the design of
existing code. Addison-Wesley Professional.
Jiang, L., Rewcastle, R., Denny, P., & Tempero, E. (2020,
June). Comparecfg: Providing visual feedback on code
quality using control flow graphs. In Proceedings of the
2020 ACM Conference on Innovation and Technology
in Computer Science Education (pp. 493-499).
Keuning, H., Heeren, B., & Jeuring, J. (2020, November).
Student refactoring behaviour in a programming tutor.
In Proceedings of the 20th Koli Calling International
Conference on Computing Education Research (pp. 1-
10).
Liu, X., & Woo, G. (2020, February). Applying code
quality detection in online programming judge.
In Proceedings of the 2020 5th International
Conference on Intelligent Information Technology (pp.
56-60).
Orr, J. W. (2020, novembro). Automatic assessment of the
design quality of student python and java programs. J.
Comput. Sci. Coll., 38(1), 27-36.
Ureel II, L. C., & Wallace, C. (2019, February). Automated
critique of early programming antipatterns.
In Proceedings of the 50th ACM Technical Symposium
on Computer Science Education (pp. 738-744).
Wasik, S., Antczak, M., Badura, J., Laskowski, A., &
Sternal, T. (2018). A survey on online judge systems
and their applications. ACM Computing Surveys
(CSUR), 51(1), 1-34.
Zhou, W., Pan, Y., Zhou, Y., & Sun, G. (2018, May). The
framework of a new online judge system for
programming education. In Proceedings of ACM turing
celebration conference-China (pp. 9-14).
CSEDU 2024 - 16th International Conference on Computer Supported Education
584