increased response time contributed to the higher quiz
accuracy in the experimental group. Future studies
should investigate whether the improved
performance was a direct result of the interaction
modality itself or simply a by-productB NM; of
slower, more deliberate responses.
Although the findings provide strong support for
sound-driven interactions in learning environments,
there are several limitations to consider. First, the
study was conducted in a controlled laboratory
setting, which may not fully replicate real-world
classroom conditions, where factors like peer
influence, background noise, and social pressure
could affect engagement and performance.
Additionally, the sample size (N = 40) was relatively
small, limiting the generalizability of the results.
Future research should expand the sample size and
conduct longitudinal studies to examine whether the
benefits of sound-based interactions persist over time.
Furthermore, while IMI provided useful self-reported
insights, alternative engagement measures such as
eye-tracking, physiological responses, or real-time
interaction analytics could offer a more objective
evaluation of engagement levels.
Overall, the findings highlight the potential of
sound-driven human-robot interactions to enhance
learning outcomes by promoting active participation
and cognitive processing. While motivation levels
remained comparable between the two groups, the
higher quiz accuracy in the experimental group
suggests that non-verbal sound interactions can be an
effective alternative to traditional input methods in
educational robotics. Future work should focus on
improving recognition accuracy and use case,
exploring multimodal adaptive learning systems, and
testing these interactions in real-world educational
settings to further validate their effectiveness.
6 CONCLUSIONS
This study investigated the impact of non-verbal
sound-driven interactions on student engagement,
motivation, and learning outcomes in a robot-assisted
quiz-based learning environment. The findings reveal
that students who interacted with the Pepper robot
using sound-based responses achieved significantly
higher quiz accuracy than those who used touch-
based interactions (t = 5.47, p < .001, Cohen’s d =
1.21), suggesting that non-verbal sound cues may
enhance cognitive processing and knowledge
retention. However, motivation and engagement
levels, as measured by the Intrinsic Motivation
Inventory (IMI), did not show significant differences
between the two groups, indicating that while sound-
based interactions improved learning outcomes, they
did not intrinsically increase student motivation
beyond traditional input methods. These results
emphasize that while multimodal interactions can
optimize learning efficiency, their ability to enhance
motivation depends on additional factors such as user
preference and system reliability.
Overall, this study highlights the potential of non-
verbal sound-driven interactions in educational
robotics, particularly in enhancing learning outcomes
through increased cognitive engagement. While
motivation levels remained similar across interaction
methods, the findings suggest that sound-based
responses can serve as an effective alternative to
touch-based inputs in quiz-based learning. As
educational technology advances, future research
should aim to design scalable, adaptive human-robot
interaction systems that cater to diverse learning
needs and optimize multimodal engagement
strategies in real-world educational settings.
REFERENCES
Andić, B., Maričić, M., Mumcu, F., & Flitcroft, D. (2024).
Direct and indirect instruction in educational robotics:
A comparative study of task performance per cognitive
level and student perception. Smart Learning
Environments, 11(12). https://doi.org/10.1186/s40561-
024-00298-6
Belpaeme, T., Kennedy, J., Ramachandran, A., Scassellati,
B., & Tanaka, F. (2018). Social robots for education: A
review. Science Robotics, 3(21), eaat5954.
https://doi.org/10.1126/scirobotics.aat5954
Buchem, I., Tutul, R., Bäcker, N. (2024). Same Task,
Different Robot. Comparing Perceptions of Humanoid
Robots Nao and Pepper as Facilitators of Empathy
Mapping. In: Biele, C., et al. Digital Interaction and
Machine Intelligence. MIDI 2023. Lecture Notes in
Networks and Systems, vol 1076. Springer, Cham.
https://doi.org/10.1007/978-3-031-66594-3_14
Ching, Y.-H., & Hsu, Y.-C. (2023). Educational Robotics
for Developing Computational Thinking in Young
Learners: A Systematic Review. In TechTrends (Vol.
68, Issue 3, pp. 423–434). Springer Science and
Business Media LLC. https://doi.org/10.1007/s11528-
023-00841-1
Fridin, M. (2014). Storytelling by a kindergarten social
assistive robot: A tool for constructive learning in
preschool education. Computers & Education, 70, 53–
64. https://doi.org/10.1016/j.compedu.2013.07.043
Grover, S., Bienkowski, M., Tamrakar, A., Siddiquie, B.,
Salter, D., & Divakaran, A. (2016). Multimodal
analytics to study collaborative problem solving in pair
programming. In Proceedings of the Sixth International
Conference on Learning Analytics & (pp. 516–