
sity for well-curated data in future research. It is more
effective to train a model capable of accurately pre-
dicting specific situations and transfer the alert logic
to the ADAS. In other words, a model should be
trained on images where the class is clearly distin-
guishable, allowing it to achieve high performance in
classifying those images. Meanwhile, in the ADAS,
this model’s predictions can be combined with an al-
gorithm that determines when to alert the driver. This
method allows the model to focus on what it does
best: analyzing images and predicting states, while
the ADAS evaluates when it is appropriate to alert the
driver based on a combination of additional factors,
such as the duration of the detected drowsiness, vehi-
cle’s speed and other environmental parameters. Fur-
thermore, XAI has provided significant insights into
the model’s behavior, which would have been impos-
sible to ascertain solely by examining the accuracy
metrics. For instance, there are cases where images
labeled as drowsy were classified as awake due to the
absence of actual signs of fatigue. This finding re-
veals that these misclassifications are a consequence
of incorrectly labeled images, underscoring the criti-
cal importance of accurate dataset labeling. This fa-
cilitates the drawing of crucial conclusions, as those
mentioned before in this section, and paves the way
for future research avenues in driver monitoring.
In conclusion, we believe that the continued ap-
plication of XAI will be crucial for not only foster-
ing transparency and building trust in AI systems but
also for refining model behavior. By revealing the in-
fluence of specific features on predictions, XAI em-
powers practitioners to make informed adjustments,
ultimately leading to more reliable and effective deep
learning applications. Despite these advances, much
remains to be done to fully achieve interpretability
in CNN models. This research represents a prelim-
inary exploration, and we anticipate that continued
efforts will bring further clarity and refinement, ad-
vancing the development of transparent and reliable
AI applications for emotion and drowsiness detection
in ADAS. Drowsiness detection remains an open and
challenging problem but it is clear that deep learning
techniques have significant potential for incorporation
into ADAS, facilitating innovations that can greatly
enhance roadway safety and reduce accidents.
ACKNOWLEDGEMENTS
This research was funded by
MCIN/AEI/10.13039/501100011033 grant numbers
PID2022-140554OB-C32, PDC2022-133684-C31.
REFERENCES
Ghoddoosian, R., Galib, M., and Athitsos, V. (2019). A
realistic dataset and baseline temporal model for early
drowsiness detection.
Goodfellow, I. J., Erhan, D., Carrier, P. L., Courville, A.,
Mirza, M., Hamner, B., Cukierski, W., Tang, Y.,
Thaler, D., Lee, D.-H., Zhou, Y., Ramaiah, C., Feng,
F., Li, R., Wang, X., Athanasakis, D., Shawe-Taylor,
J., Milakov, M., Park, J., Ionescu, R., Popescu, M.,
Grozea, C., Bergstra, J., Xie, J., Romaszko, L., Xu,
B., Chuang, Z., and Bengio, Y. (2013). Challenges
in representation learning: A report on three machine
learning contests.
Jeon, M. (2016). Don’t cry while you’re driving: Sad driv-
ing is as bad as angry driving. International Journal
of Human–Computer Interaction, 32(10):777–790.
Jeon, M., Roberts, J., Raman, P., Yim, J.-B., and Walker,
B. N. (2011). Participatory design process for an in-
vehicle affect detection and regulation system for var-
ious drivers. In The Proceedings of the 13th Inter-
national ACM SIGACCESS Conference on Computers
and Accessibility, ASSETS ’11, page 271–272, New
York, USA. Association for Computing Machinery.
Kapishnikov, A., Bolukbasi, T., Vi
´
egas, F., and Terry, M.
(2019). Xrai: Better attributions through regions.
Lorente, M. P. S., Lopez, E. M., Florez, L. A., Espino, A. L.,
Mart
´
ınez, J. A. I., and de Miguel, A. S. (2021). Ex-
plaining deep learning-based driver models. Applied
Sciences, 11(8).
Lundberg, S. and Lee, S.-I. (2017). A unified approach to
interpreting model predictions.
Mag
´
an, E., Sesmero, M. P., Alonso-Weber, J. M., and San-
chis, A. (2022). Driver drowsiness detection by apply-
ing deep learning techniques to sequences of images.
Applied Sciences, 12(3).
Ribeiro, M. T., Singh, S., and Guestrin, C. (2016). ”why
should i trust you?”: Explaining the predictions of any
classifier.
Sheykhivand, S., Mousavi, Z., Rezaii, T. Y., and Farzamnia,
A. (2020). Recognizing emotions evoked by music
using cnn-lstm networks on eeg signals. IEEE Access,
8:139332–139345.
Tamanani, R., Muresan, R., and Al-Dweik, A. (2021). Esti-
mation of driver vigilance status using real-time facial
expression and deep learning. IEEE Sensors Letters,
5(5):1–4.
Verma, B. and Choudhary, A. (2018). A framework for
driver emotion recognition using deep learning and
grassmann manifolds. pages 1421–1426.
Zepf, S., Hernandez, J., Schmitt, A., Minker, W., and Pi-
card, R. W. (2020). Driver emotion recognition for
intelligent vehicles: A survey. ACM Comput. Surv.,
53(3).
˚
Akerstedt, T. and Gillberg, M. (1990). Subjective and ob-
jective sleepiness in the active individual. The Inter-
national journal of neuroscience, 52:29–37.
Analyzing Model Behavior for Driver Emotion Recognition and Drowsiness Detection Using Explainable Artificial Intelligence
341