
camera without ir illumination. In 2008 19th Interna-
tional Conference on Pattern Recognition, pages 1–4.
IEEE.
Chen, Z. and Shi, B. E. (2018). Appearance-based gaze es-
timation using dilated-convolutions. In Asian Confer-
ence on Computer Vision, pages 309–324. Springer.
Choi, I.-H., Tran, T. B. H., and Kim, Y.-G. (2016). Real-
time categorization of driver’s gaze zone and head
pose using the convolutional neural network. In Pro-
ceedings of HCI Korea, pages 417–422.
Ewaisha, M., El Shawarby, M., Abbas, H., and Sobh, I.
(2020). End-to-end multitask learning for driver gaze
and head pose estimation. Electronic Imaging, 32:1–
6.
George, A. and Routray, A. (2016). Real-time eye gaze di-
rection classification using convolutional neural net-
work. In 2016 International Conference on Signal
Processing and Communications (SPCOM), pages 1–
5. IEEE.
Hu, J., Shen, L., and Sun, G. (2018). Squeeze-and-
excitation networks. In Proceedings of the IEEE con-
ference on computer vision and pattern recognition,
pages 7132–7141.
Jesorsky, O., Kirchberg, K. J., and Frischholz, R. W. (2001).
Robust face detection using the hausdorff distance. In
Audio-and Video-Based Biometric Person Authentica-
tion: Third International Conference, AVBPA 2001
Halmstad, Sweden, June 6–8, 2001 Proceedings 3,
pages 90–95. Springer.
Jetley, S., Lord, N. A., Lee, N., and Torr, P. H. (2018). Learn
to pay attention. arXiv preprint arXiv:1804.02391.
Karmi, R., Rahmany, I., and Khlifa, N. (2024). Gaze esti-
mation using convolutional neural networks. Signal,
Image and Video Processing, 18(1):389–398.
Konrad, R., Shrestha, S., and Varma, P. (2016). Near-Eye
Display Gaze Tracking Via Convolutional Neural Net-
works. Standford University: Standford, CA. Techni-
cal report, USA, Tech. Rep.
Milborrow, S., Morkel, J., and Nicolls, F. (2010). The muct
landmarked face database. Pattern recognition asso-
ciation of South Africa, 201(0):535.
Omori, Y. and Shima, Y. (2020). Image augmentation for
eye contact detection based on combination of pre-
trained alex-net cnn and svm. J. Comput., 15(3):85–
97.
Park, S., Zhang, X., Bulling, A., and Hilliges, O. (2018).
Learning to find eye region landmarks for remote gaze
estimation in unconstrained settings. In Proceedings
of the 2018 ACM symposium on eye tracking research
& applications, pages 1–10.
Poulopoulos, N. and Psarakis, E. Z. (2022). Deeppupil net:
Deep residual network for precise pupil center local-
ization. In VISIGRAPP (5: VISAPP), pages 297–304.
Rahmany, I., Guetari, R., and Khlifa, N. (2018). A fully
automatic based deep learning approach for aneurysm
detection in dsa images. In 2018 IEEE international
conference on image processing, applications and sys-
tems (IPAS), pages 303–307. IEEE.
Smith, B. A., Yin, Q., Feiner, S. K., and Nayar, S. K.
(2013). Gaze locking: passive eye contact detection
for human-object interaction. In Proceedings of the
26th annual ACM symposium on User interface soft-
ware and technology, pages 271–280.
Sun, L., Liu, Z., and Sun, M.-T. (2015). Real time gaze es-
timation with a consumer depth camera. Information
Sciences, 320:346–360.
Vaswani, A. (2017). Attention is all you need. arXiv
preprint arXiv:1706.03762.
Venkateswarlu, R. et al. (2003). Eye gaze estimation from
a single image of one eye. In Proceedings Ninth IEEE
International Conference on Computer Vision, pages
136–143. IEEE.
Villanueva, A., Ponz, V., Sesma-Sanchez, L., Ariz, M.,
Porta, S., and Cabeza, R. (2013). Hybrid method
based on topography for robust detection of iris cen-
ter and eye corners. ACM Transactions on Multime-
dia Computing, Communications, and Applications
(TOMM), 9(4):1–20.
Wang, K. and Ji, Q. (2017). Real time eye gaze tracking
with 3d deformable eye-face model. In Proceedings
of the IEEE International Conference on Computer
Vision, pages 1003–1011.
Wang, K. and Ji, Q. (2018). 3d gaze estimation with-
out explicit personal calibration. Pattern Recognition,
79:216–227.
Wang, X., Girshick, R., Gupta, A., and He, K. (2018). Non-
local neural networks. In Proceedings of the IEEE
conference on computer vision and pattern recogni-
tion, pages 7794–7803.
Wood, E. and Bulling, A. (2014). Eyetab: Model-based
gaze estimation on unmodified tablet computers. In
Proceedings of the symposium on eye tracking re-
search and applications, pages 207–210.
Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhudi-
nov, R., Zemel, R., and Bengio, Y. (2015). Show, at-
tend and tell: Neural image caption generation with
visual attention. In International conference on ma-
chine learning, pages 2048–2057. PMLR.
Enhancing Appearance-Based Gaze Estimation Through Attention-Based Convolutional Neural Networks
23