Authors:
Wataru Akahori
1
;
Tatsunori Hirai
2
and
Shigeo Morishima
3
Affiliations:
1
Waseda University / JST ACCEL, Japan
;
2
Komazawa University, Japan
;
3
Waseda Research Institute of Science and Engineering / JST ACCEL, Japan
Keyword(s):
Dynamic Subtitles, Eye-tracking, Region of Interest, Speaker Detection, User Experience.
Related
Ontology
Subjects/Areas/Topics:
Applications and Services
;
Computer Vision, Visualization and Computer Graphics
;
Enterprise Information Systems
;
Entertainment Imaging Applications
;
Human and Computer Interaction
;
Human-Computer Interaction
;
Image and Video Analysis
;
Multimedia Forensics
;
Visual Attention and Image Saliency
Abstract:
This paper presents a subtitle placement method that reduces unnecessary eye movements. Although methods
that vary the position of subtitles have been discussed in a previous study, subtitles may overlap the region
of interest (ROI). Therefore, we propose a dynamic subtitling method that utilizes eye-tracking data to avoid
the subtitles from overlapping with important regions. The proposed method calculates the ROI based on the
eye-tracking data of multiple viewers. By positioning subtitles immediately under the ROI, the subtitles do
not overlap the ROI. Furthermore, we detect speakers in a scene based on audio and visual information to
help viewers recognize the speaker by positioning subtitles near the speaker. Experimental results show that
the proposed method enables viewers to watch the ROI and the subtitle in longer duration than traditional
subtitles, and is effective in terms of enhancing the comfort and utility of the viewing experience.