line method failed when a slight misalignment of
the segmentation of the eye area occurred. In con-
trast, the proposed method succeeded even when the
slight misalignment occurred because the HOG fea-
ture could be extracted robustly even in case of slight
translation or rotation.
5 CONCLUSIONS
Since a Web camera is usually located outside the dis-
play while the user looks at his/her partner in the dis-
play, there is a problem that they cannot establish eye
contact with each other.
In this paper, we proposed a system for synthe-
sizing eye contact using a single camera. The pro-
posed system transformed eye areas of an user only
when the user’s gaze falls in the range that the partner
should perceive eye contact.
The training phase may impose the users a trou-
blesome task. To solve this issue, we can apply an
online gaze calibration method using click events in
daily use of a computer mouse like in (Sugano et al.,
2015) to capture the training images.
Our system runs at 5 fps for an input video with
a resolution of 1,280 × 960 pixels on a standard
consumer computer equipped with an Intel Core i7
3.59GHz CPU, and 8GB RAM. However, the system
can be faster by shrinking the input video size or par-
allelizing the process.
Our current system is not adapted for users wear-
ing glasses. Future work includes improving the gaze
classification by introductionof other features and im-
plementing the proposed method on an actual video
conferencing system.
ACKNOWLEDGEMENTS
Parts of this research were supported by MEXT,
Grant-in-Aid for Scientific Research.
REFERENCES
Anstis, S., Mayhew, J., and Morley, T. (1969). The percep-
tion of where a face or television ‘portrait’ is looking.
American J. of Psychology, 82(4):474–489.
Cortes, C. and Vapnik, V. N. (1995). Support vector net-
works. Machine Learning, 20(3):273–297.
Dalal, N. and Triggs, W. (2005). Histograms of oriented
gradients for human detection. In Proc. of the 2005
IEEE Computer Society Conf. on Computer Vision
and Pattern Recognition, volume 1, pages 886–893.
Giger, D., Bazin, J., Kuster, C., Popa, T., and Gross, M.
(2014). Gaze correction with a single webcam. In
Proc. of the 2014 IEEE Int. Conf. on Multimedia and
Expo, pages 68–72.
Kollarits, R., Woodworth, C., and Ribera, J. (1995). An eye-
contact cameras/display system for videophone appli-
cations using a conventional direct-view LCD. In Di-
gest of 1995 SID Int. Symposium, pages 765–768.
Kuster, C., Popa, T., Bazin, J., Gotsman, C., and Gross, M.
(2012). Gaze correction for home video conferencing.
ACM Trans. on Graphics, 31(6):174:1–174:6.
Muhlbach, L., Kellner, B., Prussog, A., and Romahn, G.
(1985). The importance of eye contact in videotele-
phone service. In Proc. of the 11th Int. Symposium on
Human Factors in Telecommunications, number O-4,
pages 1–8.
Saragih, J., Lucey, S., and Cohn, J. (2011). Deformable
model fitting by regularized landmark mean-shift. Int.
J. of Computer Vision, 91(3):200–215.
Solina, F. and Ravnik, R. (2011). Fixing missing eye-
contact in video conferencing systems. In Proc. of the
33rd Int. Conf. on Information Technology Interfaces,
pages 233–236.
Sugano, Y., Matsushita, Y., Sato, Y., and Koike, H. (2015).
Appearance-based gaze estimation with online cal-
ibration from mouse operations. IEEE Trans. on
Human-Machine Systems, 45(6):750–760.
Uono, S. and Hietanen, J. (2015). Eye contact perception in
the West and East: A cross-cultural study. Plos one,
10(2):e0118094.
Yang, R. and Zhang, Z. (2004). Eye gaze correction
with stereovision for video-teleconferencing. IEEE
Trans. on Pattern Analysis and Machine Intelligence,
26(6):956–960.
Yip, B. (2005). Face and eye rectification in video confer-
ence using affine transform. In Proc. of the 2005 IEEE
Int. Conf. on Image Processing, volume 3, pages 513–
516.