There is a lot of room to improve the actual
accuracy of the system - we might be able to use
more sophisticated face detection algorithms or
classifiers, and even use techniques of hallucinating
exemplars from the existing data, to make the
system more robust to noise and illumination
conditions. Nevertheless, we can strongly declare
that our objective in this paper has been reached —
it is technically possible to make a real-time robust
face recognition system running exclusively on the
low-performance hardware of the smartwatch.
Additionally, in terms of user interaction, the
experiment was important to show usability and
ergonomic issues that need to be addressed before
people with actual visual impairments are involved.
The feedback that indicates a face is being framed
needs more work so that it becomes a more precise
clue as to where the user needs to point the
smartwatch’s camera. This is important not only to
allow the system to be used as an assistive
technology, but also to alleviate the fatigue issue
reported by the participants. Other potential place
for future enhancement concerns the feedback
interface to get data from people´s faces, which still
must be made accessible for use by blind and low-
vision people.
Finally, we propose challenges for future work,
including wearable systems for objects recognition,
textual information recognition (e.g. signs, symbols)
and a gesture recognition like Porzi et al. (2013), but
processed within the smartwatch itself. Furthermore,
we will conduct experiments to better analyze the
system's energy consumption. Also, experiments
with visually impaired users will be used to further
evaluate and improve the system as an assistive
device.
ACKNOWLEDGEMENTS
The authors wish to express their gratitude to all the
volunteers who participated in the experiments in
this study, and also for Samsung Research that
loaned the hardware equipment. LSBN receives a
Ph.D. fellowship from CNPq (grant #141254/2014-
9). VRMLM receives a Ph.D. fellowship from
CAPES (grant #01-P-04554/2013). MCCB, ARR
and SKG receives a Productivity Research
Fellowship from CNPq (grants #308618/2014-9,
#304352/2012-8 and #308882/2013-0, respectively).
This work is part of a project that was approved by
Unicamp Institutional Review Board CAAE
31818014.0.0000.5404.
REFERENCES
Ahonen, T., Hadid, A., and Pietikainen, M. (2006). Face
description with local binary patterns: Application to
face recognition. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 28(12):2037–
2041.
Astler, D., Chau, H., Hsu, K., Hua, A., Kannan, A., Lei, L.
Nathanson, M., Paryavi, E., Rosen, M., Unno, H.,
Wang, C., Zaidi, K., Zhang, X., and Tang, C. (2011).
Increased accessibility to nonverbal communication
through facial and expression recognition technologies
for blind/visually impaired subjects. In The
Proceedings of the 13th International ACM
SIGACCESS Conference on Computers and
Accessibility, pages 259–260.
Belhumeur, P., Hespanha, J., and Kriegman, D. (1997).
Eigenfaces vs. fisherfaces: recognition using class
specific linear projection. IEEE Transactions on
Pattern Analysis and Machine Intelligence,
19(7):711–720.
Chen, X., Flynn, P., and Bowyer, K. (2003). PCA-based
face recognition in infrared imagery: baseline and
comparative studies. In Proceedings of the IEEE
International Workshop on Analysis and Modeling of
Faces and Gestures, pages 127–134.
Cover, T., and Hart, P.: Nearest neighbor pattern
classification. (1967). IEEE Transactions on
Information Theory, 13(1):21–27.
Dalal, N. and Triggs, B. (2005). Histograms of oriented
gradients for human detection. In IEEE Computer
Society Conference on Computer Vision and Pattern
Recognition, pages 886–893.
FreevoxTouch (2014). The only smart watch in the world
for the visually impaired. Available:
http://myfreevox.com/en/.
Fusco, G., Noceti, N., and Odone, F. (2012). Combining
retrieval and classification for real-time face
recognition. In 2013 10th IEEE International
Conference on Advanced Video and Signal Based
Surveillance, pages 276–281.
Gordon, G. (1991). Face recognition based on depth maps
and surface curvature. In SPIE1570, Geometric
methods in Computer Vision, pages 234–247.
Hadid, A., Pietikainen, M., and Li, S. (2007). Learning
personal specific facial dynamics for face recognition
from videos. In Analysis and Modeling of Faces and
Gestures: Lecture Notes in Computer Science 4778,
pages 1–15.
Kistler, D. and Wightman, F. (1992). A model of head-
related transfer functions based on principal
components analysis and minimum-phase
reconstruction. Journal of the Acoustical Society of
America, 91(3):1637–1647.
Kramer, K., Hedin, D., and Rolkosky, D. (2010).
Smartphone based face recognition tool for the blind.
In 32nd Annual International Conference of the IEEE
Engineering in Medicine and Biology Society, pages
4538–4541.
Krishna, S., Little, G., Black, J., and Panchanathan, S.
AWearableFaceRecognitionSystemBuiltintoaSmartwatchandtheVisuallyImpairedUser
11