The proposed system was able to correctly clas-
sify the input gestures in 83.75% of the cases, cor-
responding to 134 positives out of 160 inputs. Such
a result shows that compared to the standalone fuzzy
recognizer, representing the set of possible commands
and queries as a language increases the performance
of the system.
5 CONCLUSIONS
In this work we presented a system for the manage-
ment of an office environments by means of an unob-
trusive sensing device, i.e., the Kinect. Such a sensor
is equipped with a number of input/output devicesthat
make it possible to sense the user and its interaction
with the surrounding environment. We considered a
scenario where the whole environment is permeated
with small pervasive sensor devices, for this reason
the Kinect is coherently connected to a miniature fan-
less computer with reduced computation capabilities.
The control of the actuators of the AmI system
(e.g., air-conditioning, curtain and rolling shutter) is
performed by the Kinect by recognizing some simple
gestures (i.e., open/closed hands) opportunely struc-
tured by means of a grammar.
Once the hand of the user has been detected, some
local processing is done using RGB-D data and the
obtained hand region is described according to its
roundness. Such a descriptor is verified by means of
a fuzzy procedure that predicts the state of the hand
with a certain level of accuracy. Each state (i.e., open
or closed) represents a symbol of a grammar that de-
fines the corresponding commands for the actuators.
The construction of a real prototype of the moni-
toring and controlling system allowed for exhaustive
testing of the proposed method. Experimental results
showed that the system is able to perform efficiently
on a miniature computer while maintaining a high
level of accuracy both in terms of image analysis and
gesture recognition.
Although the effectiveness of the system has been
evaluated considering only two different gestures, the
addition of more symbols is straightforward. As fu-
ture work we may conceivably consider some easily
recognizable gestures involving the use of both hands
and their relative position in order to allow the defini-
tion of a more complex grammar.
ACKNOWLEDGEMENTS
This work is supported by the SMARTBUILDINGS
project, funded by POR FESR SICILIA 2007-2013.
REFERENCES
Aho, A., Lam, M., Sethi, R., and Ullman, J. (2007). Compil-
ers: principles, techniques, and tools, volume 1009.
Pearson/Addison Wesley.
Alcal´a, R., Casillas, J., Cord´on, O., and Herrera, F.
(2003). Linguistic modeling with weighted double-
consequent fuzzy rules based on cooperative co-
evolutionary learning. Integr. Comput.-Aided Eng.,
10(4):343–355.
Borenstein, G. (2012). Making Things See: 3D Vision With
Kinect, Processing, Arduino, and MakerBot. Make:
Books. O’Reilly Media, Incorporated.
Cho, J.-S. and Park, D.-J. (2000). Novel fuzzy logic control
based on weighting of partially inconsistent rules us-
ing neural network. Journal of Intelligent Fuzzy Sys-
tems, 8(2):99–110.
De Paola, A., Cascia, M., Lo Re, G., Morana, M., and
Ortolani, M. (2012a). User detection through multi-
sensor fusion in an ami scenario. In Information Fu-
sion (FUSION), 2012 15th International Conference
on, pages 2502 –2509.
De Paola, A., Gaglio, S., Lo Re, G., and Ortolani, M.
(2012b). Sensor9k: A testbed for designing and exper-
imenting with WSN-based ambient intelligence appli-
cations. Pervasive and Mobile Computing. Elsevier,
8(3):448–466.
Gulshan, V., Lempitsky, V., and Zisserman, A. (2011). Hu-
manising grabcut: Learning to segment humans us-
ing the kinect. In Computer Vision Workshops (ICCV
Workshops), 2011 IEEE International Conference on,
pages 1127 –1133.
Kean, S., Hall, J., and Perry, P. (2011). Meet the Kinect:
An Introduction to Programming Natural User Inter-
faces. Apress, Berkely, CA, USA, 1st edition.
Lai, K., Bo, L., Ren, X., and Fox, D. (2011). Sparse dis-
tance learning for object recognition combining rgb
and depth information. In Robotics and Automa-
tion (ICRA), 2011 IEEE International Conference on,
pages 4007 –4013.
Morana, M., De Paola, A., Lo Re, G., and Ortolani, M.
(2012). An Intelligent System for Energy Efficiency
in a Complex of Buildings. In Proc. of the 2nd IFIP
Conference on Sustainable Internet and ICT for Sus-
tainability.
Raheja, J., Chaudhary, A., and Singal, K. (2011). Track-
ing of fingertips and centers of palm using kinect. In
Computational Intelligence, Modelling and Simula-
tion (CIMSiM), 2011 Third International Conference
on, pages 248 –252.
Xia, L., Chen, C.-C., and Aggarwal, J. (2011). Hu-
man detection using depth information by kinect. In
Computer Vision and Pattern Recognition Workshops
(CVPRW), 2011 IEEE Computer Society Conference
on, pages 15 –22.
PECCS2013-InternationalConferenceonPervasiveandEmbeddedComputingandCommunicationSystems
34