of the latter is performed by determining the closest
template gesture.
Our method has been tested with the accelerome-
ter data of the MHAD dataset (Ofli et al., 2013) using
the Leave-One-Out cross validation on the 12 sub-
jects of the set, and compared to other state-of-the-
art classification approaches (SVM, MLP and CNN).
It achieves a 85.30% average accuracy, and outper-
forms other real-time gesture classification models,
without having any need -unlike the other approaches-
to set a time-window parameter which could be con-
strained by the constraints of an actual real-time im-
plementation, or not be suitable for the recognition
of very short gestures. The results obtained show
that most misclassifications of our method concern
gestures which share common submovements (e.g.
sitting-down, standing-up). The high accuracies ob-
tained for the other gestures of the dataset, no mat-
ter the subject performing them, or the way they are
performed, indicate that our method is robust to vari-
ations in execution of the gestures.
There are still some points on which our method
could be improved or developed further though. Pre-
liminary experiments carried out on a dataset of hand
gestures using accelerometers (uWave, (Liu et al.,
2009)) seem to confirm the fact that our method
struggles to differenciate gestures similar in terms of
sensor values and variations, as it provides accura-
cies much lower than other tested state-of-the-art ap-
proaches (SVM and Neural Networks) there. Future
works will focus on finding axis of improvements to
address this issue: in particular finding more rele-
vant state features to be tracked by the particle filter,
testing the classification problem with additional 1D
temporal sensor data, and analyze the possibility to
attribute an importance weight to each sensor in our
observation model to favour the recognition of some
gestures.
ACKNOWLEDGEMENTS
Research and development activities leading to this
article have been supported by the German Research
Foundation (DFG) as part of the research training
group GRK 1564 ”Imaging New Modalities”, and
the German Federal Ministry of Education and Re-
search (BMBF) within the project ELISE (grant num-
ber: 16SV7512, www.elise-lernen.de).
REFERENCES
Akl, A., Feng, C., and Valaee, S. (2011). A
novel accelerometer-based gesture recognition sys-
tem. IEEE Transactions on Signal Processing,
59(12):6197–6205.
Arulampalam, M. S., Maskell, S., Gordon, N., and Clapp,
T. (2002). A tutorial on particle filters for on-
line nonlinear/non-gaussian bayesian tracking. IEEE
Transactions on Signal Processing, 50(2):174–188.
Black, M. J. and Jepson, A. D. (1998). A probabilis-
tic framework for matching temporal trajectories:
Condensation-based recognition of gestures and ex-
pressions. In Computer Vision — ECCV’98: 5th Euro-
pean Conference on Computer Vision Freiburg, Ger-
many, June, 2–6, 1998 Proceedings, Volume I”, pages
909–924.
Caramiaux, B., Montecchio, N., Tanaka, A., and Bevilac-
qua, F. (2014). Adaptive gesture recognition with vari-
ation estimation for interactive systems. ACM Trans.
Interact. Intell. Syst., 4(4):18:1–18:34.
Chudgar, H. S., Mukherjee, S., and Sharma, K. (2014). S
control: Accelerometer-based gesture recognition for
media control. In Advances in Electronics, Computers
and Communications (ICAECC), 2014 International
Conference on, pages 1–6.
Doucet, A., Godsill, S., and Andrieu, C. (2000). On se-
quential monte carlo sampling methods for bayesian
filtering. Statistics and Computing, 10(3):197–208.
Gordon, N. J., Salmond, D. J., and Smith, A. F. M. (1993).
Novel approach to nonlinear/non-gaussian bayesian
state estimation. volume 140, pages 107–113.
Isard, M. and Blake, A. (1998). Condensation - conditional
density propagation for visual tracking. International
Journal of Computer Vision, 29(1):5–28.
Liu, J., Zhong, L., Wickramasuriya, J., and Vasudevan, V.
(2009). uwave: Accelerometer-based personalized
gesture recognition and its applications. Pervasive and
Mobile Computing, 5(6):657 – 675. PerCom 2009.
Liu, J. S. and Chen, R. (1998). Sequential monte carlo
methods for dynamic systems. Journal of the Ameri-
can statistical association, 93(443):1032–1044.
Mitra, S. and Acharya, T. (2007). Gesture recognition:
A survey. IEEE Transactions on Systems, Man,
and Cybernetics, Part C (Applications and Reviews),
37(3):311–324.
Ofli, F., Chaudhry, R., Kurillo, G., Vidal, R., and Bajcsy, R.
(2013). Berkeley mhad: A comprehensive multimodal
human action database. In Applications of Computer
Vision (WACV), 2013 IEEE Workshop on, pages 53–
60.
Petitjean, F. and Ganc¸arski, P. (2012). Summarizing a set
of time series by averaging: From steiner sequence to
compact multiple alignment. Theoretical Computer
Science, 414(1):76–91.
Shoaib, M., Bosch, S., Incel, O., H.Scholten, and Havinga,
P. (2015). A survey of online activity recognition us-
ing mobile phones. Sensors, 15:2059–2085.
Real-Time Gesture Recognition using a Particle Filtering Approach
401