Table 2: Average f (1) score for the cross validation test. T stands for training and V stands for validation tests. The parameters
used are shown in the table.
Distance k k
words
th
templates
th
sd
T (µ, σ) V (µ, σ)
Euclidean 200 75 0.55 2.2 (0.61, 0.09) (0.43, 0.34)
DTW 200 75 0.55 1.9 (0.68, 0.08) (0.59, 0.32)
the approach by computing the f (1) score. Table 2
summarizes the results of the cross-validation test
As it can be seen from Table 2, using the Eu-
clidean distance for comparing sequences gives poor
results in the validation datasets. As expected, using
the Dynamic Time Warping distance yields better re-
sults since DTW is more appropriate for comparing
sequences given that it allows a non-uniform align-
ment between the two sequences being compared.
6 CONCLUSIONS
This paper presents a method for detecting gesture
templates in a semi-supervised manner. The experi-
ments demonstrated that using Dynamic Time Warp-
ing as the distance measure for the k-medoids algo-
rithm gave the best results when spotting gestures.
Our results for the cross-validation test are compa-
rable to the ones obtained by Sagha et al. (Sagha
et al., 2011). Both contributions use a window of
16 samples overlapped 50% and the average as fea-
ture. The main difference between both contributions
is that Sagha et al. use the information in a single
window for inferring the gestures, whereas we use the
information in a sequence of feature values from mul-
tiple adjacent windows. Amongst the tested classi-
fiers, they found that 1-NN has the best performance
with an average f (1) score of 0.53 on subject 1. It
is important to note that they use all the upper body
sensors while we only use one sensor located on the
right forearm (RLA). We obtained a validation per-
formance of 0.59 over subject 1, which is very close
to the best results found by Sagha et al. Moreover,
our approach has the advantage of using less sensors
and of not requiring the whole dataset in memory to
classify new gestures.
Finally, by requiring ground truth only for the
template selection stage (but not to find the candi-
dates), our approach for template discovery could be
used in a system that can create its group of templates
incrementally. For example, a system which asks for
user annotation when a previously unseen template
is discovered. The method is therefore more flexible
than a fully supervised method where all the possible
gestures would have to be defined beforehand.
ACKNOWLEDGEMENTS
The authors would like to thank Daniel Roggen and
the team of the wearable computing laboratory at
ETHZ for their valuable input. This work is part of
the project SmartDays founded by the Hasler Foun-
dation.
REFERENCES
Berndt, D. J. and Clifford, J. (1994). Using Dynamic Time
Warping to Find Patterns in Time Series. In Proc. of
KDD-94: AAAI Workshop on Knowledge Discovery
in Databases, pages 359–370.
Frank, A. and Asuncion, A. (2010). UCI machine learning
repository.
Kaufman, L. and Rousseeuw, P. J. (1987). Clustering by
means of medoids. Statistical Data Analysis Based on
the L
1
-Norm and Related Methods, pages 405–416.
Ko, M. H. et al. (2005). Online context recognition in multi-
sensor systems using dynamic time warping. In Proc.
of the International Conference on Intelligent Sensors,
Sensors Networks and Information Processing., pages
283 – 288.
Lara, O. D. et al. (2012). Centinela: A human activ-
ity recognition system based on acceleration and vi-
tal sign data. Pervasive and Mobile Computing,
8(5):717–729.
MacQueen, J. B. (1967). Some methods for classification
and analysis of multivariate observations. In Proc. of
the fifth Berkeley Symposium on Mathematical Statis-
tics and Probability, volume 1, pages 281–297.
Mathie, M. et al. (2003). Detection of daily physical ac-
tivities using a triaxial accelerometer. Medical and
Biological Engineering and Computing, 41:296–301.
Rijsbergen, C. J. V. (1979). Information Retrieval.
Butterworth-Heinemann, 2nd edition.
Roggen, D. et al. (2010). Collecting complex activity
datasets in highly rich networked sensor environ-
ments. In 2010 Seventh International Conference on
Networked Sensing Systems, pages 233–240.
Sagha, H. et al. (2011). Benchmarking classification tech-
niques using the Opportunity human activity dataset.
In IEEE International Conference on Systems, Man,
and Cybernetics.
Stiefmeier, T. et al. (2008). Wearable activity tracking in car
manufacturing. Pervasive Computing, IEEE, 7(2):42–
50.
ICPRAM2013-InternationalConferenceonPatternRecognitionApplicationsandMethods
576