head in order to process information in real time.
Our approach uses a single representative gesture
sample to automatically infer a number of control
points capturing the characteristic parts of the ges-
ture. By offering an external representation of these
control points, developers can visualise and further
refine these points. Implicit support for overlapping
submatches, relaxed spatiotemporal operators and ad-
ditional programming constructs such as negation and
user-defined conditions are key factors to ease the
gesture spotting development. This includes the op-
timisation for a high recall, precision or processing
performance based on the application scenario.
The manual refinement of gesture rules helps to
achieve better results in the gesture spotting process.
By automatically inferring m control points from a
single gesture sample and compiling them into an
extensible declarative rule, we support gesture de-
velopers in obtaining the intended continuous ges-
ture recognition results. Inspired by mathematical
line simplification schemes such as B-spline curve fit-
ting (Cham and Cipolla, 1999), we plan to improve
the current angle-based control point computation.
Given the use of expert knowledge, we plan to
provide a graphical tool for three-dimensional tra-
jectories based on ideas of Holz and Feiner (Holz
and Feiner, 2009), where relaxed selection techniques
can be annotated and manipulated graphically to ease
the development process. While Holz and Feiner fo-
cussed on creating an interface for time series graphs
with a single dimension, our graphical gesture devel-
opment tool will address at least three dimensions.
As highlighted in Figure 5, the angle-based con-
trol point inferring technique is able to extract char-
acteristic points from a sample trajectory. However,
in this specific case, the control point cA is not opti-
mal and might negatively influence the spotting per-
formance. Another limitation of our current imple-
mentation is the lack of scale invariance. We can also
not choose between a sub- or supergesture spotting.
This application-dependent problem can be solved in
the post-classification process, while the gesture spot-
ting phase should focus on a high recall.
Our main goal was to improve the spotting of po-
tential gestures in continuous data streams. By only
requiring a single gesture sample and due to the possi-
bility to programmatically refine the spotting process
by loosening or tightening spatial and temporal con-
straints, we distinguish ourselves from existing spot-
ting solutions. The external declarative representation
of inferred control points has shown to be beneficial
and complementary to programming constructs such
as spatiotemporal operators, negation, user-defined
functions and the invocation of coupled recognisers.
Last but not least, due to the use of an efficient incre-
mental evaluation engine the computational overhead
of our gesture spotting approach is minimal.
ACKNOWLEDGEMENTS
The work of Lode Hoste is funded by an IWT doctoral
scholarship.
REFERENCES
Alon, J., Athitsos, V., Yuan, Q., and Sclaroff, S. (2009).
A Unified Framework for Gesture Recognition and
Spatiotemporal Gesture Segmentation. IEEE Trans-
actions on Pattern Analysis and Machine Intelligence,
31(9).
Cham, T.-J. and Cipolla, R. (1999). Automated B-Spline
Curve Representation Incorporating MDL and Error-
Minimizing Control Point Insertion Strategies. IEEE
Transactions on Pattern Analysis and Machine Intel-
ligence, 21(1).
Darrell, T. and Pentland, A. (1993). Space-Time Gestures.
In Proceedings of CVPR 1993, New York, USA.
Elmezain, M., Al-Hamadi, A., and Michaelis, B. (2009).
Hand Gesture Spotting Based on 3D Dynamic Fea-
tures Using Hidden Markov Models. Signal Process-
ing, Image Processing and Pattern Recognition, 61.
Forgy, C. L. (1982). Rete: A Fast Algorithm for the Many
Pattern/Many Object Pattern Match Problem. Artifi-
cial Intelligence, 19(1).
Holden, E.-J., Lee, G., and Owens, R. (2005). Australian
Sign Language Recognition. Machine Vision and Ap-
plications, 16.
Holz, C. and Feiner, S. (2009). Relaxed Selection Tech-
niques for Querying Time-Series Graphs. In Proceed-
ings of UIST 2009, Victoria, Canada.
Just, A. (2006). Two-Handed Gestures for Human-
Computer Interaction. PhD thesis,
´
Ecole Polytech-
nique F
´
ed
´
erale de Lausanne. Diss No. 3683.
Kadous, M. W. (1999). Learning Comprehensible Descrip-
tions of Multivariate Time Series. In Proceedings of
ICML 1999, Bled, Slovenia.
Kelly, D., McDonald, J., and Markham, C. (2011). Recog-
nition of Spatiotemporal Gestures in Sign Language
Using Gesture Threshold HMMs. Machine Learning
for Vision-Based Motion Analysis.
Lee, H.-K. and Kim, J. H. (1999). An HMM-based Thresh-
old Model Approach for Gesture Recognition. IEEE
Transactions on Pattern Analysis and Machine Intel-
ligence, 21(10).
Pittman, J. A. (1991). Recognizing Handwritten Text. In
Proceedings of CHI 1991, New Orleans, USA.
Rubine, D. (1991). Specifying Gestures by Example. In
Proceedings of SIGGRAPH 1991, Las Vegas, USA.
DeclarativeGestureSpottingusingInferredandRefinedControlPoints
149