For the evidential method, we have made various
modifications on the software so that the SVM
output is automatically presented throughout the
evidential formalism (Burger, 2006). These
modifications are available on demand.
The results are presented in Table 3, as the test
accuracy of the classical voting procedure and the
default tuning of the evidential method. The
improvement in ΔPoint is worth 1.03 points and
corresponds to an avoidance of mistakes of
%AvMis = 11.11%.
Table 3: Results for experiments 1 & 2.
Evidential method
Classical
Voting
procedure
Default
(no thumb
detection)
With Thumb
Detection
Test
Accuracy
90.7% 91.8% 92.8%
The goal of the second experiment is to evaluate
the advantage of the thumb information. For that
purpose, we add the thumb information to the
evidential method. Thus, the training set is used to
set the two thresholds, which defines the possible
distance with respect to the center of palm.
However, the thumb information is not used during
the training of the SVMs as they only work on the
Hu invariants, as explained in Figure 5. The results
with and without the thumb indicator are presented
in Table 3.
Table 4: Confusion matrix for the second method on
Corpus 2, with the Thumb and NoThumb superclasses
framed together.
0 1 2 3 4 5 6 7 8
0
12 0 0 0 0 0 0 0 0
1
0 46 0 0 0 0 0 0 1
2
0 2 23 2 0 0 0 0 0
3
0 2 0 29 2 2 1 0 0
4
0 0 0 1 32 0 0 0 1
5
0 0 0 0 0 58 0 1 0
6
0 0 2 0 0 0 41 3 0
7
0 0 0 0 0 0 1 6 0
8
0 0 0 0 0 0 0 0 23
The evidential method that uses the thumb
information provides an improvement of 2.06 points
with respect to the classical voting procedure, which
corresponds to an avoidance of 22.22% of the
mistakes. Table 4 presents the corresponding
confusion matrix for the test set: Hand shape 3 is
often misclassified into other hand shapes, whereas,
on the contrary, hand shape 1 and 7 gather a bit
more misclassification from other hand shapes.
Moreover, there are only three mistakes between
THUMB and NO_THUMB super-classes.
6 CONCLUSION
In this paper, we propose to apply a belief-based
method for SVM fusion to hand shape recognition.
Moreover, we integrate it in a wider classification
scheme which allows taking into account other
sources of information, by expressing them in the
Belief Theories formalism. The results are better
than with the classical methods (more than 1/5 of the
mistakes are avoided) and the absolute accuracy is
high with respect to the number of classes involved.
ACKNOWLEDGEMENTS
This work is the result of a cooperation supported by
SIMILAR, European Network of Excellence
(www.similar.cc).
REFERENCES
Aran, O., Keskin, C., Akarun, L., 2005. Sign Language
Tutoring Tool, EUSIPCO’05, Antalya, Turkey.
Burger , T., Aran, O., and Caplier, A., 2006. Modeling
hesitation and conflict: A belief-based approach, In
Proc. ICMLA.
Boser, B., Guyon, I., and Vapnik, V., 1995. A training
algorithm for optimal margin classifiers, In Proc. Fifth
Annual Workshop on Computational Learning Theory.
Capelle, A. S., Colot, O., Fernandez-Maloigne, C., 2004.
Evidential segmentation scheme of multi-echo MR
images for the detection of brain tumors using
neighborhood information. Information Fusion,
Volume 5, Number 3, pages 203-216.
Caplier, A., Bonnaud, L., Malassiotis, S., and Strintzis,
M., 2004. Comparison of 2D and 3D analysis for
automated Cued Speech gesture recognition,
SPECOM, St Petersburg, Russia.
Chang , C.-C., and Lin, C.-J., 2001. LIBSVM: a library
for support vector machines. Software available at
http://www.csie.ntu.edu.tw/~cjlin/libsvm
Cornett, R. O., 1967. Cued Speech, American Annals of
the Deaf.
Cortes, C., and Vapnik, V., 1995. Support-vector network,
Machine Learning 20, 273–297.
Denoeux, T., 2000. Modeling vague beliefs using fuzzy-
valued belief structures. Fuzzy Sets and Systems,
116(2):167-199.
CUED SPEECH HAND SHAPE RECOGNITION - Belief Functions as a Formalism to Fuse SVMs & Expert Systems
11