Measurement and Measurement Signal Processing
of Graz University of Technology (Austria)
demonstrated the good performance of our classifier.
This paper is organized as fellows: in section 2, we
describe the classifier construction steps, beginning
by features extraction and ending by validation, in
section 3, we present some experimental results and
we end by conclusions in section 4.
2 CLASSIFIER CONSTRUCTION
We first recall that MPEG-7 visual descriptors
(ISO\IEC, 2001) standardized for the image content
description are a compressed description of image
features, which are represented in terms of primitive
image features such as color, texture, and shape of
the image. Among the MPEG-7 visual descriptors,
we have chosen the edge histogram descriptor
(EHD) (Won et. al, 2002) as features of images to be
classified in two classes (Face/Non-Face).
To construct our classifier, an image data base is
needed for its training and testing. In our work we
used a data set of 470 images among which 214
contain human faces. Whilst for features (image
descriptors) extraction, we used the MPEG-7 XM
module (Manjunath et. al, 2001), to get the Edge
Histogram Descriptor.
2.1 Edge Histogram Descriptor (EHD)
The EHD of the MPEG-7 visual descriptors
represents the distribution of five edge types, namely
vertical, horizontal, 45- degree diagonal, 135-degree
diagonal, and non-edge types (ISO\IEC, 2001),
(Won et .al, 2002). The distribution of five edge
types is represented by 16 local edge histograms.
Each local histogram is generated from each sub-
image. A sub-image is a non-overlapping 4x4
partition of the given image space. That is, an image
is divided into non-overlapping 4x4 sub-images.
Then, each sub-image is used as a basic region to
generate an edge histogram, which consists of five
bins with vertical, horizontal, 45-degree diagonal,
135-degree diagonal and non-directional edge types.
Note that the image block may or may not have an
edge in it. If there is an edge in the block, the
counter of the corresponding edge type is increased
by one. Otherwise, the image block has monotonous
gray levels and no histogram bin is increased. After
examining all image blocks in the sub-image, the 5-
bin values are normalized by the total number of
blocks in the sub-image. Thus the sum of the
normalized five bins is not necessarily 1. Finally, the
normalized bin values are quantized for the binary
representation. Since there are 16 (4 x 4) sub-
images, each image yields an edge histogram with a
total of 80 (16 sub-images x 5 bins/sub-image) bins.
These normalized and quantized 80 bins constitute
the EHD of the
MPEG-7. That is, arranging edge
histograms for the sub-images in the raster scan
order (bloc order is done according to that of lines),
16 local histograms are concatenated to have an
integrated histogram with 80 (16x5) bins.
Once EHD descriptors have been recovered,
Independent Component Analysis (ICA) is then
applied to obtain independent parameters and keep
just pertinent information. We have fixed a
percentage of retained information in the
prewhitening step by Principle Component Analysis
(PCA).
For the classification step, we tried many
classifiers (Bayes classifiers, support vectors
machine, K-nearest neighbors,…) but we limit our
selves here to a brief description of classifiers that
gave satisfactory results: the Nearest Mean
Classifier (NMC) and Linear Fisher Discrimenant
Classifier (LFDC) . For the tests of classification
scores, the cross-validation (mainly the leave one
out) strategy has been adopted. All our tests were
performed with the Matlab Toolbox PRTools4
(http://prtools.org/).
2.2 Nearest Mean Classifier
Nearest mean classifier calculates the centers of in-
class and out-class training samples and then assigns
the upcoming samples to the closest center. This
classifier gives two distance values as output and
should be modified to produce a posterior
probability value. A common method used for K-
NN classifiers can be utilized (Arlandis et. al, 2002).
According to this method, distance values are
mapped to posterior probabilities by the formula:
(1)
where W
i
refers to the i
th
class (i=1,2.), d
mi
and d
mj
are distances from the i
th
and j
th
class means,
respectively. In addition, a second measure
recomputes the probability values below a given
certainty threshold by using the formula (Arlandis
et. al, 2002):
(2)
∑
=
=
2
1
1
/
1
)/(
j
mjmi
i
dd
xWP
N
xWP
i
i
=)/(
VISAPP 2008 - International Conference on Computer Vision Theory and Applications
330