generalization discriminationtraining data
Figure 8: Laboratory data: the training data and the two
types of test - generalization capability and discrimination
capability.
generalization across illumination is achieved, while
enough discrimination still remained because of the
high dimensionality.
5 EXPERIMENTS AND RESULTS
To validate proposed the illumination normalization
methods, we collected data under laboratory condi-
tions
1
. We collected 10 subjects, each in indepen-
dent sessions under 3 completely different illumina-
tions. The number of images per session is 1,200.
The experiments take into consideration two impor-
tant aspects of the face verification system: discrim-
ination which is closely related to the security of the
verification system, tested by different subjects un-
der the same illumination; generalization which is
closely related to the user-friendliness of the verifi-
cation system, tested by the same subject under dif-
ferent illumination. Fig. 8 illustrates the two types
of test. The user space is trained on one session of
the user data, while the background space is trained
on three public face databases, namely the BioID
database (BioID), FERET database (FERET), and
FRGC database (FRGC).
The receiver operation characteristic (ROC) is an
indication of the system performance. It can be ob-
tained by thresholding the matching scores (in our
work likelihood ratio L) of the user data and the im-
postor data. The selection of the final threshold de-
pends on the application requirement, e.g. false ac-
cept rate or false reject rate, by taking the thresh-
old corresponding to such a operation point on the
ROC. We adapt very harsh testing protocols: the user
matching scores are calculated as the likelihood ra-
tio L of the user data in all the independent sessions
with completely different illuminations, while the im-
postor matching scores are calculated as the likeli-
1
Most publicly available database do not contain enough
number of images per user to train a user-specific space.
Our larger database is still under construction, and the data
used in this paper are available on request.
hood ratio L of all the other 9 subjects under ex-
actly the same illuminations as the training data. We
test the illumination normalization methods in 6 dif-
ferent schemes: (1) shifting and rescaling every fea-
ture vector to zero mean and unit variance (NORM1)
(2) horizontal Gaussian derivative filter (HF), fol-
lowed by NORM1; (3) original LBP filtering (LBP-
256); (4) simplified LBP filtering (LBP-9); (5) hori-
zontal Gaussian derivative filter, followed by original
LBP filtering (HF+LBP-256); (6) horizontal Gaussian
derivative filter, followed by simplified LBP filtering
(HF+LBP-9). Fig. 9 (a) shows the ROCs of the 6 illu-
mination normalization methods, along with the equal
error rates (EER) of the verification. In all the tests, a
Gaussian horizonal filter with width σ
x
= 5, σ
y
= 1 is
applied to the face images of size 100× 100. The filter
extracts fine horizontal information while discarding
vertical information.
The experiments show that when only NORM1
is applied, the verification performance is poor, indi-
cating that different illuminations make large differ-
ences across face images of the same subject. The
same is true for horizontal Gaussian derivative filter
followed by NORM1, as illumination intensities also
make large differences on the feature vectors. The
two LBP filters have better verification performance,
while horizontal Gaussian derivative filter followed
by LBP filters (especially LBP-9) yields the best ro-
bustness to illumination. This experiment setting pro-
vided a way to validate and compare these illumina-
tion normalization methods. Although the harshness
of the test puts forward high requirements on the illu-
mination normalization methods, the results in Fig. 9
(a) do illustrate the potence of our solutions. Exper-
iments on larger databases are still being done for a
more comprehensive report.
The algorithm were also tested on the Yale
database B (Georghiades et al., 2001), which contains
the images of 10 subject, each seen under 576 view-
ing conditions (9 poses × 64 illuminations). For each
subject, the user data are randomly partitioned into
80% for training, and 20% for testing. The data of the
other 9 subjects are used as the impostor data. We also
test the six different illumination schemes, as shown
in Fig. 9 (b). In this experiment, it can be noticed that
horizontal Gaussian derivative filter does not further
improve the performance, which can be explained by
the fact that the Yale database B contains very ex-
treme illuminations, which cause deep shadows and
strong edges in horizontal directions. Our laboratory
data are more realistic.
A STUDY ON ILLUMINATION NORMALIZATION FOR 2D FACE VERIFICATION
47