rameters. All classifiers have 100 features and the dif-
ference between these classifiers is the training image
resolution. Two classifiers are trained by resizing the
images to a resolution of 16×12 and 32×24, respec-
tively. For these classifiers the sampling matrix is
T
rect
=
1 0
0 1
.
This sampling matrix is common practice and means
that the five Haar-like features (Fig. 1) are calculated
at all image coordinates.
The third classifier uses the quincunx sampling
method. To perform this sampling, a minimal reso-
lution of 40 × 30 is required. After resizing the im-
age just these positions are used for feature calcula-
tion which are determined by the quincunx sampling
matrix
T
quin
=
2 0
−1 2
.
It is to mention that the minimal size of the Haar-like
features is set to be 2 × 2. Table 1 shows the number
of sampling points and features that are extracted dur-
ing the training process. The performance results of
Table 1: Trained classifiers.
Resolution Sampling Points Features
16× 12, T
rect
192 15· 10
3
32× 24, T
rect
768 260· 10
3
40× 30, T
quin
300 160· 10
3
the different classifiers are illustrated by using ROC
curves as shown in Fig. 7. The results reveal that the
best classification performance is obtained by using
the resolution 40 × 30 with the sampling method and
unsatisfying performance by the resolution 16 × 12.
Even though the best classifier’s feature pool is signif-
icantly smaller than the number of features used for
the classifier with resolution 32 × 24 the results are
slightly better. This strengthens the assumption that
the proposed sampling method is valid and moreover
can even improve classification performance without
increasing the computational load during the training
process.
Summing up, an approach has been introduced
to generate a sampling grid to determine reasonable
positions for calculating the Haar-like features. On
the one hand the number of features is reduced by
around 40% and the classification accuracy is in-
creased. These advantages are due to the better uti-
lization of positions for feature calculation which are
adapted to the properties of the training images. One
aspect that should be included in future work is to
transfer this methodology directly to the Haar-like
10
−3
10
−2
10
−1
0.8
0.85
0.9
0.95
1
false positive rate
true positive rate
16x12
32x24
40x30, T
quin
Figure 7: ROC curve of three equally trained classifiers us-
ing two different cartesian and the quincunx sampling.
features to further reduce the computational complex-
ity without losings in accuracy.
REFERENCES
Lienhart, R., Kuranov, A., and Pisarevsky, V. (2002). Em-
pirical analysis of detection cascades of boosted clas-
sifiers for rapid object detection. Technical report,
Mic. Research Lab, Intel Corporation, Santa Clara,
CA 95052, USA.
Ohm, J.-R. (2004). Multimedia Communication Technol-
ogy. Springer, Berlin, Heidelberg, Germany.
Overett, G. and Petersson, L. (2007). Boosting with multi-
ple classifier families. Proc. of IEEE Intelligent Vehi-
cles Symposium, pages 1039–1044.
Ponsa, D., Lopez, A., Lumbreras, F., Serrat, J., and Graf,
T. (2005). 3d vehicle sensor based on monocular vi-
sion. In Proc. of the 8th Int. IEEE Conf. on Intelligent
Transportation Systems, Vienna, Austria.
Russell, B., Torralba, A., and Freeman, W. T. (2005). La-
belme image database. http://labelme.csail.mit.edu.
Sun, Z., Bebis, G., and Miller, R. (2004). On-road vehicle
detection using optical sensors: A review.
Viola, P. and Jones, M. (2001). Rapid object detection us-
ing a boosted cascade of simple features. In Accepted
Conf. on Computer Vision and Pattern Recognition.
AN ANALYSIS OF SAMPLING FOR FILTER-BASED FEATURE EXTRACTION AND ADABOOST LEARNING
185