3D Descriptor for an Oriented-human Classification from Complete
Point Cloud
Kyis Essmaeel, Cyrille Migniot and Albert Dipanda
LE2I-CNRS, University of Burgundy, Dijon, France
Keywords:
Human Classification, Histogramm of Oriented Normals, 3D Point Cloud.
Abstract:
In this paper we present a new 3D descriptor for the human classication. It is applied over a complete point
cloud (i.e 360
◦
view) acquired with a multi-kinect system. The proposed descriptor is derived from the His-
togram of Oriented Gradient (HOG) descriptor : surface normal vectors are employed instead of gradients,
3D poins are expressed on a cylindrical space and 3D orientation quantization are computed by projecting
the normal vectors on a regular polyhedron. Our descriptor is utilized through a Support Vector Machine
(SVM) classifier. The SVM classifier is trained using an original database composed of data acquired by
our multi-kinect system. The evaluation of the proposed 3D descriptor over a set of candidates shows very
promising results. The descriptor can efficiently discriminate human from non-human candidates and pro-
vides the frontal direction of the human with a high precision. The comparison with a well known descriptor
demonstrates significant improvements of results.
1 INTRODUCTION
Human detection has been an important research sub-
ject in computer vision for many years. It is used in
a wide variety of applications including health mon-
itoring, driving assistance, video games and behav-
ior analysis. It is particularly a challenging prob-
lem for many reasons. Pose, color and texture sig-
nificantly vary from one person to another, besides
the complexity of the working environment represents
another challenge to overcome. While most of the ap-
proaches for human detection rely on color-image, the
recent advances in depth sensor technology provided
additional solutions. The introduction of affordable
and reliable depth sensors like the kinect from Mi-
crosoft has dramatically increased the interest of these
technologies and is leading to a huge number of ap-
plications using such sensors. Human detection was
one of the first domains to use this new technology
and exploit its benefits. Depth information is most of
the time used to reduce the computation cost. How-
ever the descriptiveness of the 3D shape of the human
envelop was never really exploited.
There are two main categories of methods for human
detection: descriptor/classifier (Figure 1) and match-
ing templates. In the first category, HOG (Histogram
of Oriented Gradients ) (Dalal and Triggs, 2005) is
considered as one of the most successful descriptor
for 2D image human detection. It is used most of the
time with SVM as a classifier. The HOD (Histogram
of Oriented Depths) (Spinello and Arras, 2011; Choi
et al., 2013) is a well-known adaptation of the HOG
which is applied on depth images. HOD locally en-
codes the direction of depth changes and relies on a
depth-informed scale-space search. In fact it uses the
depth array as a 2D image to apply the HOG pro-
cess. Hence 3D data are not exploited in their first
forms, which makes them difficult to apply in scenar-
ios where multiple sources of information are com-
bined to produce the 3D data like in a multi-sensor
system. The Relational Depth Similarity Features
(RDSF) (Ikemura and Fujiyoshi, 2011) arise the same
problem as before. The RDSF calculate the degrees
of similarity between all of the combinations of rect-
angular regions inside a detection window in a sin-
gle depth image only. The second category of meth-
ods rely on matching one or many templates of cer-
tain body-parts in 2D data (images) or 3D data (point
clouds). The Ω-shape of the head and shoulders of a
human body are an example of descriptive templates
(Tian et al., 2013). To compare it to the data, Xia (Xia
et al., 2011) uses chamfer distance and Choi (Choi
et al., 2011) uses the Hamming distance.
In this paper we propose a human classification
method that operates on point clouds and exploits
uniquely the 3D features of the human without using
Essmaeel, K., Migniot, C. and Dipanda, A.
3D Descriptor for an Oriented-human Classification from Complete Point Cloud.
DOI: 10.5220/0005679803530360
In Proceedings of the 11th Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2016) - Volume 4: VISAPP, pages 353-360
ISBN: 978-989-758-175-5
Copyright
c
2016 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved
353