Figure 1: Main features of the used RMB LiDARs, and po-
sitioning of the sensors in the experiments.
ture surveillance systems.
The rest of the paper is organized as follows: Sec-
tion 2. provides some information about related work
in the field of gait recognition, Section 3. presents
a brief introduction to our gait recognition method
using Rotating Multi-Beam LiDAR sensor. Section
4. gives quantitative results about the accuracy of
each sensor in the different gait sequences. In Section
5. experiments on activity recognition are presented.
Conclusion is provided in Section 6.
2 RELATED WORK
Gait recognition has been extensively studied in the
recent years (Zhang et al., 2011). The proposed meth-
ods can be divided into two categories: model based
methods, which fit models to the body parts and ex-
tracts features and parameters like joint angles and
body segment lengths, and model free methods, where
features are extracted from the body as a whole ob-
ject. Due to the characteristics and the density of point
clouds generated by a Rotating Multi-Beam LiDAR
sensor, like the Velodyne HDL-64E or the VLP-16,
robust generation of detailed silhouettes are hard to
accomplish, so we decided to follow a model free ap-
proach as the model based methods need precise in-
formation on the shape of body parts, such as head,
torso, thigh etc. as described in (Yam and Nixon,
2009), which are often missing in RMB LiDAR-based
environments.
There are many gait recognition approaches pub-
lished in the literature which are based on point clouds
(Tang et al., 2014; Gabel et al., 2012; Whytock et al.,
2014; Hofmann et al., 2012), yet they use the widely
adopted Kinect sensor which has limited range and
a small field-of-view and is less efficient for applica-
tions in real life outdoor scenarios than LiDAR sen-
sors. Also the Kinect provides magnitudes higher
density than an RMB LiDAR, so the effectiveness of
these approaches are questionable in our case.
The Gait Energy Image (Han and Bhanu, 2006),
originally proposed for optical video sequences, is of-
ten used in its basic (Shiraga et al., 2016) or improved
version (Hofmann et al., 2012), since it provides a
robust feature for gait recognition. In (G´alai and
Benedek, 2015) many state-of-the-art image based
descriptors were tested for RMB LiDAR point cloud
streams, proposed methods for both optical images
(Kale et al., 2003) and point clouds were evaluated.
(Tang et al., 2014) uses Kinect point clouds and cal-
culates 2.5D gait features: Gaussian curvature, mean
curvature and local point density which are combined
into 3-channel feature image, and uses Cosine Trans-
form and 2D PCA for dimension reduction, but this
feature needs dense point clouds for curvature calcu-
lation, thus not applicable for RMB LiDAR clouds.
(Hofmann et al., 2012) adopts the image aggregation
idea behind the Gait Energy Image and averages the
pre-calculated depth gradients of a depth image cre-
ated from the Kinect points. This method proved to
be more robust for sparser point clouds, yet it was
outperformed by the Lidar-based Gait Energy Image,
which is described in Section 3. in detail.
2.1 Gait Databases
The efficiency of the previously proposed methods
are usually tested on public gait databases like the
CMU Mobo (Gross and Shi, 2001), the CASIA
(Zheng et al., 2011) or the TUM-GAID (Hofmann
et al., 2014) database. However these datasets were
recorded with only a single person present at a time,
with limited background motion and illumination is-
sues, which constraints are often not fulfilled in re-
alistic outdoor scenarios. To overcome the domina-
tion of such databases (Benedek et al., 2016) pub-
lished the SZTAKI-LGA-DB dataset recorded with
RMB LiDAR sensor in outdoor environments. Dur-
ing the experiments presented in Section 4 we fol-
lowed the same approach by recording the point cloud
sequences.
2.2 Devices Used in Our Experiments
The LiDAR devices used here are the Velodyne HDL-
64E and VLP-16 sensors, shown in Fig. 1. The HDL-
64E sensor has a vertical field-of-view of 26.8° with
64 equally spaced angular subdivisions, and approx-
imately 120 metres range providing more than two
million points per second. The VLP-16 has 30° verti-
cal field-of-view, 2° vertical resolution and a range of