information being available from other views.
A method of aligning multiple scans from vari-
ous viewpoints to ensure the 3D scene model com-
pleteness for complex and unstructured underground
environments is discussed by (Craciun et al., 2010).
A technique for extracting features from urban buil-
dings by fusing camera and lidar data is presented by
(Becker and Haala, 2007) but it also fails to address
this problem. (Frueh et al., 2004) proposed a method
in which the point cloud is used to generate a 3D mesh
that is then classified as foreground or background.
Large holes in the background layer, caused by oc-
clusions from foreground layer objects, are then filled
by planner or horizontal interpolation. However, such
an approach may result in false features in case of in-
sufficient repetitions or lack of symmetry (Li et al.,
2011). In our work, we aim to resolve this problem
by using a new approach combining both multiple
views and multiple passages such that first multiple
views are exploited to fill in some of the holes in the
3D point cloud followed by the use of multiple scans
of the same environment obtained at different days
and hours of the day to fill in the remaining holes and
completing the missing regions of the urban cartogra-
phy. This ensures that the resulting 3D point cloud
of the cartography is most accurate containing only
the exact and actual permanent features/regions. An
overview of the method is presented in Algorithm 1.
2 3D SCAN REGISTRATION
Different features and robust landmarks extracted
from 3D images as points of interest and as refer-
ences for image mapping and scan registration have
commonly been used for different multi-sessional
SLAM (Simultaneous Localisation And Mapping) al-
gorithms. This approach works well in simple repeti-
tive paths. But some more complex situations can be
found in urban environments where the selected fea-
tures/regions can be occluded. When the data acqui-
ring vehicle enters from different directions, then the
path is not repetitive. As a result, the selected fea-
tures/regions may not be readily visible, etc. Thus,
in order to cater for this problem, the method of di-
rect geo-referencing of 3D LiDAR points is found
most suitable in our case. The method uses integrated
GPS/IMU data to directly orient laser data from its
local reference frame to the mapping reference frame
(WGS84).The advantage of using this method is that
the transformation between the local laser reference
frame and the mapping reference frame is known at
any given moment (as long as the laser is synchro-
nized), independently if the laser is collecting data
(a)
(b)
(c)
Figure 1: (a) The Vis Center’s LiDAR Truck. (b) Optech
LiDAR/GPS system along with IMU mounted on a rigid
frame. (c) The different viewing angles of the mounted Li-
DAR systems.
in a static mode or in kinematic mode. Thus, the
laser can be used as a pushbroom sensor sweeping the
scene with profiles while fixing the scan angles as the
vehicle moves.
The data that we have used to evaluate our work
are the dynamic data set of the 3D Urban Data Chal-
lenge 2011, which contains dynamic scenes from
downtown Lexington, Kentucky, USA obtained from
the Vis Center’s (University of Kentucky) LiDAR
Truck containing two Optech LiDAR sensor heads
(high scan frequency up to 200 Hz), a GPS, an iner-
tial measurement unit and a spherical digital camera
as shown in Figure 1.
3 CLASSIFICATION OF 3D
URBAN ENVIRONMENT
We classify the urban environment into 2 main ca-
tegories: Permanent objects and Temporary objects.
In order to achieve this, the 3D point cloud is first
segmented into objects which are then classified into
basic object classes. Once classified into these basic
classes, they are then grouped under one of the 2 men-
tioned categories. Although several methods have
been proposed for the classification of urban environ-
ments, we have used one of the most recent meth-
ods (Aijazi et al., 2012) for this task. This method
presents a super-voxel based approach in which the
3D urban point cloud is first segmented into vox-
els and then converted into super-voxels. These are
then clustered together using an efficient Link-Chain
method to form objects. These objects are then clas-
sified using local descriptors and geometrical features
into 6 main classes: {Road, Building, Car, Pole, Tree,
Unclassified}. The salient features of this method are
data reduction, efficiency and simplicity of approach.
The 3D point cloud obtained from each of the two
mounted LiDAR sensors is divided into the 6 object
ICPRAM2013-InternationalConferenceonPatternRecognitionApplicationsandMethods
406