resentation built by an existing SLAM method, but
without increasing its complexity.
Three different sensors have been analyzed and
characterized to achieve the proposed task: a Kinect
RGB-D sensor, a ToF (SR3000) optical camera and
a Tilting LRF (Hokuyo). Appearance-based informa-
tion obtained typically using texture mapping, are not
considered in this paper. In the following section,
we will describe some of the most interesting related
works, followed by a section dedicated to analyze the
main characteristics of the mentioned sensors. In sec-
tion 4, are presented, some evaluations in order to be
able to chose the correct sensor for the task, and in
section 5 we present experimental results, using the
PR2 robot.
2 RELATED WORKS
In recent years, 3D modeling and mapping has be-
come one of the most interesting subjects of research
all along the world. 3D sensors allow to extract the
richness of geometric features, presents in most of the
environments. The construction of 3D models could
be done in many different ways, depending on the
type of environment, the sensor used and applications.
In (Trevor et al., 2012) the problem of 3D model-
ing is considered as a part of SLAM techniques. In
this work, it is used a 3D sensor (a tilting LRF or a
Kinect like sensor) to extract 3D planar surfaces, that
combined with 2D segments obtained from a 2D scan-
ner at base of a mobile robot, are used to build a map
using the GTSAM library (Dellaert and Kaess, 2006).
In this way, 2D lines and 3D planes with a high level
representation and easy to be annotated with seman-
tic information are a good combination to create an
accurate map with high level features.
In (N
¨
uchter and Hertzberg, 2008) as a part of a
6D SLAM method, point clouds are acquired using a
rotating LRF and registered using ICP. Planes are ex-
tracted from the global 3D point cloud by a RANSAC
method and then with the use of a constraint network
these planes are semantically annotated, i.e. walls,
floor or ceiling.
The problem of environment modeling can be re-
solved considering that robots are already localized,
as in (An et al., 2012). In this work, authors concen-
trate more on the computational part, by proposing a
method for fast planar faces detection using 2D lines
extracted from a tilting LRF over a mobile robot. The
proposed method works in real-time and only stores
the initial and end point of each 2D line to construct
the 3D model.
In (Klaess et al., 2012), it is built a 3D map using
a set of 3D laser scanners acquired at different po-
sitions (stop-and-go method); poses are provided by
the use of the gmapping method (Grisetti et al., 2007).
Then, the global point cloud is refined off-line, by ICP
methods. Finally using surfels (surface elements) the
global dense map is reduced, to be treatable by the
robot.
In (Rusu et al., 2009) a pan rotating LRF has been
used to acquire a point cloud, that it is used to get
a high level semantic model of a kitchen environ-
ment. The model is built off-line and it is used a ma-
chine learning algorithm to classify objects and label-
ing them with semantic information.
In (Wolf and Sukhatme, 2008) a tilting LRF has
been used to get 3D data, machine learning methods
have been applied to classify environment to naviga-
ble and non navigable zones. In (Douillard et al.,
2010) a LRF has used to build a hybrid 3D outdoor
environment model using elevation level and planar
faces. As there is also other works use 3D sensors
to model objects and for surface reconstruction as in
(Newcombe et al., 2011) and (Lai et al., 2011), the
modeled objects are used to build semantic maps or
for pattern and objects recognition, its applications
are generally for image of color and depth recogni-
tion or to help robot to recognize and grasp daily used
objects.
So, as we have seen, there is many works from
here and there that use 3D data for multiple applica-
tions from object to environment modeling, we can
extend to say cities modeling as generalization for
outdoor modeling like (Wolf and Sukhatme, 2008)
and (Douillard et al., 2010). Any way, as we has
already mentioned our goal is the modeling of large
scale indoor or man made environments, where most
of works have used a LRF to acquire 3D data.
Other work have concentrated on the evaluation
of sensors as in (Sturm et al., 2012) or (Henry et al.,
2012), they have studied the case of like Kinect sen-
sor and its use in SLAM. In (Smisek et al., 2011) an
evaluation for Kinect ,ToF camera and stereo vision
has been done.
The results and the conclusions of the evaluation is
different from work to an other, because of the differ-
ence between application and the performance needed
in each application.
In this work, has been considerate the three sen-
sors: a Kinect like sensor, a ToF camera and a LRF
over a tilting platform. We evaluate and present their
performances for 3D modeling in large scale indoor
environments.
ComparisonofActiveSensorsfor3DModelingofIndoorEnvironments
443