and manipulation actions: complete meshes for
known objects (de Figueiredo et al., 2015), recon-
structed meshes for unknown objects (Aleotti et al.,
2012), and object part clusters modeled with su-
perquadrics (Faria et al., 2012), also for unknown ob-
jects. Despite these representations allowing for good
precision in representing the shape of the objects, they
suffer from high-dimensionality and varying descrip-
tion length, thus being hard to define a representation
of object categories suitable for generalization. Fur-
thermore, the noise present in the sensor data may
negatively influence representations with large num-
ber of parameters. Given the current perception tech-
nology, the most robust and simple features to repre-
sent objects, match their similarity with others, and
provide a basis for the definition of categories, must
be low-dimensional and rely on gross features, to pre-
vent over-fitting.
2.2 Shape Completion
In recent years shape completion using a single view
has been extensively studied, typically in robotics
grasping applications. Usually multiple object par-
tial views are acquired from different viewpoints, us-
ing 3D range cameras, and the gathered point clouds
are then registered and aligned together in a com-
mon reference frame. The Iterative Closest Point al-
gorithm (Besl and McKay, 1992) and efficient vari-
ants (Rusinkiewicz and Levoy, 2001) are often used to
compute the alignment transformations and to build
a complete object shape model (Chen and Medioni,
1992). However, when only a single view is avail-
able and/or it is not possible to acquire several views
due to time constraints or scenario/robot restrictions
the shape completion problem becomes harder and
some assumptions or pattern analysis must be made.
In this direction, a wide range of ideas have been
proposed including fitting the visible object surface
with primitive shapes such as cylinders, cones, paral-
lelepipeds (Marton et al., 2009; Kuehnle et al., 2008)
or with more complex parametric representations like
superquadrics (Biegelbauer and Vincze, 2007).
Closely related to our shape completion approach,
Thrun and Wegbreit (Thrun and Wegbreit, 2005) pro-
posed a method based on the symmetry assumption.
This method considers 5 basic and 3 composite types
of symmetries that are organized in an efficient entail-
ment hierarchy. It uses a probabilistic model to evalu-
ate and decide which are the completed shapes, gener-
ated by a set of hypothesized symmetries, that best fit
the object partial view. More recently Kroemer et al.
(Kroemer et al., 2012) proposed an extrusion-based
completion approach that is able to deal with shapes
that symmetry-based methods cannot handle. The
method starts by detecting potential planes of sym-
metry by combining the Thrun and Wegbreit method
with Mitra et al.’s fast voting scheme (Mitra et al.,
2006). Given a symmetry plane, an ICP algorithm
is used to decide the extrusion transformation to be
applied to the object partial point cloud. Despite the
fact that these methods were shown to be robust to
noise and were able to deal with a wide range of ob-
ject classes, they are inherently complex in terms of
computational effort and thus, not suitable in real-
time. Nevertheless, to simplify this problem, one can
take advantage of common scenario structures and
objects properties that are usually found in daily en-
vironments. They mostly involve man-made objects
that are typically symmetric and standing on top of
planar surfaces. For example, Bohg et al. (Bohg et al.,
2011) took advantage of the table-top assumption and
the fact that many objects have a plane of reflection
symmetry. Starting from the work of Thrun and Weg-
breit (Thrun and Wegbreit, 2005) and similar in spirit
to Bohg et al. (Bohg et al., 2011), we propose a new
computationally efficient shape completion approach
which translates a set of environmental assumptions
into a set of approximations, allowing us to recon-
struct the object point cloud in real-time, given a par-
tial view of the object.
3 METHODOLOGIES
The role of our algorithm is to obtain a semantic de-
scription of the perceived objects in terms of their
pose, symbolic parts and probability distributions
over possible object categories. The object segmenta-
tion step (Muja and Ciocarlie, ) is followed by part de-
tection and object category estimation, which rely on
a full object point cloud. When only a partial view of
the object is available, we employ a symmetry-based
methodology for object shape completion. Next, the
extraction of semantical parts is based on the object’s
dimensions along the main geometrical axes and can
be achieved by bounding-box analysis via PCA. The
low dimensional and efficient representation obtained
guides the division of each object into a set of seman-
tical parts, namely, top, middle, bottom, handle and
usable area. This reduces the search space for robot
grasp generation, prediction and planning. The next
subsections explain our symmetry-based method for
shape-completion and the division of the completed
point cloud into a set of semantical parts.
VISAPP 2017 - International Conference on Computer Vision Theory and Applications
566