a ROI can be determined. While there is no quan-
tative definition of saliency, it is often considered in
terms of the principles of perceptual organization and
the laws of Gestalt psychology (Lowe, 1985). Ele-
ments are grouped together according to the phenom-
ena of Proximity, Similarity, Closure, Continuation,
Symmetry and Familiarity. Since many of the group-
ing phenomena found to be important in human visual
perception are related to discontinuities, an approach
to object segmentation based on edge detection is ap-
propriate.
Attempts at object segmentation based on grouping
of edge-segments (Elder and Zucker, 1996; Kiranyaz
et al., 2006; Wang et al., 2005; Stahl and Wang,
2007; Mahamud et al., 2003) roughly follow these
steps: First the input images are preprocessed before
running an edge detection algorithm over them, the
resulting edge maps are traced into edge-segments,
these edge-segments are further processed. Finally, a
search algorithm is applied to graph representations of
the edge-segments to find closed contours. Some pro-
cessing steps of our method are inspired by a related
method by Ferreira, Kiranyaz, and Gabbouj (hence-
forth referred to as FKG) (Kiranyaz et al., 2006),
which in turn,has some similarities to a method by
Elder and Zucker (Elder and Zucker, 1996).
3 THE ALGORITHM
3.1 Preprocessing
Reduction to the Region of Interest (ROI). The
method starts by reducing the image to the region
of interest (ROI). Only that part of the image is
processed further. Knowledge of that ROI can ei-
ther be generated by manually labelling or by using
additional sensors like laser-scanners or PMD 3D-
cameras. They provide information about the distance
to the object, allowing a discrimination of object and
background. Then again those sensor usually have
a much lower spatial resolution. So usually only a
bounding box of the object’s position can be deter-
mined. At best a rough contour of the object at a much
lower resolution can be found.
3.2 Prominence-Map Generation
Detection of Weak and Strong Edges. In a first
step edges of different strength are detected. The one
approach is to detect edges in different scales of a
scale space. However due to gaussian filtering edges
can move . FKG uses a cascade of bilateral filters to
Figure 1: Edge maps generated with 4 different thresholds
(top) and the corresponding Prominence map (bottom).
overcome this problem, but Runtime is high and pa-
rameters are heavily dependent on the object in the
image. We generate edge maps with different levels
of detail by increasing the edge detection thresholds
of the Canny Edge Detector (Canny, 1986) in k iter-
ations. This can be viewed as an approximate dis-
cretization of edge pixel strengths.
Combining Edges to a Prominence Map. Edge
maps are then merged into a single prominence map.
Each edge pixel i
x,y
is assigned a prominence value
p
x,y
∈
{
1,...,k
}
corresponding to the k increasing
Canny edge detection thresholds after which that edge
pixel still remains detected as an edge. Figure 1 shows
the prominence map generated from combining edge
maps with 4 different thresholds.
Morphological Thinning. Since Canny edge de-
tection does not guarantee edge-segments with a
thickness of exactly one pixel, a thinning process is
applied to the prominence map to reduce that thick-
ness to one pixel. This simplifies the subsequent edge
tracing process.
The Use of A Priori Knowledge. If additionally to
the bounding box also a rough contour of the object is
available, this is used as additional a priori knowledge
to boost the segmentation process in two ways: First
the prominence values of edge pixels in a region on
and near the contour of that mask can be increased
and are so more likely to be incorporated in the final
contour. Second the search space can be reduced by
discarding pixels which are not in that region.
3.3 Generating a Graph Representation
Tracing of Edge-segments. The process of tracing
is the first step in the transition from individual edge
pixels on the prominence map to a graph represen-
tation. To generate a list of edge-segments ES
j
∈
IMAGAPP 2009 - International Conference on Imaging Theory and Applications
44