replacing the external local force with a force field
derived from the image edge map.
Yan et al present the first fully automated system
for ear biometrics using 3D shape. There are two
major parts of the system: automatic ear region
segmentation and 3D ear shape matching. Starting
with the multimodal 3D + 2D image acquired in a
profile view, the system automatically finds the ear
pit by using skin detection, curvature estimation, and
surface segmentation and classification. After the ear
pit is detected, an active contour algorithm using
both colour and depth information is applied to
outline the visible ear region.
Except the Yan's method that automatically
approached ear-segmentation using 3-D information
of the profile images, the works based on edge
detection algorithms present automated approaches
for ear-segmentation using only 2-D images of the
ear. In this case, the edge image of the ear and
surroundings is first obtained then distinctive
contours are recognized. It is experimentally
concluded that the contour which includes the
maximum number of pixels, points out to the edge
of outer ear. Therefore the ear boundary can be
found easily according to that contour. However,
this technique suffers two main problems. First
problem is caused by the dependence of edge
detection algorithms to thresholding value. Second
one arises from the discontinuity of contours
obtained from the edge detector. In some cases,
several pixels of a same contour, which is detectable
using human’s eyes, are disconnected therefore; the
separated contours are recognized wrongly. This
effect is illustrated in figure 1.
(a) (b) (c)
Figure 1: (a). A profile image from USTB database, (b).
The edge image using “canny” operator with threshold 0.2
(c). The edge image with threshold 0.1 (the red line and
ellipse show the discontinuity of contours).
These problems motivated us to use a method
leading to obtain more reliable regions (instead of
edge contours) free from threshold selection.
Providing this purpose, we used topographic map of
ear image and label it. Each pixel in ear image is
described by one of the twelve topographic labels.
No predefined threshold is needed to label the image
in this manner. On the other hand, more geometric
properties of the ear image are yielded. Three labels,
namely, ridge, convex hill and convex saddle hill
which are selected experimentally, are combined in
order to extract the ear outline. More consideration
in outer ear shape shows that this part of ear is fully
described by the convex hill and convex saddle hill
labels. Moreover the ridge label in the ear image is
similar to the image edges, but the extraction of the
ridge label is independent of threshold value. Figure
2 shows the each of those labels extracted from a
test ear. It is experimentally concluded that the
connected component obtained from the
combination of topographic labels consisting of the
maximum number of pixels matches to the image of
outer ear. As a result the boundary of this region
demonstrates the ear in an ear image.
(A) (B) (C) (D)
Figure 2: (A). Selected image from USTB database, (B).
Green part shows the convex hill label, (C). Green part
indicates the convex sadlle hill, (D). Green part shows
convex hill and blue part shows the convex saddle hill
label, all extracted from selected ear image.
2 TOPOGRAPHIC LABELS
Using topographic models for representing images
has been reported in computer vision literature
((M.Haralick, 1983), (J.Wang, 2007)). This method
is classified in appearance based category. The main
advantage of using topographic features for
representation respect to intensity is its robustness to
lighting condition (L.Wang, 1993).
Consider a grey scale ear image as a surface in 3D
space where x and y axes are along image
dimensions. The value of surface at pixel (x,y) is the
pixel intensity f(x,y). Depending on the topographic
property of surface at each pixel one of twelve
topographic labels in Figure 3 is assigned to the each
pixel. In order to label intensity image based on
topographic property let us consider the input image
as a continues function f(x,y). The topographic label
at each pixel of image is determined using first and
second order derivatives on surface f.
EAR SEGMETATION USING TOPOGRAPHIC LABELS
187