Distante, 2006) ,(Spagnolo, D'Orazio, Leo, Distante,
2005).
The following step is to aggregate pixels belonging
to the same moving object in order to build a higher
logical level entity named region or blob. Detected
regions are the input of a color based probabilistic
tracking procedure. The main aim of the tracking
procedure is to analyze temporally the displacements
of each moving region in order to manage
overlapping or occlusion when the following
decision making procedures could otherwise be
misleading .
The tracking procedure exploits appearance and
probabilistic models, suitably modified in order to
take into account the shape variations and the
possible region of occlusion (Cucchiara, Grana &
Tardini, 2004). Using the procedures outlined each
object is localized in the 2D image plane and is
temporally tracked. Tracking information is the
input to the two procedure dealing with the
automatic recognition of suspicious human
behaviors.
The first procedure deals with the problem of
detection of forbidden area violation.
This procedure consists of two steps: firstly the 3D
localization of moving regions is obtained using an
homographic transformation (Hartley,R., Zisserman,
A., 2003); then object positions on the ground plane
are compared with those labeled as forbidden in the
foregoing calibration procedure. If a match occurs
the algorithm generates an alarm.
The second procedure deals with the problem of
recognition of abandoned and removed objects.
In the literature usually these two issues are not
distinguished, and they are dealt with in a similar
way. So, detecting an abandoned/removed object
becomes a tracking problem, with the aim of
distinguishing moving people from static objects
left/removed by human people (see (Connell, 2004)
and (Spengler & Schiele, 2003) for good reviews).
In this work, instead, the goal is to distinguish
between these two cases: so a classic tracking
problem now becomes a pattern recognition
problem. The reliability of the algorithm is strictly
related to the ability to find/not find correspondences
between patterns extracted in different images.
The approach implemented starts from the
segmented image at each frame. If a blob is
considered as static for a certain period of time (we
have chosen to consider a blob as static if its
position does not change for 5 seconds, but this
value is arbitrary and does not affect the algorithm),
it is passed to the module for removed/abandoned
discrimination. By analyzing the edges, the system is
able to detect the type of static regions as abandoned
object (a static object left by a person) and removed
object (a scene object that is moved). Primarily, an
edge operator is applied to the segmented binary
image
t
F
to find the edges of the detected blob. The
same operator is applied to the current gray level
image
t
I
.
To detect abandoned or removed objects a matching
procedure of the edge points in the resulting edged
images is introduced. To perform edge detection, we
have used the Susan algorithm (Smith, 1992), which
is very fast and has optimal performances. The
matching procedure physically counts the number of
edge points in the segmented image that have a
correspondent edge point in the corresponding gray
level image. Additionally, a searching procedure
around those points is introduced to avoid mistakes
due to noise or small segmentation flaws. Finally if
the matching measurement
FI
t
M is greater than a
certain value th
a
experimentally selected, it means
that the edges of the object extracted from the
segmented image have correspondent edge points in
the current grey level image and it is labeled as an
abandoned object by the automatic system.
Otherwise, if
FI
t
M
has a small value, typically less
then a given threshold th
r
, it means that the edges of
the foreground region do not match with edge points
in the current image, so it is labelled by the
automatic system as an object of the background that
has been removed. For values of
FI
t
M
between
these two thresholds the system is not able to decide
on the nature of the object.
3 EXPERIMENTAL RESULTS
The experiments were performed in both the
Messapic Civic Museum of Eganthia.
The museum has many rooms containing important
evidence of the past: the smallest archeological finds
are kept under lock in proper showcases but the
largest ones are exposed without protection. The
areas around the unprotected finds are no-go zones
for visitors and are defined with cord. Only a visual
control can ensure that visitors don’t step over the
cord in order to touch the finds or to see them in
more detail.
The proposed framework was tested to detect
forbidden entry into protected areas of the museum
and to recognize removed and abandoned objects in
the monitored areas.
VISAPP 2007 - International Conference on Computer Vision Theory and Applications
528