DETECTION AND TRACKING OF
MULTIPLE MOVING OBJECTS IN VIDEO
Wei Huang and Jonathan Wu
Department of Electrical and Computer Engineering, University of Windsor
Windsor, Ontario, N9B 3P4, Canada
Keywords: Motion detection, tracking, partial occlusion, color, texture, DCT, inexact graph matching.
Abstract: This paper presents a method for detecting and tracking multiple moving objects in both outdoor and indoor
environments. The proposed method measures the change of a combined color-texture feature vector in each
image block to detect moving objects. The texture feature is extracted from DCT frequency domain. An
attributed relational graph (ARG) is used to represent each object, in which vertices are associated to an
object’s sub-regions and edges represent spatial relations among the sub-regions. Object tracking and
identification are accomplished by matching the input graph to the model graph. The notion of inexact graph
matching enables us to track partially occluded objects. The experimental results prove the efficiency of the
proposed method.
1 INTRODUCTION
The efficient detection and tracking of multiple
moving objects is currently one of the most active
research topics in computer vision. It has many
applications such as visual surveillance, human-
machine interfaces, video communication, and so
on.
As for motion detection, the background
subtraction technique is a popular method. In
(
Stauffer and Grimson, 2000), the pixel value was
modeled by a mixture of weighted K Gaussian
distributions to support multiple backgrounds.
(
Elgammal et al., 2002) used a nonparametric kernel
density model by estimating the probability of pixel
intensity directly from a set of recent intensity
values.
As to the tracking method, the most widely used
cues in object tracking are color, spatial position,
shape and motion. In (
Xu et al., 2004), five significant
features were used, including velocity, size, elliptic-
fit aspect ratio, orientation, and dominant color.
(
Brasnett et al., 2005) demonstrated that the combined
color and texture cues provided a good tracking
result that was more accurate than the two cues
individually.
In this paper we introduce a new motion
detection method which does not compute any
model of the background. We measure the change of
a combined color-texture feature vector in each
image block within a time window and then directly
obtain moving objects by statistically analyzing the
change. For effective tracking, the attributed
relational graph is used to represent each moving
object. A combined color-texture-position feature
vector is used to describe each object’s sub-regions,
which are associated to the vertices of the ARG.
Inexact graph matching enables us to track and
identify partially occluded objects. In the discussion
below, we calculate the color-texture combined
feature vector for motion detection in Section 2.1,
and then we explain the details of detecting moving
objects using eigenspace decomposition and
statistical analysis in Section 2.2. Section 2.3
describes how to construct the attributed relational
graph to represent the detected object. Section 2.4
gives the details of identifying objects using inexact
graph matching technique. We show experimental
results for real image sequences in Section 3.
Conclusions are given in Section 4.
2 PROPOSED ALGORITHM
2.1 The Color-Texture Feature
Approach for Motion Detection
In (Latecki et al., 2004), an idea was introduced that
the texture vectors are very likely to have a large
spread when a moving object is passing through a
492