colored blob. In this step, we choose the biggest
region as an object-colored blob.
2.2 Adaptive Learning System
Training is an off-line procedure that does not affect
the on-line performance of the tracker. Nevertheless,
the compilation of a sufficiently representative
training set is a time-consuming and labor-intensive
process. To cope with this problem, an adaptive
training procedure has been developed. Training is
performed on a small set of seed images for which a
human provides ground truth by defining object-
colored regions. Following this, detection together
with hysteresis thresholding is used to continuously
update the prior probabilities P(o), P(c) and P(c|o)
based on a larger image data set. The updated prior
probabilities are used to classify pixels of these
images into object-colored and non-object-colored
ones. The final training of the classifier is then
performed based on the training set resulting from
user editing. This process for adapting the prior
probabilities P(o), P(c) and P(c|o) can either be
disabled as soon as the achieved training is deemed
sufficient for the purposes of the tracker, or continue
as more input images are fed to the system.
The success of the color detection depends
crucially on whether or not the luminance conditions
during the on-line operation of the detector are
similar to those during the acquisition of the training
data set. Despite the fact that using the UV color
representation model has certain luminance
independent characteristics, the object color detector
may produce poor results if the luminance
conditions during on-line operation are considerably
different to those used in the training set. Thus, a
means of adapting the representation of object-
colored image pixels according to the recent history
of detected colored pixels is required. To solve this
problem, object color detection maintains two sets of
prior probabilities (Zabulis et al, 2009). The first set
consists of P(o), P(c), P(c|o) that have been
computed off-line from the training set. The second
is made up of
)(oP
W
,
)(cP
W
,
)|( ocP
W
,
corresponding to the P(o), P(c), P(c|o) that the
system gathers during the W most recent frames
respectively. Obviously, the second set better
reflects the “recent” appearance of object-colored
objects and is therefore better adapted to the current
luminance conditions. Object color detection is then
performed based on the following moving average
formula:
(|) (|) ((1 ) (|))
AW
oc Poc P oc
γ
=+−
,
(2)
where
)|( coP
A
represents the adapted probability
of a color c being an object color. P(o|c) and
)|( coP
W
are both given by Equation (1) but
involve prior probabilities that have been computed
from the whole training set [for P(o|c)] and from the
detection results in the last W frames [for
)|( coP
W
].
is a sensitivity parameter that controls the
influence of the training set in the detection process
)10(
. If
1
, then the object color detection
takes into account only the training set (35 images in
the off-line training set), and no adaptation takes
place; if
is close to zero, then the object color
detection becomes very reactive, relying strongly on
the recent past for deriving a model of the immediate
future. W is the number of history frames. If W value
is too high, the length of history frames will be too
long; if W value is set too low, the history for
adaptation will be too short. In our implementation,
we set
= 0.8 and W = 5 which gave good results
in the tests that have been carried out.
Thus, the object color probability can be determined
adaptively. By using on-line adaptation of object
color probabilities, the classifier is easily able to
cope with considerable luminance changes, and also
it is able to segment the object even in the case of a
dynamic background.
3 RESULTS
In this section, representative results from our
experiment are shown. Figure 2 provides a few
representative snapshots of the experiment. The
reported experiment is based on a sequence that has
been acquired with USB camera at a resolution of
320x240 pixels. This process is done in real time
and on-line. The experimental room is outdoor area
(balcony). Note that the training set (35 images in
the off-line training set) was collected from the
indoor area. So the luminance change makes it much
more challenging.
The left window depicts the input images. The
middle window shows the output images. The right
window represents the object probability map in the
U and V axis in color model, as depicted in Figure 3.
In the initial stage (frame 15), when the
experiment starts, the object color probability does
not converge to a proper value. In other words, the
color probability is scattering. So the segmented
output cannot be achieved well because it uses only
from the off-line data set which the lighting is
SIMULTECH 2011 - 1st International Conference on Simulation and Modeling Methodologies, Technologies and
Applications
202