portance weight of a particle is calculated as the prod-
uct of the weights associated to each percept class.
From a theoretical point of view, the smoothing per-
formed in Equation 5 is a violation of the Markov as-
sumption, in fact past and current measurements for
objects in the same class are correlated by this up-
date method. This can be seen as an incremental es-
timate of the importance weight of a given particle
with respect to a series of measurements of a certain
object, rather than considering each measurement in-
dependently. For this reason, this method can easily
filter noise in the measurements, as well as limit the
effects of outliers, since it takes a series of consis-
tent measurements to affect significantly the distribu-
tion. However, this approach as presented in (R
¨
ofer
and J
¨
ungel, 2003) has some problems:
1. having ∆
up
6= ∆
dn
introduces an unnecessary
bias in the importance weights;
2. since measurements relative to different objects
are put in the same class, and the particles keep
their “memory” indefinitely, this mechanism in-
correctly correlates measurements which should
be independent.
While fixing the first problem is trivial, the same is not
true about the second; the correct solution would be
to assign a different percept class to each perceived
object, so that they can be treated as statistically in-
dependent. This however is not always possible, be-
cause in some cases the robot is observing different
objects which are undistinguishable, like walls in vi-
sion based localization, or range measurements with
laser scanners. Thus, here we introduce an “aging
law” which fades out the memory of past measure-
ments in the importance weights, thus reducing the in-
correct correlations among percepts of the same class.
This corresponds to an assumption of temporal co-
herence of a series of measurements, i. e. we expect
“bursts” of measurements relative to a certain object
to appear for a short time. So, if we define a set Γ of
percept classes, let w
i,j
t
be the importance weight for
the percept class j ∈ Γ relative to the i-th particle at
time t, α
j
∈ [0..1] an aging factor for class j, then the
result of the aging process ˆw
i,j
t
can be calculated as:
ˆw
i,j
t
= w
i,j
t
+ (1 − w
i,j
t
) · α
j
(6)
This function asymptotically fades an importance fac-
tor toward 1 (i. e. that percept class does not affect the
global importance factor of that particle), with a speed
controlled by the parameter α
j
, which can be set ac-
cording to the frequency of independent observations
of a given percept class. For α tending to 0, the par-
ticles keep “memory” of old percepts for a long time,
while for α tending to 1 the particle weights are re-
set in each iteration like in the SIR filter. This is also
beneficial in case a series of past measurements for a
given class resulted in a very low weight, which with-
out aging would affect negatively the global weight
of the particle, regardless of the values of the other
percept classes. Finally, after the aging step (Equa-
tion 6), given a ¯w
i,j
t
calculated from the measurement
model, we can apply the temporal smoothing step:
w
i,j
t
=
( ˆw
i,j
t−1
+ ∆
j
) if( ¯w
i,j
t
> ˆw
i,j
t−1
+ ∆
j
)
( ˆw
i,j
t−1
− ∆
j
) if( ¯w
i,j
t
< ˆw
i,j
t−1
− ∆
j
)
¯w
i,j
t
else
w
i
t
=
Y
j∈Γ
w
i,j
t
(7)
3.2 “lazy” Resampling
The main problem of the approach described in
(R
¨
ofer and J
¨
ungel, 2003) is the resampling step: par-
ticles are copied in the new distribution χ with a fre-
quency proportional to their importance weight w
i
t
,
as in the Step 3 of Algorithm 1, but the weights w
i,j
t
are not reset to 1 afterward, this because otherwise it
would not be possible to use them to filter the mea-
surements at the following iterations of the algorithm.
So, supposing that a particle with importance w
i
t
is
copied n
i
> 1 times in the target distribution χ, the
probability density in that location will increase by a
factor of n
i
. Consequently, after the resampling step
the particle distribution does not approximate the true
posterior anymore: the probability density will artifi-
cially increase in areas of high likelihood, while it will
further decrease in areas of low likelihood. The re-
sult is that this filter converges very quickly and tends
to focus most of the particles around the main peak
of the probability distribution; in practice the filter
works very well as long as the noise can be filtered
by its sensor smoothing mechanism, but when there
is a high level of ambiguity in the sensor data, for ex-
ample if no unique landmark is observed in a long
time, then the particles will start to quickly oscillate
among different peaks of the true posterior distribu-
tion, rather than assume a multi-modal distribution.
This problem can be fixed however, if we know the
number of times n
i
that a certain particle i has been
copied. This can be used to normalize the weights in
the target distribution, so that it again approximates
the posterior:
˜w
i
t
=
w
i
t
n
i
t
⇒
Q
j∈Γ
w
i,j
t
Q
j∈Γ
ν
i
t
⇒
Y
j∈Γ
w
i,j
t
ν
i
t
ν
i
t
=
c
q
n
i
t
; c = kΓk
(8)
ICINCO 2008 - International Conference on Informatics in Control, Automation and Robotics
96