Range Finder (LRF) is utilized in this work. Commu-
nication between the mobile robot and the computer
hosting the cameras is accomplished through a wi-fi
connection.
Leg Detection
Foreground Segmentation
(Detection)
HOG based Person
Detection
Multi-Person Tracking
2D SICK Laser
(on robot)
Flea2 RGB
cameras (wall mounted)
Figure 2: Multi-person detection and tracking system block
diagram.
Figure 2 shows block diagram of the envisaged
multi-person perceptual system. It has two main
parts. The first part deals with automated multi-
person detection. The second part is dedicated for
multi-person tracking. It takes all detections as in-
put and fuses them in a Particle Filtering framework.
Each of these parts are discussed in detail in subse-
quent sections. It is worth mentioning here that the
entire system is calibrated with respect to a global ref-
erence frame. Both the intrinsic and extrinsic param-
eters of the fixed cameras are known and in addition
the mobile robot has localization module that local-
izes its pose with respect to the reference frame using
laser scan segments.
3 MULTI-PERSON DETECTION
The perceptual functionalities of the entire system are
based on various detections. The detection modules
are responsible for automatically detecting persons in
the area. Differentperson detection modalities are uti-
lized depending on the data provided by each sensor.
Leg Detection with LRF: the LRF provides hori-
zontal depth scans with a 180
o
FOV and 0.5
o
reso-
lution at a height of 38cm above the ground. Person
detection, hence, follows by segmenting leg patterns
within the scan. In our implementation a set of ge-
ometric properties characteristic to human legs and
outlined in (Xavier et al., 2005) are used.
Person Detection from Wall Mounted Cameras:
to detect persons using the wall mounted cameras,
two different modes are used. First, a foreground
segmentation using a simple Σ-∆ background subtrac-
tion technique (Manzanera, 2007) is used. The mo-
bile robot is masked out of the foreground images us-
ing its position from its localization module. Second,
Histogram of Oriented Gradients (HOG) based per-
son detection (Dalal and Triggs, 2005) is used. This
method makes no assumption of any sort about the
scene or the state of the camera (mobile or static). It
detects persons in each frame using HOG features.
Both detections are projected to yield ground posi-
tions, (x, y)
G
with associated color appearance infor-
mation in the form of HSV histograms (P´erez et al.,
2002), of individuals in the area.
4 MULTI-PERSON TRACKING
Multi-person tracking in our context, is concerned
with the problem of tracking a variable number of
persons, possibly interacting, in the ground plane.
The literature in multi-target tracking contains differ-
ent approaches but when it comes to tracking mul-
tiple interacting targets of varying number (Khan
et al., 2005) has clearly shown that Reversible
Jump Markov Chain Monte Carlo - Particle Filters
(RJMCMC-PFs) are more appealing taking perfor-
mance and computational requirements into consid-
eration. Inspired by this, we have used RJMCMC-
PF, adapted to our cooperativeperceptual strategy, for
multi-person tracking driven by the various heteroge-
neous detectors. The actual detectors are: the LRF
based person detector, the foreground segmentation
(detection) and HOG based detections from each wall
mounted camera. Implementations choices crucial to
any RJMCMC-PF are briefly discussed below.
State Space: the state vector of a person i in hy-
pothesis n at time t is a vector encapsulating the id
and (x, y) position of an individual on the ground
plane with respect to a defined coordinate base, x
n
t,i
=
{Id
i
, x
n
t,i
, y
n
t,i
}.
Proposal Moves: RJMCMC-PF accounts for the
variability of the tracked targets by defining a variable
dimension state space. Proposal movespropose a spe-
cific move on each iteration to guide this variable state
space exploration. In our implementation, four sets of
proposal moves, m = {
Add
,
Update
,
Remove
,
Swap
},
are used. The choice of the proposals privileged in
each iteration is determined by q
m
, the jump move
distribution. These values are determined empir-
ically and are set to {0.15, 0.8, 0.02, 0.03} respec-
tively. Equation 1 shows computation of the accep-
tance ratio, β, of a proposal X
∗
at the n
th
iteration.
It makes use of the jump move distribution, q
m
; pro-
posal move distribution, Q
m
(), associated with each
move; the observation likelihood, π(X
n
t
); and the in-
teraction model, Ψ(X
n
t
).
VISAPP2013-InternationalConferenceonComputerVisionTheoryandApplications
412