pears in the IR camera, so a vehicle model for tracking
the vehicle is generated on the spot. In our proposed
system, tracking uses the cross-correlation peak be-
tween the vehicle model image and the image being
tracked, so the vehicle model is the area surrounding
the vehicle in the video. It should be noted here that
the appearance of the vehicle changes during track-
ing. The cause of the apparent change is a geometric
change due to the rotation of the vehicle and the per-
spective projection of the camera, and a change of the
maid pattern due to the optical environment change.
Our tracking target is assumed to be a straight high-
way, and the geometric change due to the rotation of
the vehicle is small, but in order to track from 20m
to 140m far from the camera, the geometric change
due to perspective projection is Must be considered.
In the case of an IR camera, the pixel value is not an
absolute amount of heat radiation, but is obtained as
a relative value with respect to the surrounding pixel
values, so that the gain and offset of the luminance
pattern are not unchanged. For this reason, the ve-
hicle model uses a pattern on the rear end surface of
the vehicle that is nearly perpendicular to the camera
optical axis with little change in shape in perspective
projection from near to far. Use of zero-mean normal-
ized cross-correlation has the advantage that changes
in gain and offset do not affect the correlation value.
Note that vehicle model (template image) should be
shrinked as the vehicle moves away.
q(u,v) =
∑
x,y
T
i
(x,y) · (I(u+ x,v+ y) −
¯
I)
q
∑
x,y
T
i
(x,y)
2
∑
x,y
(I(u + x, v+ y) −
¯
I)
2
(1)
where q(u,v), I(u,v),
¯
I, T
i
(x,y) and
¯
T are the simi-
larity score at image location u,v, the image value at
location u,v, the mean of I(u, v), the model (template)
value, and the mean of T(x,y) respectively. Note that
T is adjusted so as to
¯
T = 0.
A new tracking location in the i-th image is given
by the following equation:
x,y = arg max
x,y
q
x,y
(2)
4 EXPERIMENTS AND
DISCUSSION
The proposed system was prototyped and its perfor-
mance was evaluated on a highway. The system was
setup over a four-lane highway in Kobe City’s Mi-
natojima, where sensors and cameras are temporarily
placed on a pedestrian bridge that crosses the road,
data is acquired and saved in a file system, and the
evaluation was done off-line. The tracking success
rate and the accuracy was verified. The IR cam-
era was a traffic surveillance camera made by FLIR
(image resolution 640 x 480 pixels, wavelength 10
mm), and the 2D-LiDAR was make by SICK. Fig-
ure 5 shows the correspondence between the scene of
the test site as seen from the aerial photograph and
the image taken with the IR camera. From a few tens
of minutes of video taken on this site, 100 vehicles
were selected randomly, and the rear end position of
each vehicle in each frame was manually marked as
the ground truth.
Figure 5: Captured image of the test site and the corre-
sponding aerial photo. (taken from Google map).
Figure 8 shows the tracking success rate with re-
spect to the distance from the camera. Tracking suc-
cess is the apparent tracking (Figure, red line) where
the tracking is continued in the algorithm, and the true
success rate (Figure, blue line) being tracked within
a certain range of error from the ground truth. A
tracking failure of several percent was observed near
the distance from the camera exceeding 70m, and in-
creased to about 20% at 150m, which ended the track-
ing. From this, it can be concluded that this system
can track up to a distance of 70m from the vicinity of
the camera with high reliability. There is also a 3%
tracking failure from close to 70m. The reason is that
the initial model (template) was generated almost cor-
rectly, but the search failed in the next frame. This is
considered to be because the template magnification
prediction in the next frame has failed, and it is esti-
mated that there is an algorithmic problem in the por-
tion that predicts the template magnification through
the velocity measurement and map-to-camera coordi-
nate conversion matrix.
5 CONCLUSION AND FUTURE
WORK
Our proposal demonstrates the effectiveness of on-
the-fly vehicle model generation with vehicle detec-
tion using LiDAR. So far, a data set that enables IR
vehicle image tracking only from images has not been