object is seen at an oblique angle. In the second
sequence the feature-based algorithm performs bet-
ter than the extended ESM algorithm (see figure 4).
However neither algorithm can track the whole se-
quence. Our tracking approach on the other hand
successfully tracks both sequences entirely, because it
changes the tracking algorithm used at the right mo-
ment. We obtained similar results on all synthetic
sequences we simulated. Since there are no blur-
ring, illumination changes or noise in the synthetic
sequences it is not possible to show how our system
deals with these conditions. Therefore we also per-
formed many real-world experiments using different
objects.
Figure 5 shows some experiments on real sequences
made with a tea box and a candy box under vary-
ing tracking conditions. The images show how our
system deals with partial occlusions (b,d,f,g,j), illu-
mination changes (c), changes in scale (b,g,i,h) and
severely oblique viewing angles (k,l). This shows that
the proposed algorithm is able to deal with dynamic
scenarios and solve the major limitations of classical
tracking algorithms such as partial occlusions, illu-
mination changes and fast object movement. We can
also see that it is possible to robustly overlay virtual
objects in order to perform Augmented Reality.
5 CONCLUSION
We presented a tracking system which intelligently
combines template-based and feature-based tracking.
The contributions are the extension of the ESM al-
gorithm, the formulation of the feature-based track-
ing and the FSM for deciding which algorithm to use
for the current frame. The system has been tested on
real-world sequences as well as on simulations and
performs at high frame rates on a standard PC.
Compared to other algorithms proposed in the litera-
ture we achieve a higher frame rate and more robust-
ness to fast object motions. Our approach also gives
good results in the face of partial occlusions and illu-
mination changes.
REFERENCES
Baker, S., Dellaert, F., and Matthews, I. (2001). Aligning
images incrementally backwards. Technical report,
Robotics Institute, Carnegie Mellon University.
Baker, S. and Matthews, I. (2001). Equivalence and effi-
ciency of image alignment algorithms. In IEEE Int.
Conf. on Computer Vision and Pattern Recognition,
pages 1090–1097.
Bay, H., Tuytelaars, T., and van Gool, L. (2006). SURF:
Speeded up robust features. European Conference on
Computer Vision.
Benhimane, S. and Malis, E. (2004). Real-time image-
based tracking of planes using efficient second-order
minimization. In IEEE/RSJ Int. Conf. on Intelligent
Robots Systems, pages 943–948.
Bouthemy, P. (1989). A maximum likelihood framework
for determining moving edges. IEEE Trans. Pattern
Anal. Mach. Intell., 11(5):499–511.
Fischler, M. and Bolles, R. (1981). Random sample consen-
sus: A paradigm for model fitting with applications to
image analysis and automated cartography. Commu-
nications of the ACM, 24(6):381–395.
Hager, G. and Belhumeur, P. (1998). Efficient region track-
ing with parametric models of geometry and illumina-
tion. IEEE Trans. on Pattern Analysis and Machine
Intelligence, 20(10):1025–1039.
Harris, C. and Stephens, M. (1988). A combined corner and
edge detector. In Proceedings of the 4th Alvey Vision
Conf., pages 147–151.
Hartley, R. I. and Zisserman, A. (2004). Multiple View Ge-
ometry in Computer Vision. Cambridge University
Press, second edition.
Lepetit, V., Lagger, P., and Fua, P. (2005). Randomized
trees for real-time keypoint recognition. In IEEE Int.
Conf. on Computer Vision and Pattern Recognition,
pages 775–781.
Lowe, D. (2004). Distinctive image features from scale-
invariant keypoints. International Journal of Com-
puter Vision, 60(2):91–110.
Lucas, B. and Kanade, T. (1981). An iterative image regis-
tration technique with application to stereo vision. In
Int. Joint Conf. on Artificial Intelligence, pages 674–
679.
Masson, L., Dhome, M., and Jurie, F. (2004). Robust real
time tracking of 3d objects. In Int. Conf. on Pattern
Recognition, pages 252–255.
Pressigout, M. and Marchand, E. (2005). Real-time planar
structure tracking for visual servoing: a contour and
texture approach. In IEEE/RSJ Int. Conf. on Intelli-
gent Robots and Systems.
Vacchetti, L., Lepetit, V., and Fua, P. (2004). Combin-
ing edge and texture information for real-time accu-
rate 3d camera tracking. In Proceedings of the Third
IEEE and ACM International Symposium on Mixed
and Augmented Reality, pages 48–57.
Wuest, H., Vial, F., and Stricker, D. (2005). Adaptive line
tracking with multiple hypotheses for augmented re-
ality. In Proceedings of the Fourth IEEE and ACM In-
ternational Symposium on Mixed and Augmented Re-
ality, pages 62–69.
Zhang, Z., Deriche, R., Faugeras, O., and Luong, Q.-T.
(1994). A ro bust technique for matching two uncal-
ibrated images through the recovery of the unknown
epipolar geometry. Technical Report 2273, INRIA.
VISAPP 2007 - International Conference on Computer Vision Theory and Applications
332