detection, gives a reliable feature tracker for the 3D
reconstruction and camera pose estimation process.
The time for the whole feature tracking
algorithm is insignificant compared with the camera
solving process. For example, a video of 340 frames
with a resolution of 704x576 needs approximately 5
seconds to search and track 300 features.
4.2 Testing the 3D Tracker
The strategy used to test the accuracy of the camera
pose estimation algorithm consists in comparing the
position of the features in the image with the
corresponding projections of the 3D points.
The next graph shows the mean of the error
measured along 100 frames.
0
0,05
0,1
0,15
0,2
0,25
0,3
0,35
0,4
0,45
0,5
1 6 11 16 21 26 31 36 41 46 51 56 61 66 71 76 81 86 91 96
Frame
Projection error (pixels)
Figure 8: Projection error.
The time needed to perform the 3D tracking
process is approximately one second per frame. This
is very far from the maximum of 40ms needed to run
the process in real time, but this is mainly because it
is implemented in Matlab.
5 CONCLUSIONS
This work covers all the processes involved in an
augmented video application. The method does not
need any knowledge of the augmented scene or user
interaction except in the registration step.
The advantage of this type of system is that any
video can be augmented imposing only a few
restrictions on it. Additionally, any user without
experience can augment videos in an easy way
because all the process is automatic.
In the first part of the work, a 2D feature tracker
has been developed. This tracker has proven to be
accurate enough for many applications, like 3D
reconstruction or camera pose estimation and it can
work in real time in a standard PC. This fact makes
the tracker suitable for surveillance, human
computer interaction or any application that needs
real time response.
Secondly, the designed 3D tracker can add
virtual objects to real videos. It depends heavily on
the accuracy of the feature tracker but the tests
demonstrate that the result is satisfactory under
normal conditions. On the other hand, actually the
prototype works under Matlab so the time needed to
run the tracker is very high. Thus, an immediate
objective is to translate the code into another
language, like C++. However, the proposed
algorithm is not proper for running in real time
because of the outlier search and the key frame
reconstruction based algorithm.
REFERENCES
Bouguet, J.-Y., "Pyramidal Implementation of the Lucas
Kanade Feature Tracker", Intel Corporation, Technical
Report 2000.
Cornelis, K., "From uncalibrated video to augmented
reality": Katholike Universiteit Leuven, 2004.
DeMenthon, D. & Davis, L., (1995). Model-Based Object
Pose in 25 lines of code. International Journal of
Computer Vision, 15, Pp. 123-141.
Hartley, R. & Zisserman, A., (2000). Multiple View
Geometry in computer vision: Cambridge University
Press.
Kalman, R., (1960). A New Approach to Linear Filtering
and Prediction Problems. Journal of Basic
Engineering, 82, Pp. 35-45.
Kato, H. & Billinghurst, M., (1999). Marker Tracking and
HMD Calibration for a video-based Augmented
Reality Conferencing System. In International
Workshop on Augmented Reality (IWAR), Pp. 85-94.
San Francisco, USA.
Mendonca, P. & Cipolla, R., (1999). A simple technique
for self-calibration. In IEEE Conference on Computer
Vision and Pattern Recognition, Pp. 112-116. Fort
Collins, Colorado.
Milgram, P., Takemura, H., Utsumi, A., & Kishino, F.,
(1994). Augmented Reality: A Class of Displays of the
Reality-Virtuality Continuum. In SPIE Conference on
Telemanipulator and Telepresence Technologies, Pp.
282-292. Boston, USA, October 31 - November 4.
Shi, J. & Tomasi, C., (1994). Good Features To Track. In
IEEE Conference on Computer Vision and Pattern
Recognition, Pp. 593-600. Seattle, Washington.
Torr, P., "A Structure and Motion Toolkit in Matlab",
Microsoft Research, Cambridge, UK, Technical
Report MSR-TR-2002-56, 2002.
VISAPP 2007 - International Conference on Computer Vision Theory and Applications
522