Figure 5: Camera moving front and back along a fixed
3.30m studio rail. On the left, three images taken from
the beginning, middle and the end of the rail. Green re-
gions illustrated selected points. On the right, comparison
between dense tracking and keyframe tracking. In dense
tracking drift increases in time, but keyframe tracking main-
tains small bounded error. A person is moving in the scene
during the last four cycles.
ence frame and the current frame. To remove the
global drift in incremental tracking, the closest pre-
recorded keyframe was chosen to be the motion refer-
ence. The system was designed to be an affordable so-
lution for TV broadcasting studios relying only on the
Kinect sensor and a commodity laptop. The proposed
approach performs robustly in a standard benchmark,
where KinectFusion has problems with planar sur-
faces and limited voxel grid resolution. Our future
work will address the practical issues how studio staff
and camera men can use our computer vision system
in live broadcasts. Moreover, combination of the best
properties of our approach and KinectFusion will be
investigated.
REFERENCES
Audras, C., Comport, A. I., Meilland, M., and Rives, P.
(2011). Real-time dense rgb-d localisation and map-
ping. In Australian Conference on Robotics and Au-
tomation. Monash University, Australia, 2011.
Baker, S. and Matthews, I. (2004). Lucas-kanade 20 years
on: A unifying framework. Int. J. Comput. Vision,
56(3):221–255.
Bouguet, J.-Y. (2010). Camera calibration toolbox
for matlab. http://www.vision.caltech.edu/bouguetj/
calib doc.
Comport, A., Malis, E., and Rives, P. (2007). Accu-
rate quadri-focal tracking for robust 3d visual odome-
try. In IEEE Int. Conf. on Robotics and Automation,
ICRA’07, Rome, Italy.
Davison, A., Reid, I., Molton, N., and Stasse, O. (2007).
MonoSLAM: Real-time single camera SLAM. PAMI,
29:1052–1067.
Dobbert, T. (2005). Matchmoving: The Invisible Art of
Camera Tracking. Sybex.
Henry, P., Krainin, M., Herbst, E., Ren, X., and Fox, D.
(2012). RGB-D mapping: Using Kinect-style depth
cameras for dense 3D modeling of indoor environ-
ments. The International Journal of Robotics Re-
search, 31(5):647–663.
Herrera, C., Kannala, J., and Heikkila, J. (2012). Joint depth
and color camera calibration with distortion correc-
tion. IEEE PAMI, 34(10).
Kato, H. and Billinghurst, M. (1999). Marker tracking and
hmd calibration for a video-based augmented reality
conferencing system. In Proceedings of the 2nd Inter-
national Workshop on Augmented Reality (IWAR 99),
San Francisco, USA.
Klein, G. and Murray, D. (2007). Parallel tracking and map-
ping for small ar workspaces. Proceedings of the In-
ternational Symposium on In Mixed and Augmented
Reality (ISMAR), pages 225–234.
Ma, Y., Soatto, S., Kosecka, J., and Sastry, S. (2004). An in-
vitation to 3-D vision: from images to geometric mod-
els, volume 26 of Interdisciplinary applied mathemat-
ics. Springer, New York.
Newcombe, R., Lovegrove, S., and Davison, A. (2011a).
Dtam: Dense tracking and mapping in real-time. In
ICCV, volume 1.
Newcombe, R. A., Izadi, S., Hilliges, O., Molyneaux, D.,
Kim, D., Davison, A. J., Kohli, P., Shotton, J., Hodges,
S., and Fitzgibbon, A. (2011b). Kinectfusion: Real-
time dense surface mapping and tracking. ISMAR,
pages 127–136.
Rusu, R. B. and Cousins, S. (2011). 3D is here: Point Cloud
Library (PCL). In IEEE International Conference on
Robotics and Automation (ICRA), Shanghai, China.
Snavely, N., Seitz, S. M., and Szeliski, R. (2006). Photo
tourism: Exploring photo collections in 3d. In ACM
TRANSACTIONS ON GRAPHICS, pages 835–846.
Press.
Sturm, J., Magnenat, S., Engelhard, N., Pomerleau, F., Co-
las, F., Burgard, W., Cremers, D., and Siegwart, R.
(2011). Towards a benchmark for rgb-d slam evalua-
tion. In Proc. of the RGB-D Workshop on Advanced
Reasoning with Depth Cameras at Robotics: Science
and Systems Conf. (RSS), Los Angeles, USA.
Triggs, B., McLauchlan, P., Hartley, R., and Fitzgibbon, A.
(2000). Bundle adjustment - a modern synthesis. In
Triggs, B., Zisserman, A., and Szeliski, R., editors, Vi-
sion Algorithms: Theory and Practice, volume 1883
of Lecture Notes in Computer Science, pages 298–
372. Springer-Verlag.
Tykkala, T. M., Audras, C., and Comport, A. (2011). Direct
iterative closest point for real-time visual odometry. In
ICCV Workshop CVVT.
VISAPP2013-InternationalConferenceonComputerVisionTheoryandApplications
252