Improving Stereo Vision Odometry for Mobile Robots
in Outdoor Dynamic Environments
Dan Pojar and Sergiu Nedevschi
Department of Computer Science, Technical University Cluj-Napoca, Cluj-Napoca, Romania
Keywords: Visual Odometry, Stereo Vision, Localization, Mobile Robot, Real-Time.
Abstract: This article presents a method for localization able to provide the pose in 3D using stereo vision. The
method offers a better and inexpensive alternative to classical localization methods such as wheel odometry
or GPS. Using only a calibrated stereo camera, the method integrates both optical flow based motion
computation and SURF features detector for stereo reconstruction and motion computation. Robustness is
obtained by finding correspondences using both feature descriptors and RANSAC inlier selection for the
reconstructed points. Least squares optimization is used to obtain the final computed motion. World scale
pose estimation is obtained by computing successive motion vectors characterized through their orientation
and magnitude. The method involves fast algorithms capable to function at real time frequency. We present
results supporting global consistency, localization performance and speed as well as the robustness of the
approach by testing it in unmodified, real life, very crowded outdoor dynamic environments.
1 INTRODUCTION
Recent trends in mobile robotics deal with a
fundamental requirement of any robot, the
possibility of localizing itself. As robots moved from
a highly deterministic environment the proposed
solutions for localization had to deal with more and
more difficult scenarios.
Solutions that used both custom infrastructure
and expensive sensor configurations exist already.
Wheel odometry is the most commonly encountered
solution that allows easy and cheap localization but
is reliable only for a few tens of meters, at best, due
to accumulated measurement errors.
A different alternative that has previously been
explored but which only recently has been shown to
provide better results relies on using cameras as
sensors. Our paper focuses on the type of approaches
named structure from motion. Different methods can
also use different types of cameras, but most work is
based on monocular or stereo cameras.
Structure from motion methods allow recovering
both scene geometric structure and camera extrinsic
as well as intrinsic parameters from sets of images of
the scene taken from different poses.
2 RELATED WORK
One of the most cited approaches by (Nister, 2006)
presents solutions for both monocular and stereo
setups. The 5 point algorithm or in the case of a
stereo camera, a 3 point perspective method referred
therein is enough to compute the relative pose
change. Other work presented by (Konolige and
Agrawal, 2007) is also based on stereo vision. In this
case, the authors use 3D triangulation from stereo.
A similar approach used among other purposes
for the Boston Dynamics Big Dog robot is given by
(Howard, 2008). Other sensors such as expensive
IMU are used to offer a reference. (Scaramuzza,
2009) presents a solution tested on an
omnidirectional camera that uses only one point to
compute the motion hypothesis for RANSAC outlier
rejection. This is possible because the motion model
is restricted to planar motion. A simplified stereo
motion model which can be directly computed is
presented in (Jeong, 2010). The author uses robot
wheel odometry to correct errors that occur from
stereo based motion computation.
In our previous work (Pojar, 2010) we presented
an approach that simplified to planar motion, similar
to (Scaramuzza, 2009) using a stereo camera.
476
Pojar D. and Nedevschi S..
Improving Stereo Vision Odometry for Mobile Robots in Outdoor Dynamic Environments.
DOI: 10.5220/0004043404760480
In Proceedings of the 9th International Conference on Informatics in Control, Automation and Robotics (ICINCO-2012), pages 476-480
ISBN: 978-989-8565-22-8
Copyright
c
2012 SCITEPRESS (Science and Technology Publications, Lda.)