the point-based approach, here it fails four times be-
cause too few SIFT feature matches are found. Con-
trary, our algorithm processes the whole sequence.
We observe, that with our algorithm the reconstructed
trajectory resembles the true trajectory the most, since
it has almost a rectangular form and less variation in
the height dimension then the other approaches. This
indicates that our rotation estimation is more robust.
Please note that due to inaccurate distance measure-
ments of our robot, the overall scaling of the trajecto-
ries is not correct.
The average runtime per image on the “Office Cir-
cle” sequence is for the triplet approach 3277 ms, for
the point-based approach 146 ms and for our approach
85 ms.
We can conclude that our method is better suited
for the visual odometry task than the “Triplet” ap-
proach. We also demonstrated that in low textured
environments a line-based approach can function as
replacement for point-based relative pose estimation.
6 CONCLUSION
In this paper, we presented a novel relative pose esti-
mation scheme using lines. In our approach, we esti-
mate the 3D line directions through a clustering step
of parallel lines in the world and use this information
throughout the whole processing pipeline. The direc-
tion information is used to guide the line matching
and to calculate the relative rotation. We also pre-
sented a visual odometry using our relative pose esti-
mation.
As the 3D line direction estimation is such an im-
portant step, we evaluated it on synthetic data and
showed how the usage of “direction priors” in a se-
quential processing boosts the accuracy.
Furthermore, we compared our relative pose esti-
mation to the state-of-the-art approach from Elqursh
and Elgammal (Elqursh and Elgammal, 2011). We
showed that our method outperforms theirs in terms
of accuracy and runtime. Especially the runtime can
be reduced from seconds to milliseconds.
The visual odometry system is evaluated and com-
pared to the state-of-the-art for line-based relative
pose estimation (Elqursh and Elgammal, 2011) and a
comparable point-based algorithm (Nister, 2004). We
showed that our method is better applicable in tex-
tureless indoor scenarios than both other approaches.
In the future, we want to extend our approach to a
complete SLAM system. To reach this goal, we need
to relax the restriction to small motions in the guided
matching step. How this could be done is future work.
REFERENCES
Antone, M. E. and Teller, S. (2000). Automatic recovery
of relative camera rotations for urban scenes. In IEEE
Conference on Computer Vision and Pattern Recogni-
tion. CVPR 2000, pages 282–289.
Bartoli, A. and Sturm, P. (2005). Structure-from-motion
using lines: Representation, triangulation, and bundle
adjustment. Computer Vision and Image Understand-
ing, 100(3):416–441.
Bazin, J., Demonceaux, C., Vasseur, P., and Kweon, I.
(2010). Motion estimation by decoupling rotation and
translation in catadioptric vision. Computer Vision
and Image Understanding, 114(2):254–273.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977).
Maximum likelihood from incomplete data via the em
algorithm. Journal of the Royal Statistical Society. Se-
ries B (Methodological), pages 1–38.
Duda, R. O. and Hart, P. E. (1972). Use of the hough trans-
formation to detect lines and curves in pictures. Com-
munications of the ACM, 15(1):11–15.
Elqursh, A. and Elgammal, A. (2011). Line-based relative
pose estimation. In 2011 IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR), pages
3049–3056.
Fischler, M. A. and Bolles, R. C. (1981). Random sample
consensus: a paradigm for model fitting with appli-
cations to image analysis and automated cartography.
Communications of the ACM, 24(6):381–395.
Gioi, R. v., Jakubowicz, J., Morel, J.-M., and Randall, G.
(2010). Lsd: A fast line segment detector with a false
detection control. IEEE Transactions on Pattern Anal-
ysis and Machine Intelligence, 32(4):722–732.
Gower, J. C. and Dijksterhuis, G. B. (2004). Procrustes
problems, volume 3. Oxford University Press.
Hartley, R. I. (1994). Projective reconstruction from line
correspondences. In Proceedings of IEEE Conference
on Computer Vision and Pattern Recognition, pages
903–907.
Hartley, R. I. and Zisserman, A. (2004). Multiple View Ge-
ometry in Computer Vision. Cambridge University
Press, ISBN: 0521540518, second edition.
Hirose, K. and Saito, H. (2012). Fast line description for
line-based slam. In Bowden, R., Collomosse, J., and
Mikolajczyk, K., editors, British Machine Vision Con-
ference 2012, pages 83.1–83.11.
Ko
ˇ
seck
´
a, J. and Zhang, W. (2002). Video compass. In Goos,
G., Hartmanis, J., Leeuwen, J., Heyden, A., Sparr, G.,
Nielsen, M., and Johansen, P., editors, Computer Vi-
sion — ECCV 2002, volume 2353 of Lecture Notes
in Computer Science, pages 476–490. Springer Berlin
Heidelberg, Berlin, Heidelberg.
Lowe, D. G. (2004). Distinctive image features from scale-
invariant keypoints. International Journal of Com-
puter Vision, 60(2):91–110.
Nister, D. (2004). An efficient solution to the five-point
relative pose problem. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 26(6):756–770.
VISAPP 2016 - International Conference on Computer Vision Theory and Applications
432