superpixels to generate the coarse 3D primitives (su-
persurfels) composing the predicted model guarantee
the preservation of most of the relevant and meaning-
ful information from the observed scene.
The reconstruction system proposed, Supersurfel-
Fusion, based on this representation, rebuild a low-
resolution but pertinent 3D model with high speed
performance and low memory usage. The use of this
system is of interest for robots which do not need a
very accurate but rather an efficient, fast and light-
weight 3D map generation, enabling them to perform
other tasks at the same time without consuming too
much of the resources available.
As future works, we would like to integrate our
own camera tracking solution to this system, as frame
to model registration strategy, and make it robust to
dynamic environments by detecting moving objects
at a superpixel level and tracking them. Further opti-
mization can be achieved to speed up the computation
of supersurfels attributes by minimizing branch diver-
gence.
REFERENCES
B
´
odis-Szomor
´
u, A., Riemenschneider, H., and Gool, L. V.
(2014). Fast, approximate piecewise-planar modeling
based on sparse structure-from-motion and super-
pixels. In 2014 IEEE Conference on Computer Vision
and Pattern Recognition, pages 469–476.
Curless, B. and Levoy, M. (1996). A volumetric method for
building complex models from range images. In Pro-
ceedings of the 23rd Annual Conference on Compu-
ter Graphics and Interactive Techniques, SIGGRAPH
’96, pages 303–312, New York, NY, USA. ACM.
Davis, J. V. and Dhillon, I. (2006). Differential entropic
clustering of multivariate gaussians. In Proceedings
of the 19th International Conference on Neural Infor-
mation Processing Systems, NIPS’06. MIT Press.
Handa, A., Whelan, T., McDonald, J., and Davison, A.
(2014). A benchmark for RGB-D visual odometry,
3D reconstruction and SLAM. In IEEE Intl. Conf. on
Robotics and Automation, ICRA.
Henry, P., Fox, D., Bhowmik, A., and Mongia, R. (2013).
Patch volumes: Segmentation-based consistent map-
ping with rgb-d cameras. In 3DV. IEEE Computer
Society.
Izadi, S., Kim, D., Hilliges, O., Molyneaux, D., Newcombe,
R., Kohli, P., Shotton, J., Hodges, S., Freeman, D.,
Davison, A., and Fitzgibbon, A. (2011). Kinectfusion:
Real-time 3d reconstruction and interaction using a
moving depth camera. In Proceedings of the 24th An-
nual ACM Symposium on User Interface Software and
Technology. ACM.
Karras, T. (2012). Maximizing parallelism in the con-
struction of bvhs, octrees, and k-d trees. In Procee-
dings of the Fourth ACM SIGGRAPH / Eurographics
Conference on High-Performance Graphics, EGGH-
HPG’12. Eurographics Association.
Kay, T. L. and Kajiya, J. T. (1986). Ray tracing complex
scenes. In Proceedings of the 13th Annual Confe-
rence on Computer Graphics and Interactive Techni-
ques, SIGGRAPH ’86. ACM.
Keller, M., Lefloch, D., Lambers, M., Izadi, S., Weyrich,
T., and Kolb, A. (2013). Real-time 3d reconstruction
in dynamic scenes using point-based fusion. In 2013
International Conference on 3D Vision - 3DV 2013.
Mur-Artal, R. and Tard
´
os, J. D. (2017). ORB-SLAM2: an
open-source SLAM system for monocular, stereo and
RGB-D cameras. IEEE Transactions on Robotics.
Ren and Malik (2003). Learning a classification model for
segmentation. In Proceedings Ninth IEEE Internatio-
nal Conference on Computer Vision.
Salas-Moreno, R. F., Glocken, B., Kelly, P. H. J., and Da-
vison, A. J. (2014). Dense planar slam. In 2014 IEEE
International Symposium on Mixed and Augmented
Reality (ISMAR), pages 157–164.
Steinbr
¨
ucker, F., Sturm, J., and Cremers, D. (2011). Real-
time visual odometry from dense rgb-d images. In
2011 IEEE International Conference on Computer Vi-
sion Workshops (ICCV Workshops).
Sturm, J., Engelhard, N., Endres, F., Burgard, W., and Cre-
mers, D. (2012). A benchmark for the evaluation of
rgb-d slam systems. In Proc. of the International Con-
ference on Intelligent Robot Systems (IROS).
Thomas, D. and Sugimoto, A. (2013). A flexible scene
representation for 3d reconstruction using an rgb-d
camera. In 2013 IEEE International Conference on
Computer Vision, pages 2800–2807.
Whelan, T., Kaess, M., Fallon, M., Johannsson, H., Leo-
nard, J., and McDonald, J. (2012). Kintinuous: Spati-
ally extended kinectfusion. In RSS Workshop on RGB-
D: Advanced Reasoning with Depth Cameras, Syd-
ney, Australia.
Whelan, T., Leutenegger, S., Moreno, R. S., Glocker, B.,
and Davison, A. (2015). Elasticfusion: Dense slam
without a pose graph. In Proceedings of Robotics:
Science and Systems.
Yamaguchi, K., McAllester, D., and Urtasun, R. (2014). Ef-
ficient joint segmentation, occlusion labeling, stereo
and flow estimation. In ECCV.
A Coarse and Relevant 3D Representation for Fast and Lightweight RGB-D Mapping
831