gorithm by using the depth information provided by
a LIDAR sensor to guide the stereo correspondence
process by limiting the disparity search interval. The
advantage of the method is that it generates fused
and dense point clouds which can then be used for
a more accurate detection of protruding obstacles in
the scene. The principle of fusion is that the more
reliable LIDAR sensor which however has a low res-
olution guides the less reliable but higher resolution
sensor. The obstacle detection is performed in two
ways on that data: by data clustering and by using a
CNN.
The results presented in Subsection 6.1 show that
the fused point clouds lead to a better 2D clustering
result, where the accuracy in detection can be up to
27% better than in the case where the Block Matching
point clouds are used. The simple clustering is not
able to detect which type of obstacle is there, only
that it is present.
The CNN is able to detect different classes. Our
results show that it is able to accurately detect humans
and trees in the fused point clouds while distinguish-
ing between them. The guidance through the LIDAR
information significantly improves the detection F1
score of the trained networks.
ACKNOWLEDGEMENTS
This work was supported by the project SAFE Per-
ception, funded by the Danish Innovation Fund.
REFERENCES
Aeschimann, R. and Borges, P. V. K. (2015). Ground or
obstacles? detecting clear paths in vehicle navigation.
In 2015 IEEE Int. Conf. on Robotics and Automation
(ICRA), pages 3927–3934.
Badino, H., Huber, D., and Kanade, T. (2011). Integrat-
ing lidar into stereo for fast and improved disparity
computation. In 3D Imaging, Modeling, Processing,
Visualization and Transmission, Int. Conf. on, pages
405–412.
Bradski, G. (2000). The opencv library. Dr. Dobb’s Jour-
nal: Software Tools for the Professional Programmer,
25(11):120–123.
Dagum, L. and Menon, R. (1998). Openmp: an industry
standard api for shared-memory programming. IEEE
computational science and engineering, 5(1):46–55.
English, A., Ross, P., Ball, D., and Corke, P. (2014). Vi-
sion based guidance for robot navigation in agricul-
ture. In 2014 IEEE Int. Conf. on Robotics and Au-
tomation (ICRA), pages 1693–1698.
Kragh, M. F., Christiansen, P., Laursen, M. S., Larsen, M.,
Steen, K. A., Green, O., Karstoft, H., and Jørgensen,
R. N. (2016). Fieldsafe: Dataset for obstacle detection
in agriculture. CoRR.
Long, J., Shelhamer, E., and Darrell, T. (2014). Fully
convolutional networks for semantic segmentation.
CoRR.
Nakamura, K., Kimura, M., Anazawa, T., Takahashi, T., and
Naruse, K. (2016). Investigation of weeding ability
and plant damage for rice field weeding robots. In
2016 IEEE/SICE International Symposium on System
Integration (SII), pages 899–905.
Qi, C. R., Su, H., Mo, K., and Guibas, L. J. (2016). Pointnet:
Deep learning on point sets for 3d classification and
segmentation. CoRR.
Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T.,
Leibs, J., Wheeler, R., and Ng, A. Y. (2009). Ros: an
open-source robot operating system. In ICRA work-
shop on open source software, page 5. Kobe.
Redmon, J. and Farhadi, A. (2016). Yolo9000: Better,
faster, stronger. CoRR.
Romero, A. R., Borges, P. V. K., Elfes, A., and Pfrunder, A.
(2016). Environment-aware sensor fusion for obsta-
cle detection. In 2016 IEEE Int. Conf. on Multisensor
Fusion and Integration for Intelligent Systems (MFI),
pages 114–121.
Rusu, R. B. and Cousins, S. (2011). 3D is here: Point cloud
library (PCL). In Robotics and automation (ICRA),
2011 IEEE Int. Conf. on, pages 1–4. IEEE.
Sharma, S. and Borse, R. (2016). Automatic agricul-
ture spraying robot with smart decision making. In
Corchado Rodriguez, Juan Manueland Mitra, S. and
Thampi, Sabu M.and El-Alfy, E.-S., editors, Intel-
ligent Systems Technologies and Applications 2016,
pages 43–758. Springer International Publishing.
Somanath, G., Cohen, S., Price, B., and Kambhamettu, C.
(2013). Stereo+kinect for high resolution stereo cor-
respondences. In 3D Vision, Int. Conf. on, pages 9–16.
Song, X., Shan, H., Liu, H., and Guo, J. (2016). An under-
actuated end-effector design for fruit picking. In 2016
23rd Int. Conf. on Mechatronics and Machine Vision
in Practice (M2VIP), pages 1–5. IEEE.
Suvei, S.-D., Bodenhagen, L., Kiforenko, L., Christiansen,
P., Jørgensen, R. N., Buch, A. G., and Kr
¨
uger, N.
(2016). Stereo and active-sensor data fusion for im-
proved stereo block matching. In Int. Conf. Image
Analysis and Recognition, pages 451–461. Springer.
Teichmann, M., Weber, M., Zoellner, M., Cipolla, R., and
Urtasun, R. (2016). Multinet: Real-time joint seman-
tic reasoning for autonomous driving. CoRR.
Zhang, F., Clarke, D., and Knoll, A. (2014). Vehicle detec-
tion based on lidar and camera fusion. In IEEE 17th
Int. Conf. on Intelligent Transportation Systems, pages
1620–1625.
Stereo and LIDAR Fusion based Detection of Humans and Other Obstacles in Farming Scenarios
173