Authors:
Oguz Kedilioglu
1
;
Markus Lieret
1
;
Julia Schottenhamml
2
;
Tobias Würfl
2
;
Andreas Blank
1
;
Andreas Maier
2
and
Jörg Franke
1
Affiliations:
1
Institute for Factory Automation and Production Systems, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), 91058 Erlangen, Germanyy
;
2
Pattern Recognition Lab, Friedrich-Alexander- Universität Erlangen-Nürnberg (FAU), 91058 Erlangen, Germany
Keyword(s):
Image Segmentation, Object Recognition, Neural Networks, Deep Learning, Robotics, Autonomous Mobile Robots, Flexible Automation, Warehouse Automation.
Abstract:
Automated guided vehicles (AGV) are nowadays a common option for the efficient and automated in-house
transportation of various cargo and materials. By the additional application of unmanned aerial vehicles (UAV)
in the delivery and intralogistics sector this flow of materials is expected to be extended by the third dimension
within the next decade.
To ensure a collision-free movement for those vehicles optical, ultrasonic or capacitive distance sensors are
commonly employed. While such systems allow a collision-free navigation, they are not able to distinguish
humans from static objects and therefore require the robot to move at a human-safe speed at any time. To overcome these limitations and allow an environment sensitive collision avoidance for UAVs and AGVs we provide a solution for the depth camera based real-time semantic segmentation of workers in industrial environments.
The semantic segmentation is based on an adapted version of the deep convolutional neural network
(CNN)
architecture FuseNet. After explaining the underlying methodology we present an automated approach for the
generation of weakly annotated training data and evaluate the performance of the trained model compared to
other well-known approaches
(More)