paved roads. There are two main contributions. First,
we have combined seven public driving and robotics
datasets, which together contain a large variety of
road types, and trained a HRNet-W48 network on this
data to achieve robust road segmentation. Second,
we have developed a path-finding algorithm and im-
proved its robustness to incorrect road segmentation
in several ways, allowing for automated control of a
vehicle-mounted PTZ camera, which can handle road
crossings and forks.
Our experimental results have shown that individ-
ual driving datasets contain insufficient variety to al-
low training of a robust road segmentation system
for all road types. Combining seven different pub-
lic datasets and adding just a small number of semi-
automatically labeled images greatly improves the
performance to the point that all roads in the test
dataset can be accurately segmented. Although this
paper is primarily concerned with camera control, the
robust road segmentation for a broad class of roads
can also be of interest for work in autonomous driv-
ing.
Overall, we conclude that robust path planning on
any type of road is feasible, but will still require com-
parable extensive datasets that autonomous driving re-
search has generated for urban environments over the
past few years. Until then, combining many existing
datasets is a good alternative for generalization.
REFERENCES
Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler,
M., Benenson, R., Franke, U., Roth, S., and Schiele,
B. (2016). The Cityscapes Dataset for Semantic Urban
Scene Understanding. In CVPR 2016.
Giusti, A., Guzzi, J., Ciresan, D. C., He, F. L., Rodriguez,
J. P., Fontana, F., Faessler, M., Forster, C., Schmid-
huber, J., Caro, G. D., Scaramuzza, D., and Gam-
bardella, L. M. (2016). A Machine Learning Ap-
proach to Visual Perception of Forest Trails for Mo-
bile Robots. IEEE Robotics and Automation Letters,
1(2):661–667.
Gonz
´
alez, D., P
´
erez, J., Milan
´
es, V., and Nashashibi, F.
(2016). A Review of Motion Planning Techniques for
Automated Vehicles. IEEE Transactions on Intelli-
gent Transportation Systems, 17(4):1135–1145.
Grigorescu, S., Trasnea, B., Cocias, T., and Macesanu,
G. (2020). A survey of deep learning techniques
for autonomous driving. Journal of Field Robotics,
37(3):362–386.
Jiang, P., Osteen, P., Wigness, M., and Saripalli, S. (2020).
RELLIS-3D dataset: Data, benchmarks and analysis.
arXiv 2011.12954v2.
Maturana, D., Chou, P.-W., Uenoyama, M., and Scherer,
S. (2018). Real-Time Semantic Mapping for Au-
tonomous Off-Road Navigation. In Springer Proceed-
ings in Advanced Robotics, pages 335–350.
Nakagomi, H., Fuse, Y., Nagata, Y., Hosaka, H., Miyamoto,
H., Yokozuka, M., Kamimura, A., Watanabe, H., Tan-
zawa, T., and Kotani, S. (2021). Forest road sur-
face detection using LiDAR-SLAM and U-Net. 2021
IEEE/SICE International Symposium on System Inte-
gration, pages 727–732.
Neuhold, G., Ollmann, T., Bulo, S. R., and Kontschieder,
P. (2017). The Mapillary Vistas Dataset for Semantic
Understanding of Street Scenes. In ICCV 2017.
Paden, B.,
ˇ
C
´
ap, M., Yong, S. Z., Yershov, D., and Fraz-
zoli, E. (2016). A survey of motion planning and con-
trol techniques for self-driving urban vehicles. IEEE
Transactions on Intelligent Vehicles, 1(1):33–55.
Qureshi, A. H., Simeonov, A., Bency, M. J., and Yip, M. C.
(2019). Motion planning networks. In Proceedings -
IEEE International Conference on Robotics and Au-
tomation, volume 2019-May, pages 2118–2124.
Ronneberger, O., Fischer, P., and Brox, T. (2015). U-Net
: Convolutional Networks for Biomedical. In Med-
ical Image Computing and Computer-Assisted Inter-
vention (MICCAI), pages 234–241.
Shelhamer, E., Long, J., and Darrell, T. (2016). Fully con-
volutional networks for semantic segmentation. IEEE
Transactions on Pattern Analysis and Machine Intel-
ligence, 39(4):640–651.
Sun, K., Zhao, Y., Jiang, B., Cheng, T., Xiao, B., Liu, D.,
Mu, Y., Wang, X., Liu, W., and Wang, J. (2019). High-
resolution representations for labeling pixels and re-
gions. arXiv 1904.04514v1.
Valada, A., Oliveira, G. L., Brox, T., and Burgard, W.
(2016). Deep Multispectral Semantic Scene Under-
standing of Forested Environments Using Multimodal
Fusion. In International Symposium on Experimental
Robotics (ISER).
Varma, G., Subramanian, A., Nboodiri, A., and Chandraker,
M. (2019). IDD : A Dataset for Exploring Problems of
Autonomous Navigation in Unconstrained Environ-
ments. In IEEE Winter Conference on Applications
of Computer Vision (WACV).
Wigness, M., Eum, S., Rogers, J. G., Han, D., and Kwon,
H. (2019). A RUGD Dataset for Autonomous Navi-
gation and Visual Perception in Unstructured Outdoor
Environments. IEEE International Conference on In-
telligent Robots and Systems, pages 5000–5007.
Yadav, S., Patra, S., Arora, C., and Banerjee, S. (2017).
Deep CNN with Color Lines Model for Unmarked
Road Segmentation. In IEEE International Confer-
ence on Image Processing (ICIP), pages 585–589.
Yu, F., Chen, H., Wang, X., Xian, W., Chen, Y., Liu, F.,
Madhavan, V., and Darrell, T. (2020). BDD100K: A
Diverse Driving Dataset for Heterogeneous Multitask
Learning. In CVPR 2020, pages 2633–2642.
Zendel, O., Honauer, K., Murschitz, M., Steininger, D.,
and Dom
´
ınguez, G. F. (2018). WildDash - Creating
Hazard-Aware Benchmarks. In Proceedings of ECCV,
pages 407–421.
Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017). Pyra-
mid Scene Parsing Network. In CVPR.
Robust Path Planning in the Wild for Automatic Look-Ahead Camera Control
561