both Gradient Vector Flow (GVF) and Balloon Para-
metric Active Contour models. This technique was
not very accurate to detect lanes at the edges of the
road and the lane predicted interfered with the road
curbs. The same author published a lane detection
algorithm (Kumar et al., 2014) based on an inten-
sity threshold as well as Region Of Interest (ROI) to
limit the number of processed LiDAR data. Subse-
quently, the data was converted to a 2D image, with
a linear dilation to complete rubbed off lanes. The
algorithm was tested over 93 roads and managed to
detect the markings in 80 roads. The authors claimed
that the failures in detecting all the test roads was due
to road wipings and erosion, causing small intensi-
ties and low densities to be received by the LiDAR.
In (Guan et al., 2014) the point cloud is segmented
into horizontal blocks. These blocks are then used
to detect the edges (road curbs) based on differences
in elevation in order to determine the surface as well
as the road boundaries. The authors claim to accom-
plish a success of 0.83 of correctness to detect lane
markings. In (Thuy and Le
´
on, 2010) the authors de-
veloped a lane marking detection algorithm based on
a dynamic threshold. First, the data points received
from the LiDAR is processed using Probability Den-
sity Function (pdf), and the maximum reflectively is
matched with the highest values of the pdf. The dy-
namic threshold is applied to this reflectively data,
since the lane markings are the ones that return high
reflectively (due to their color gradient). The author
of (Yan et al., 2016) transformed the segmented points
into scan lines based on scanner angle. Consequently,
the road data is determined based on a ”Height Dif-
ference” (HD). The road limits have been identified
with a moving least square, that only accepts certain
points that lie in a certain threshold set by the authors.
Besides a classification on intensity values, the au-
thors proposed using an ”Edge Detection and Edge
Constraints” (EDEC) technique that detects fluctua-
tions in the intensity. This method should minimize
the noise in the detected lanes. The algorithm was
tested on data from Jincheng highway China, the au-
thors claim that they have accomplished an accuracy
level of 0.9.
1.1.2 Camera Lane Detection
Camera based lane marking detection research is a
main research field and currently heavily studied.
Therefore in this section, only a brief overview of re-
lated works is provided. The lane detection algorithm
in (Mu and Ma, 2014) converts the raw images to grey
scale and applies a Otsu’s method for thresholding the
image. Sobel is used to detect the lane markings. The
results shows that it is effective for incomplete lane
markings and for fluctuations in the environment’s il-
lumination. (Li et al., 2014) uses Canny edge detec-
tion technique which results in a binary image. Then
Hough transform is implemented in order to detect the
straight lines from the image. In contrast, the author
of (Haque et al., 2019) uses thresholding based on
gradients and the HLS color space. Followed by the
perspective transformation they apply a sliding win-
dow algorithms. The centroids of the windows a fi-
nally composed to a lane.
1.2 Our Approach
For the approach presented in this paper, we tackle
the main problem where lane detection in urban areas
often fails, since the curving of the lane runs out of
the scope from the camera. In addition most of the
algorithms are mostly designed for straight lanes and
not for sharp curves. Thus a reliable lane information,
one of the basics for autonomous driving, is not guar-
anteed. In this paper we use a new type of lane mark-
ing which was developed by 3M (3M, 2021). This
lane marking enhances the contrast (artificial light-
dark boundary) for camera systems and the reflecting
of light beam from a LiDAR with 3D arranged retro
reflective elements. The intensities of the point data
can directly be used as a feature for the segmenta-
tion. Complex filters are not required to extract the
information from the lane marking and misinterpre-
tations are minimized. With transferring the LiDAR
points into the 2D area, the lanes are then extracted
through dynamic horizontal and vertical sliding win-
dows, which finally leads to the relevant points. For a
better comparison we take the raw points into account
and will not apply any filters for a smooth represen-
tation. It will also provide a comparison between the
detection based on camera and LiDAR. All the mea-
surements are done on the test field of the University
of Applied Science (HTW) Dresden. Since only one
lane is equipped with this new lane marker, a compar-
ison to the conventional lane marking can be given.
2 FUNDAMENTALS
2.1 LiDAR
A LiDAR emits infrared coherent light from a laser to
its environment. The energy is decisive for the clas-
sification of the sensor into the protective classes and
results from the integral of the pulse over time. Due to
the optics of the LiDAR, the light beam diverges and
spreads flat, depending on the distance. This means
that less light power is radiated onto the object in a
ICINCO 2021 - 18th International Conference on Informatics in Control, Automation and Robotics
70