ing chapter), because the former is contracting while
the latter is expanding.
Therefore, the method developed within this work
is also based on monocular pixel motion. The deci-
sion was against a stereo system to save one camera.
Additionally, the baseline of a stereo system would be
quite small because of the limited space on a motor-
cycle. This in turn restricts the sensing range signif-
icantly. Another advantage of using motion features
is the independence of an objects appearance, e.g. at
night the appearance changes significantly compared
to day-time as a vehicle can only be identified by its
front-lights.
To the knowledge of the authors, there exists no
such vision-based system for motorcycles yet. As the
main difference in vehicle dynamics between car and
motorcycle is the ability to ride in leaning position,
(Schlipsing et al., 2012) proposed a method to esti-
mate the roll angle of a motorcycle. The idea is to
transfer existing assistance systems from the car do-
main, like lane detection or obstacle detections which
requires such a roll-angle compensation. Obviously,
this represents a convenient solution to make use of
already available technologies. The disadvantage is
that all post-processing depends on the reliability of
the roll-angle compensation, which might not be de-
sirable in sense of error propagation and independent
running applications.
In the remainder of this paper, the approaching ve-
hicle indication is described at first in Section 2. The
implementation of the system on a motorcycle and ex-
perimental results under rainy, dark and high speed
conditions are discussed in Section 3. Finally, the dis-
cussion and conclusion section summarizes the out-
comes and explains remaining challenges.
2 APPROACHING VEHICLE
DETECTION
Mounting a camera to the rear of a vehicle causes a
contracting pixel motion in the image sequence if the
vehicle moves forward. This means that all pixels
move towards a focus of contraction if the scene is
static. The magnitude of each motion vector in the
image mainly depends on the corresponding depth of
a pixel in real world coordinates. If an object is mov-
ing in the scene, the measured pixel motion of the ob-
ject is a combination of the ego-motion and the object
motion. This results in zero motion if the object is
moving with the same speed in the same direction as
the ego-vehicle.
As soon as the object velocity is larger than the
ego-vehicle velocity, the pixel motion pattern of the
object becomes upscaling with a flow field mov-
ing away from a focus of expansion. This means,
during ego-vehicle movement, an approaching ob-
ject causes an expanding flow field while static back-
ground causes a contracting flow field. This makes
both patterns distinguishable by their scaling factor
(greater or lower than 1). In turn, if the object drives
with lower speed as the ego-vehicle, background and
object motion are both contracting, i.e. both scaling
factors are lower than 1.
If the ego-vehicle additionally undergoes a rota-
tional motion, the projected motion pattern on the im-
age plane is overlayed with a vertically translating
component in case of pitching or a rotational com-
ponent in case of rolling, whereas the magnitude of
a motion vector is independent of the corresponding
depth of a pixel. However, scaling factors are not in-
fluenced by rotational motion. The rolling component
is of special interest for this application, as such a mo-
tion occurs only for motorcycles and not for cars.
In the following, potentially approaching object
motion is detected by simply checking the scale fac-
tor in a local neighborhood. This is an efficient pre-
processing step to reduce the number of non-relevant
motion vectors for fitting a geometric motion model
for approaching vehicles in a second step. The main
advantage of this approach is that no ego-motioncom-
pensation needs to be done at all, which avoids the
influence of errors from an additional pre-processing
step.
2.1 Pre-selection of Motion Information
For motion estimation, a sparse pixel motion estima-
tion method has been applied. Sparse means that mo-
tion vectors
~
v
i
= (u
i
,v
i
)
T
with corresponding homo-
geneous pixel coordinates x
i
= (x
i
,y
i
,1)
T
are only
computed at well structured regions. Compared to
dense motion estimations, which compute motion
vectors for every pixel, such methods can save signif-
icant computational effort. The method used here is
the pyramid implementation of the Lucas and Kanade
optical flow estimation (Bouguet, 2000), available in
the OpenCV library (OpenCV, 2013). The big advan-
tage of the pyramid implementation compared to the
standard Lucas and Kanade approach is the ability to
cover large pixel displacements by propagating over
different image resolutions.
To decide whether a motion vector corresponds to
an approaching vehicle or to the background, at least
two neighboring motion vectors are required to com-
pute their scaling factor. As opposed to dense motion
vector fields, the neighborhood within a sparse mo-
tion vector field is not clearly defined.
MonocularRearApproachIndicatorforMotorcycles
475