An important part in the obstacle detection
process is the separation of the obstacle points from
the road points. Most of the roadway obstacle
detection methods are based on the flat road
assumption (Weber, 1995), (Williamson, 1998). This
is a poor model since deviations from the flat road
may be as large as or larger than the obstacles we
wish to detect. In consequence the road objects
separation and the 3D objects position estimation
cannot be done. Therefore the non-flat road
assumption is compulsory for a robust object
detection method. In literature this assumption was
introduced by non-flat road approximation by series
of planar surface sections (Hancock, 1997),
(Labayrade, 2002) or by modeling of the non-flat
roads by higher order surfaces (Goldbeck, 1999),
(Aufrere, 2001). For instance the methods presented
in (Aufrere, 2001), (Aufrere, 2000), (Takahashi,
1996) are fitting the parameters of a 3D clothoid
model of the road lane using a monocular image and
supplementary lane geometry constraints.
Our approach presented in this paper will model
the vertical profile of the road surface with such a
clothoid curve fitted directly on the detected 3D road
surface points. These 3D road points are detected
using a high accuracy stereovision method
(Nedevschi, 2004). The obtained vertical profile will
be used for the road-obstacle separation process in
order to have a proper grouping of the 3D points in
obstacles and precise estimation of their 3D position
in the driving environment.
2 ENVIRONMENT MODEL
All 3D entities (points, objects) are expressed in the
world coordinates system, which is depicted in
figure 1.a. This coordinates system, has its origin on
the ground in front of the car, the X axis is always
perpendicular on the driving heading direction, the
Y axes is perpendicular on the road surface and the
Z axis coincides with the driving heading direction.
The ego-car coordinates system has its origin in the
middle of the car front axis, and the tree coordinates
are parallel with the tree main axes of the car. The
world coordinates system is moving along with the
car and thus only a longitudinal and a vertical offset
between the origins of the two coordinates system
exists (vector T
EW
from Figure 1). The relative
orientation of the two coordinates systems (R
EW
rotation matrix) will change due to static and
dynamic factors. The loading of the car is a static
factor. Acceleration, deceleration and steering are
dynamic factors, which also cause the car to change
pitch and roll angles with respect to the road surface.
To obtain the pitch and roll angles and the car height
we measure the distance between the car’s chassis
and wheels because the wheels are on the road
surface. Four sensors are mounted between the
chassis and wheels arms and the car height (T
X
) and
the pitch(R
X
) and roll (R
Z
) angles are computed.
Figure 1.a shows also the position of the left and
the right cameras in the ego-car coordinate system.
The position is completely determined by the
translation vectors T
CE
i
and the rotation matrices
R
CE
i
. These parameters are essential for the stereo
reconstruction process and for the epipolar line
computation procedure. In order to estimate them an
offline camera calibration procedure is performed
after the cameras are mounted and fixed on the car
using a general-purpose calibration technique. Due
to the rigid mounting of the stereo system inside the
car these parameters are considered to be
unchangeable during driving.
The stereo reconstruction is performed in the car
coordinates system. The coordinates XX
E
=[X
E
, Y
E
,
Z
E
]
T
of the reconstructed 3D points in the ego-car
coordinates system can be expressed in the world
coordinate system as XX
W
=[X
W
, Y
W
, Z
W
]
T
using
the following updating equation:
)(
EWEW
TXXRXX +⋅=
EW
(1)
where T
EW
and R
EW
are the instantaneous relative
position and orientations of the two coordinates
system and are computed from the damper height
sensors by adding an offset to the initial value
(established during camera calibration). The
transformation between the rotation vector and its
corresponding rotation matrix is given by the
Rodrigues (Trucco, 1998) formulas.
)(
],0,[
],,0[
00
00
EWEW
ZXEWEWEWEW
YEWEWEWEW
Rodrigues
const
rR
RRrrrr
TTTTT
=
+=+=
+=+=
δδδ
δδ
(2)
The objects are represented as cuboids, having a
position (in the world coordinate system), size,
orientation and velocity, as in figure 1.b. The
position (X, Y, Z) and velocity (v
X
and v
Z
) are
expressed for the central lower point C of the object.
ICINCO 2004 - ROBOTICS AND AUTOMATION
12