
 
An important part in the obstacle detection 
process is the separation of the obstacle points from 
the road points. Most of the roadway obstacle 
detection methods are based on the flat road 
assumption (Weber, 1995), (Williamson, 1998). This 
is a poor model since deviations from the flat road 
may be as large as or larger than the obstacles we 
wish to detect. In consequence the road objects 
separation and the 3D objects position estimation 
cannot be done. Therefore the non-flat road 
assumption is compulsory for a robust object 
detection method. In literature this assumption was 
introduced by non-flat road approximation by series 
of planar surface sections (Hancock, 1997), 
(Labayrade, 2002) or by modeling of the non-flat 
roads by higher order surfaces (Goldbeck, 1999), 
(Aufrere, 2001). For instance the methods presented 
in (Aufrere, 2001), (Aufrere, 2000), (Takahashi, 
1996) are fitting the parameters of a 3D clothoid 
model of the road lane using a monocular image and 
supplementary lane geometry constraints.  
Our approach presented in this paper will model 
the vertical profile of the road surface with such a 
clothoid curve fitted directly on the detected 3D road 
surface points. These 3D road points are detected 
using a high accuracy stereovision method 
(Nedevschi, 2004). The obtained vertical profile will 
be used for the road-obstacle separation process in 
order to have a proper grouping of the 3D points in 
obstacles and precise estimation of their 3D position 
in the driving environment. 
2 ENVIRONMENT MODEL 
All 3D entities (points, objects) are expressed in the 
world coordinates system, which is depicted in 
figure 1.a. This coordinates system, has its origin on 
the ground in front of the car, the X axis is always 
perpendicular on the driving heading direction, the 
Y axes is perpendicular on the road surface and the 
Z axis coincides with the driving heading direction. 
The ego-car coordinates system has its origin in the 
middle of the car front axis, and the tree coordinates 
are parallel with the tree main axes of the car. The 
world coordinates system is moving along with the 
car and thus only a longitudinal and a vertical offset 
between the origins of the two coordinates system 
exists (vector T
EW
 from Figure 1). The relative 
orientation of the two coordinates systems (R
EW
 
rotation matrix) will change due to static and 
dynamic factors. The loading of the car is a static 
factor. Acceleration, deceleration and steering are 
dynamic factors, which also cause the car to change 
pitch and roll angles with respect to the road surface. 
To obtain the pitch and roll angles and the car height 
we measure the distance between the car’s chassis 
and wheels because the wheels are on the road 
surface. Four sensors are mounted between the 
chassis and wheels arms and the car height (T
X
) and 
the pitch(R
X
) and roll (R
Z
) angles are computed. 
Figure 1.a shows also the position of the left and 
the right cameras in the ego-car coordinate system. 
The position is completely determined by the 
translation vectors T
CE
i
 and the rotation matrices 
R
CE
i
. These parameters are essential for the stereo 
reconstruction process and for the epipolar line 
computation procedure. In order to estimate them an 
offline camera calibration procedure is performed 
after the cameras are mounted and fixed on the car 
using a general-purpose calibration technique. Due 
to the rigid mounting of the stereo system inside the 
car these parameters are considered to be 
unchangeable during driving. 
The stereo reconstruction is performed in the car 
coordinates system. The coordinates XX
E
 =[X
E
, Y
E
, 
Z
E
]
T
 of the reconstructed 3D points in the ego-car 
coordinates system can be expressed in the world 
coordinate system as XX
W
 =[X
W
, Y
W
, Z
W
]
T
 using 
the following updating equation: 
 
  
)(
EWEW
TXXRXX +⋅=
EW
     (1) 
 
where  T
EW
 and R
EW
 are the instantaneous relative 
position and orientations of the two coordinates 
system and are computed from the damper height 
sensors by adding an offset to the initial value 
(established during camera calibration). The 
transformation between the rotation vector and its 
corresponding rotation matrix is given by the 
Rodrigues (Trucco, 1998) formulas. 
 
)(
],0,[
],,0[
00
00
EWEW
ZXEWEWEWEW
YEWEWEWEW
Rodrigues
const
rR
RRrrrr
TTTTT
=
+=+=
+=+=
δδδ
δδ
   (2) 
 
The objects are represented as cuboids, having a 
position (in the world coordinate system), size, 
orientation and velocity, as in figure 1.b.  The 
position (X, Y, Z) and velocity (v
X
 and v
Z
) are 
expressed for the central lower point C of the object. 
  
 
 
 
ICINCO 2004 - ROBOTICS AND AUTOMATION
12