(Chum et al., 2005; Yang et al., 2005) use plane-to-
plane homography for detecting a plane. Motion seg-
mentation with layers is proposed in refs. (Wang and
Adelson, 1994; Weiss, 1997). Brox et al. proposed
an algorithm for image segmentation by the level set
method (Brox et al., 2006). We use the dominant-
plane model for segmenting multiple planar areas.
Since the dominant plane is a planar area in the robot
workspace and it corresponds to the largest part of an
image, our algorithm does not require any restrictions
on the camera motion or geometric configurations be-
tween the camera and objects. The hierarchical detec-
tion of dominant planes allows the robot to achieve
spatial reasoning without any three-dimensional re-
construction of the scene, since the dominant plane
is a binary feature on the image plane. Furthermore,
since the dominant plane is a binary feature, the algo-
rithm is robust against the outliers that are derived in
the process of optical-flow computation.
2 DOMINANT PLANE AND
OPTICAL FLOW
2.1 Dominant Plane
We define the dominant plane in an image.
Definition 1 . The dominant plane as the planar area
in the robot workspace corresponding to the largest
part of an image or at least the half of an image.
Similarly with the previous paper (N.Ohnishi
and Imiya, 2005), we accept the following five
assumptions.
Assumptions
1. The ground plane is the planar area.
2. The camera mounted on the mobile robot is look-
ing downward.
3. The robot observes the world using the camera
mounted on itself for navigation.
4. The camera on the robot captures a sequence of
images since the robot is moving.
5. Obstacles occupy at most 1/2 region in an image
captured by the robot.
Therefore, if there are no obstacles around the robot,
and since the robot does not touch the obstacles, the
ground plane corresponds to the dominant plane in the
image observed through the camera mounted on the
mobile robot. Assuming that the dominant plane in
the image corresponds to the ground plane on which
the robot moves, the detection of the dominant plane
enables the robot to detect the feasible region for nav-
igation in its workspace.
2.2 Optical Flow on the Dominant Plane
Assuming that the camera displacement is small, on
each layer the corresponding points x = (x,y)
⊤
and
x
′
= (x
′
,y
′
)
⊤
in the dominant planes between a pair
of successive two images are connected with an affine
transform such that x
′
= Ax + b, where A and b are
a 2× 2 affine-coefficient matrix and a 2-dimensional
vector.
We can estimate the affine coefficients using
the RANSAC-based algorithm (Fischler and Bolles,
1981). Using estimated affine coefficients, we can es-
timate optical flow on the dominant plane ˆx = ( ˆx, ˆy)
⊤
,
ˆx = Ax +b −x, for all points x in the image. We call
ˆx the planar flow on the lth layer , and ˆx = (x, y,t)
the planar flow field at time t, which is a set of planar
flow ˆx computed for all pixels in an image.
If an obstacle exists in front of the robot, the pla-
nar flow on the image plane differs from the optical
flow on the image plane. Since the planar flow vector
ˆx is equal to the optical flow vector ˙x on the dominant
plane, we use the difference between these two flows
to detect the dominant plane. We set ε to be the toler-
ance of the difference between the optical flow vector
and the planar flow vector. Therefore, for the optical
flow equation ∇I
⊤
˙x + ∂
t
I = 0 of an image I observed
at time t if the inequality
| ˙x − ˆx| < ε,
s.t. ˆx = (Ax + b)− x, ∇I
⊤
˙x + ∂
t
I = 0 (1)
is satisfied, we accept point x as a point on the domi-
nant plane (N.Ohnishi and Imiya, 2005).
Our algorithm is summarised as follows:
1. Compute optical flow field u(x,y,t) from two suc-
cessive images.
2. Compute affine coefficients of the transform Ax +
bby random selection of three points.
3. Estimate planar flow field ˆu(x,y,t) from affine co-
efficients.
4. Match the computed optical flow field u(x,y,t)
and estimated planar flow field ˆu(x
l
,y
l
,t) using
eq. (1).
5. Assign the points | ˙x − ˆx| < ε as the dominant
plane. If the dominant plane occupies less than
half the image, then return to step 2.
6. Output the dominant plane d(x,y,t) as a binary
image.
VISAPP 2010 - International Conference on Computer Vision Theory and Applications
436