same vertical-axes coordinates of vertex which is y
v
=
y
FOE
/2 but they have different horizontal-axes coor-
dinates v = K
r
|y
r
|(y
r
−y
FOE
| (see Fig.5(a)). In case
of moving vehicles, the most important translation
in terms of amplitude is toward the Z direction, that
is T
X
≈ 0 and T
Y
≈ 0, it means that x
FOE
≈ 0 and
y
FOE
≈0. So all parabolas share the same vertex point
which is the origin (see Fig.5(b)). Consequently, all
parabolas share the same form v = ay
2
with a an un-
known value. In order to detect parabolas, we propose
a consensus voting process that leads to the estimation
of parameter a.
Finally, the straight lines corresponding to the lat-
eral planes are detected using the Hough transform
after removing pixels that are already labeled as be-
longing to an horizontal or a vertical plane. For point
of extension FOE, knowing that all translation mo-
tions will converge to or diverse from that point. A
voting space where each optical flow draws a straight
line on image are created where the point which has
the most passages is the point of extension. Normally,
this point does not deviate much from the center of
image under our assumption.
4 EXPERIMENTS
To prove the validity of our approach, experiments
are made using first the optical flow ground truth pro-
vided by KITTI and then with the optical flow estima-
tion algorithm proposed by (Sun et al., 2010). For this
first study, only the sequences where translational mo-
tion is dominant are considered. Figures 6 to 8 show a
few examples, where the results of c-velocity and uv-
velocity are put side by side for each kind of plane.
All voting spaces are created using the absolute value
of optical flow for uv-velocity. When using the opti-
cal flow ground truth (top), we got expected results:
planes –especially the horizontal ones (see Fig. 6)–
are correctly detected. Since the optical flow ground
truth is not dense, we focus only on the vehicle and
the road planes (Fig.6,8) since their attributes appear
clearly on the voting space (Fig.4).
By using the optical flow computed from (Sun
et al., 2010) (bottom of figures), the results are not
as good as those we got with ground truth in terms
of precision, occlusion handling (see Fig.6), but the
voting spaces still reveal enough the expected curves
like the one we see in Fig.3. When using the ground
truth, the horizontal plane gives the most reliable re-
sults since it is always available on the image (it cor-
responds to the road). The detection of lateral planes
depends on the scene context, whether it has enough
points to vote for a parabola. Using the Hough trans-
form, the obstacle plane seems to be unstable, since,
for instance, a car always contains many planes. It
means that we have to consider voting space coopera-
tion as future work. However, for some scenes, when
the line appears clearly like in Fig.8, the obstacle can
be detected correctly.
As we can see on Fig.6,7, the uv-velocity give al-
most the same performance as c-velocity whatever the
optical flow, but it avoids expensive calculations like
square-root or exponential and intermediate value c.
5 CONCLUSION
This paper has shown how to detect 3D planes in a
Manhattan world using a specific voting space called
uv-velocity. Its construction is based on the exploita-
tion of optical flow intrinsic properties of moving
planes and more particularly on iso-velocity curves.
Results on ground-truth optical flows prove the effi-
ciency of our new concept, when planes have enough
pixels on the image to be detected. Experiments show
that the precision of the results depends on the the
quality of the input optical flow. In theory, the inter-
ference of other plane models on voting spaces will
not cause much side effects on curve detection be-
cause there contribution in the voting space is low and
could be eliminated by a simple threshold. In practice,
we show that these interferences can complicate line
and parabola detection. One of our futures works is
then to propose a cooperation strategy between voting
spaces. Moreover, since the quality of optical flow is
directly related to the spread of the line or the parabo-
las in the voting space, it is possible to find a met-
ric to find the parabola and refine the optical flow at
the same time (Mai et al., 2017). Finally, rotational
motion will be investigated in next steps to make the
results more general.
REFERENCES
Azuma, T., Sugimoto, S., and Okutomi, M. (2010). Ego-
motion estimation using planar and non-planar con-
straints. In 2010 IEEE Intelligent Vehicles Sympo-
sium, pages 855–862.
Bouchafa, S. and Zavidovique, B. (2012). c-Velocity:
A Flow-Cumulating Uncalibrated Approach for 3d
Plane Detection. International Journal of Computer
Vision, 97(2):148–166.
Cai, Z., Fan, Q., Feris, R. S., and Vasconcelos, N. (2016). A
Unified Multi-scale Deep Convolutional Neural Net-
work for Fast Object Detection. In Computer Vision –
ECCV 2016, pages 354–370. Springer, Cham.
Exploiting Optical Flow Field Properties for 3D Structure Identification
463