method does not rely on edge information in images,
and does not use image prior learned from the train-
ing dataset. Thus, it works properly even if there is
no strong edge information in reflection, and does not
suffer from the domain shift problem.
2 RELATED WORK
In recent years, various approaches have been pro-
posed for image signal separation. These methods fall
into two classes, i.e. single image based methods and
multiple image based methods.
The single image based methods are ill-posed, and
hence it is necessary to combine a priori knowledge in
these methods. Levin et al.(Levin et al., 2004) pro-
posed a technique for separating reflections so that
the brightness gradient of the recovered image ap-
proaches to that of the image in the natural image
database. More recently, deep learning techniques
are used for learning image features of reflection and
separating reflections from images (Fan et al., 2017;
Zhang et al., 2018; Wan et al., 2018). These meth-
ods enable us to separate reflections more accurately
and faster than before. However, since these methods
learn the image features of reflection based only on
the training data, an enormous amount of training data
is required to deal with various types of reflections in
various scenes. As a result, these methods often suffer
from the domain shift problem caused by the limited
number of data. On the contrary, our method is based
on the imaging model of reflection and does not rely
on training dataset, so it does not suffer from the do-
main shift problem.
The multiple image based methods use the prop-
erty that the motion of the background scene is dif-
ferent from that of the reflection on glass windows.
Yu and Brown (Yu and Brown, 2013) proposed a
method for separating images into foreground and
background by matching image features using SIFT.
Xue et al. (Xue et al., 2015) recovered dense motion
fields from sparse motion fields obtained by edge in-
formation, and showed that the dense motion fields
enable us to separate foreground and background im-
ages more accurately. Nandoriya et al. (Nandoriya
et al., 2017) also used image edge information for
obtaining initial motion field and for separating fore-
ground and background information in video frames.
However, all these methods require edge information
for obtaining motion fields in images, and hence they
cannot separate foreground and background informa-
tion accurately, when we do not have abrupt change
in intensity and cannot obtain image features in input
images. On the contrary, our method does not require
Figure 2: In-vehicle camera observes not only background
scene, but also reflection of objects in vehicle.
Figure 3: Observed image which contains reflection from
objects in vehicle, i.e. two gray squares.
abrupt change in images and does not need to extract
image features for separating image signals.
3 TRANSMISSION AND
REFLECTION
We first consider an imaging model of an in-vehicle
camera, in which reflected light and transmitted light
are simultaneously captured in a single image.
In the image of an in-vehicle camera, we often ob-
serve light reflected by the windshield of the vehicle
I
B
as well as light transmitted through the windshield
I
D
. Fig. 2 shows the scene where two types of light I
B
and I
D
enter the in-vehicle camera, and Fig. 3 shows
observed image I
R
by the camera. The image I
R
in-
cludes regions that are darker than the surrounding
area. This is the reflection caused by objects in the
vehicle.
Xue et al. formulated that the observed image is
expressed by the alpha blending of the background
image and the foreground image at a fixed ratio as
follows:
I
R
(x) = (1− α(x))I
B
(x) + α(x)I
D
(x) (1)
where x = [x, y]
T
is an image pixel, and α(x) denotes
the mixing ratio at x, which ranges from 0 to 1.
It seems that Eq. (1) is correct, but actually it is
physically wrong, since in this model the light I
B
from
the outside of the vehicle is attenuated with (1− α),
which does not happen in reality. Reflection is caused