object shape from multiple images taken at different
light source positions(HIGO, 2010). These methods
can simultaneously estimate object shape and reflec-
tion characteristics under conditions where the light
source is known. Similar to light source informa-
tion estimation, a neural network-based method has
also been proposed to simultaneously estimate ob-
ject shape, reflection characteristics, and light dis-
tribution from a single image taken of an indoor
scene(Sengupta et al., 2019). However, this method
also requires a large amount of training data to
achieve appropriate learning.
Therefore, in this paper, we consider the use of
neural networks to represent object shape and light
distribution without using training data and to simul-
taneously estimate object shape, reflection charac-
teristics, and light distribution from time-series im-
ages. This method aims to realize inverse rendering
with properties that are intermediate between physics-
based vision and learning-based vision.
3 OBSERVATION MODEL
3.1 General Rendering Model
First, we discuss the general rendering model. Ren-
dering is the computation of light reflections in a
scene based on specular reflections on object surfaces,
shadows, inter-reflections between objects, etc. The
rendering equation proposed by Kajiya(Kajiya, 1986)
is well-known as a basic mathematical model. There-
fore, rendering in computer graphics corresponds to
solving this rendering equation.
Assuming that no light penetrates into the interior
of the object, the reflectance at the observation point
x as f (x,
⃗
ω,
⃗
ω
′
). This reflectance indicates the ratio
of rays incident from the
⃗
ω
′
direction reflected in the
⃗
ω direction. The angle between the incident direc-
tion
⃗
ω
′
) and the plane normal direction ⃗n is shown by
θ. The set of ray directions incident on point x is de-
noted by Ω. In this case, the observed intensity L
o
is
expressed by the rendering equation as follows:
L
o
(x,
⃗
ω) = L
e
(x,
⃗
ω)+
Z
Ω
f
r
(x,
⃗
ω
′
,
⃗
ω)L
i
(x,
⃗
ω
′
)cosθd
⃗
ω (1)
where L
e
is the amount of light emitted from the point
⃗x. This equation indicates that the observed intensity
is determined by the integral of the reflected light in-
cident from all directions and the light L
e
(x,
⃗
ω) emit-
ted from the object in the
⃗
ω direction. Therefore, the
observed intensity information includes not only the
reflectance characteristics at the point, but also infor-
mation about the surrounding lighting environment.
Therefore, if we can analyze this information appro-
priately, we can reconstruct various types of informa-
tion from the intensity information.
3.2 Intensity Model
Next, we describe the image observation model used
in this study. In equation (1), L
o
is the luminance
emitted from the observation point x on the object sur-
face in the scene in the direction ω, and L
e
is the radi-
ance of light emitted from the object interior. Omega
is the direction of incidence of the light at the obser-
vation point x, which coincides with the hemisphere.
Here, L
e
can be assumed to be zero regardless of the
direction because the light is rarely emitted from the
interior of a typical object. The cosθ, is the just in-
ner product of the normal vector ⃗n and the direction
of incidence
⃗
ω
′
. Thus, the following equation, which
focuses only on the reflection component, is used as
the observation model of the image in this study.
L
r
(x,
⃗
ω) =
Z
Ω
f
r
(x,
⃗
ω
′
,
⃗
ω)L
i
(x,
⃗
ω
′
)(⃗n ·
⃗
ω
′
)d
⃗
ω (2)
When using such an observation model, light distribu-
tion and object shape information for the entire scene,
as well as reflectance property information, are re-
quired to render the image. Conversely, it is possi-
ble to estimate these information from the observed
intensity information.
3.3 Light Distribution Representation
Next, we describe the representation model of light
distribution used in this study. As described in the
equation (2), the light distribution around the object
is necessary to determine the intensity observed on
the object’s surface. However, it is difficult to directly
express this as a continuous quantity. Therefore, the
In this study, we use a geodesic dome as shown in
Fig 1. A geodesic dome is an approximate model of
a sphere composed of triangular patches, which can
discretely represent the spread of points on a sphere.
The vertices are distributed with equal density on the
sphere, so it is possible to represent light distribution
equally at isosceles angles. In this study, we assume
that a point light source exists at the center of gravity
of these triangular patches, and represents light distri-
bution by changing the intensity of each light source.
In this study, the temporal variation of light distribu-
tion is represented by changing the intensity of each
light source on the sphere according to the time.
VISAPP 2023 - 18th International Conference on Computer Vision Theory and Applications
468