methods by stereo camera measurement (D.
Stoyanov et al., 2005), laser scanning (M. Hayashibe
et al., 2006), shape from shading methods (T.
Okatani et al., 1997), illumination model (P.
Sánchez-González et al., 2008) and shape from
motion methods (K. Deguchi et al., 1996) (T.
Nagakura et al., 2007) have been presented for
clinical applications of endoscope. Stereo camera
measurement and laser scanning have drawback for
endoscopic operations in terms of the size of
endoscope because it needs two or more optical
system or additional laser devices. Shape from
shading makes use of the Lambertian reflectance
model that brightness is constant regardless of the
observed angle. But it is difficult to satisfy the
constraint in many cases of real organ surface. Shape
from motion solves for 3-D shape by using the
relative motion of objects from the camera. However,
for achieving 3-D measurement, it requires the
perturbational camera motion for image acquisition
of different positions.
We propose a configuration for 3-D shape
recovery of endoscopic images based on shape from
focus (SFF). Figure 1 is a schematic diagram of
concept of our method. In SFF method, shape is
obtained with the use of the image sequence
partially in-focus taken by changing the focused
position. The analysis of focus to estimate depth
from the camera to object has been used for the
automatically focusing camera system. The auto
focus method from focus information using the
Fourier transform is proposed by Horn(B. K. P.
Horn, 1968). Several methods concerning the
analysis of focus criterion are compared by
Krotkov(E. P. Krotkov, 1987). And the method of
recovering shape from focus has been presented by
Nayar(S. K. Nayar et al., 1994) and Subbarao(M.
Subbarao et al., 1994).
Although SFF is not available for many cases, it
is appropriate for endoscopic inspection and therapy
because of several reasons. Considering invasion
and pain of patients, endoscopes need to be narrow.
SFF is a method for extracting 3-D shape for single
camera and can make its hardware configuration
compact. Unlike the stereo method, SFF needs
neither matching problem nor occlusion, so the
application to a wide-angle lens suitable for an
endoscope is easy. Another account is that lighting
condition is easily controlled inside of organ rather
than outside world. It is an advantage to fulfill strict
conditions for SFF.
2 METHOD
2.1 Shape from Focus
SFF is a method of measuring shape by using focus
information. The focused position depends on the
camera parameters such as lens position or image
sensor.
The geometry of the defocusing and focusing can be
expressed by Figure 2. In this Figure, di is the
distance between the lens and sensor plane, df is the
distance between the lens and the focal plane when
the focus is perfectly focused. In case of df’>df or
df’<df, namely, if object is not placed on Focused
point, blur circle is formed. SFF method is an
application of this principle.
Figure 2: Formation of focused and defocused images.
The steps of SFF process are following. First, two or
more pictures are taken by changing the image
sensor position, the lens position or the object
position. In this paper, lens position is varied for
image sequence. Second, the focus measures of each
pixel in the each image are compared, and the
camera parameters are estimated by which the
photographed object is perfectly focused. Finally,
we can obtain the depth of each point from the
camera parameters.
2.2 Computation of Focus Measure
In SFF, it is important to evaluate the degree of
focus measure, and have proposed various methods.
In this paper, we use to evaluate focus measure
using High-Pass Filter (HPF) is introduced by
Krotkov (E. P. Krotkov, 1987) and Nayar (S. K.
Nayar et al., 1994). The blur image is represented
image intensity function i(x, y). i(x, y) is expressed
by the convolution of Point Spread Function(PSF)
h(x, y) and perfectly focused image i
f
(x, y).
),(),(),( yx
f
iyxhyxi ∗=
(1)
In equation (1), the symbol “
∗
” denotes convolution.
By considering the defocusing process in the
frequency domain, we obtain the following equation.
VISAPP 2009 - International Conference on Computer Vision Theory and Applications
412