5 PROPOSED APPROACH
Taking into account the needed precision, which im-
plies a high resolution camera, and the demanding im-
age processing due to the unstable conditions of the
image, a dynamic look-and-move approach has been
adopted. A particle filter has also been added to the
system to manage the uncertainties of the vision sys-
tem.
Next lines will describe the general structure of
the system, as well as the vision module, pose estima-
tion, the particle filter and the grasping algorithm.
5.1 System Modeling and Architecture
In the described scenario, the space can be repre-
sented by P ∈ ℜ
6
, a set of three positions and three
orientations, where P = [x,y, z, α,β, γ]
T
. In the same
way, this scenario will be composed of two different
frames, the robot frame r and the camera frame c.
Given those two frames, the homogeneous transfor-
mation matrix, denoted by
r
T
c
, transforms poses from
frame c to frame r as:
P
r
=
r
T
c
P
c
(5)
The error of the positioning task involved in the
grasping process is represented by vector E ∈ ℜ
6
which represents the difference between the pose of
the object P
r
o
in the robot frame and the pose of the
end-effector P
r
e
in the robot frame (6). The grasping
process can be seen as a minimization of this error
that will be fulfilled when |E| = 0.
E = P
r
e
− P
r
o
=
x
r
e
− x
r
o
y
r
e
− y
r
o
z
r
e
− z
r
o
α
r
e
− α
r
o
β
r
e
− β
r
o
γ
r
e
− γ
r
o
(6)
For pose estimation, position-based visual servo-
ing systems extract features from the acquired images
and estimate the pose of the object P
r
o
and perform
the corrections. Even so, the described scenario intro-
duces uncertainties in the feature extraction step (il-
lumination, metallic workpiece...), introducing errors
in the pose estimation. To deal with this problem, the
use of a particle filter is proposed. From each image, a
set of n feature vectors F
i
= { f
1
, f
2
,..., f
m
}
i=1...n
will
be extracted for the pose estimation, each of them re-
lated with a specific image analysis procedure. Each
of those n vectors will be a hypothesis of the values
of the m features used for the pose estimation, as it
will not be possible to have a unique feature vector
extracted from each image due to the uncertainties in
the image.
From each feature vector F
i
, P
c
o
i
and P
r
o
i
will be
calculated,
P
c
o
i
= PE(F
i
) (7)
P
r
o
i
=
r
T
c
P
c
o
i
(8)
where P
c
o
i
is the i-th hypothesis of the pose of the
object in the camera frame, P
r
o
i
is the i-th hypothesis
of the pose of the object in the robot frame and pose
estimation function PE is the function that relates a
set of features with a pose of the object in the camera
frame.
Figure 2: Dynamic position based look-and-move structure
with particle filter.
Those n poses, P
r
o
i=1..n
will be the observations of
the particle filter, which will output the final pose es-
timation of the object in the robot frame P
r
o
. This final
pose will be used to calculate the error E between the
object and the end-effector, used to calculate the next
robot movement. Fig. 2 shows the structure of the
proposed Visual Servoing system.
Next lines will describe the feature extraction,
pose estimation, particle filtering and grasping algo-
rithm of the grasping process.
5.2 Feature Extraction
As stated before, one of the challenges of the pre-
sented scenario is the feature extraction for pose es-
timation. The metallic nature of the grasping device
and the illumination problems make it difficult to de-
tect the different features (edges, corners, holes) pre-
cisely. Taking also into account the perspective of the
camera through the grasping process, the image fea-
tures used for pose estimation, shown in Fig. 3, are:
• The center of the three holes (1, 2, 3) of the grasp-
ing device. Only the pixels of the center of the
holes are included, excluding the size and dimen-
sions of the holes, due to the difficulties of extract-
ing their contour precisely.
• The inclination of the left edge (4) of the grasping
device.
To detect those image features different thresh-
olds, edge detection algorithms and filters are used.
Even so, in some images it is not possible to deter-
mine the exact position of the three holes’ centres as
ParticleFilteringforPositionbased6DOFVisualServoinginIndustrialEnvironments
163