In our approach, users can work with color, tex-
ture and shape hierarchically to refine retrieval. These
three families of features are not mixed together be-
cause they are independant. For example if a user
wants to find ”red cars” in a collection, color and
shape have to be used. Texture will not be useful in
this case. When you work with only one feature vec-
tor where the three features are mixed, useless fea-
tures influence the final decision while they are not
supposed to.
More and more methods are based on offline clas-
sification of feature vectors to build a visual search
tree to browse the collection online. In our system,
a query-by-visual-example method (Boujemaa et al.,
2003) is used because time computing limitation is
not really important in our retrieval process due to the
high speed of binary computation.
3 PROPOSED ARCHITECTURE
Our system is based on binarization of classical fea-
tures. There are two steps in the proposed system:
offline and online. Let’s consider an image collection
C containing N images noted I
i
where i = 1..N.
In the offline step (no user connected to the CBIR
system), each image I
i
of the collection C is trans-
formed from RGB to Lab colorspace. Lab colorspace
was chosen because distances computed in this space
correspond to real perception of distances between
colors. Then a multiresolution analysis (Calderbank
et al., 1998) is computed at three resolution levels.
Several classical features are extracted in color, tex-
ture and shape feature vectors. The binarization pro-
cess is described further and leads to three binary sig-
nature per image: s
C
i
, s
T
i
and s
S
i
.
The size of our signatures is 32-bits so that XOR
operations can be processed into the microprocessor
internal registers. Each bit in s
C
i
, s
T
i
and s
S
i
represents
a property which is true (1) or false (0). Thus each
signature is a set of binary properties for the image I
i
.
Figure 1 presents our query-by-example architec-
ture. The binary extracted signature of the request im-
age I
R
is compared to every image I
i
of the collection
C and results are displayed on the user screen, sorted
by increasing distance.
Features are organized into a 32-bits binary sig-
nature vector. For an image I
i
, there are three binary
signature vectors corresponding to color (s
C
i
), texture
(s
T
i
) and shape (s
S
i
). Bits in signatures represent the
fact that the considered image satisfies a certain prop-
erty or not.
• Color: Color properties are based on ”a” and ”b”
maps values of ”Lab” colorspace. There are 32
Figure 1: Architecture of the proposed system.
properties tested in every 32-bits color binary sig-
natures. For instance, the first bit is to check prop-
erty: — Does the mean value of ”a” colormap at
the coarser resolution is greater than 64 ? —. A
value of 1 indicates this property is satisfied for
this image, a value of 0 means it is not satisfied.
So by associating several properties, our signature
contains a checklist of color properties.
• Texture: Binary properties for texture are mainly
based on the study of wavelets energy (square
value of each coefficient) through the three differ-
ent levels of resolution. For instance, the first bit
is to check property: — Does the mean energy of
”L” colormap for the coarser resolution is greater
than 128 ? —.
• Shape: Shape properties are extracted from image
contours of the ”L” colormap (by a laplacian edge
detector). For example, a typical property is: — Is
there any continuous contour of the object longer
than 30 pixels ? —.
So the entire process of binarization consists in
transforming real world questions into binary an-
swers. The underlying problem is the choice of prop-
erties.
Of course the list of properties is not exhaustive
and any kind of question whose answer is yes (1) or
not (0) is a potential binary property to use in our sys-
tem. Once binary properties have been chosen, a sim-
ilarity (or dissimilarity) metric must be used to com-
pute distances between images, i.e. between signature
vectors.
4 SIMILARITY COMPUTING
In order to evaluate distances between request image
I
R
and collection images I
i
, a metric must be defined.
We need a measurement method to tell how two bi-
nary signatures s
R
(request) and s
i
(i
th
image in the
collection) are similar (bit per bit). Therefore we want
a similarity measure where the distance value will be
the number of similar bits in the considered signa-
VISAPP 2007 - International Conference on Computer Vision Theory and Applications
238