at finding images relevant to a query and thus to the
information need expressed by the query. The rela-
tionship between any image in the database and a par-
ticular query can be expressed by a relevance value.
This relevance value relies on the user-perceived sat-
isfaction of his/her information need. The relevance
value can be interpreted as a mathematical probabil-
ity (a relevance probability). The notion of relevance
probability is not unique because different interpre-
tations have been given by different authors. In this
paper a relevance probability π(I) is a quantity which
reflects the estimation of the relevance of the image I
with respect to the user’s information needs. Initially,
every image in the database is equally likely, but as
more information on the user’s preferences becomes
available, the probability measure concentrates on a
subset of the database. The iterative relevance feed-
back scheme proposed in the present paper is based
on logistic regression analysis for ranking a set of im-
ages in decreasing order of their evaluated relevance
probabilities.
Logistic regression is based on the construction
of a linear model whose inputs, in our case, will be
the image characteristics extracted from a certain im-
age I and whose output is a function of the relevance
probability of the image in the query π(I). In logis-
tic regression analysis, one of the key features to be
established is the order of the model to be adjusted.
The order of the model must be in accordance with
the reasonable amount of feedback images requested
from the user. For example, it is not reasonable for
the user to select 40 images in each iteration; a feed-
back of 5/10 images would be acceptable. This re-
quirement leads us to group the image features into n
smaller subsets. The outcome of this strategy is that
n smaller regression models must be adjusted: each
sub-model will produce a different relevance proba-
bility π
k
(I) (k = 1 . . . n). We then face to the ques-
tion of how to combine the π
k
(I) in order to rank the
database according to the user’s preferences. OWA
(ordered weighted averaging) operators which were
introduced by Yager in 1988 (Yager, 1988) provides
a consistent and versatile way of aggregating multiple
inputs into one single output.
Section 2 explains the logistic regression approach
to the problem. Next, in section 3 the aggregation op-
erators used in our work are introduced. Section 4
describes the low level features extracted from the im-
ages and used to retrieve them. An crucial part of this
work, the proposed algorithm, is described in detail in
section 5. After that, in section 6 we present experi-
mental results which evaluate the performance of our
technique using real-world data. Finally, in section 7
we extract conclusions and point to further work.
2 LOGISTIC REGRESSION
MODEL
At each iteration, a sample is evaluated by the user
selecting two sets of images: the examples or posi-
tive images and the counter-examples or negative im-
ages. Let us consider the (random) variable Y giving
the user evaluation where Y = 1 means that the image
is positively evaluated and Y = 0 means a negative
evaluation.
Each image in the database has been previously
described by using low level features in such a way
that the j-th image has the k-dimensional feature vec-
tor x
j
associated. Our data will consist of (x
j
, y
j
),
with j = 1, . . . , k where x
j
is the feature vector and y
j
the user evaluation (1= positive and 0= negative). The
image feature vector x is known for any image and
we intend to predict the associated value of Y . The
natural framework for this problem is the generalized
linear model. In this paper, we have used a logistic
regression where P(Y = 1 | x) i.e. the probability that
Y = 1 (the user evaluates the image positively) given
the feature vector x, is related with the systematic part
of the model (a linear combination of the feature vec-
tor) by means of the logit function. Generalized lin-
ear models (GLMs) extend ordinary regression mod-
els to encompass non-normal response distributions
and modeling functions of the mean. Most statisti-
cal software has the facility to fit GLMs. Logistic
regression is the most important model for categor-
ical response data. Logistic regression models are
also called logit models. They have been successfully
used in many different areas including business appli-
cations and genetics. For a binary response variable
Y and p explanatory variables X
1
, . . . , X
p
, the model
for π(x) = P(Y = 1 | x) at values x = (x
1
, . . . , x
p
) of
predictors is
logit[π(x)] = α + β
1
x
1
+ . . . + β
p
x
p
(1)
where logit[π(x)] = ln
π(x)
1−π(x)
. The model can also be
stated directly specifying π(x) as
π(x) =
exp(α + β
1
x
1
+ . . . + β
p
x
p
)
1 + exp(α + β
1
x
1
+ . . . + β
p
x
p
)
. (2)
The parameter β
i
refers to the effect of x
i
on the log
odds that Y = 1, controlling the other x
j
. The model
parameters are obtained by maximizing the likelihood
equations.
In the first steps of the procedure, we have a major
difficulty when having to adjust a global regression
model in which we take the whole set of variables into
account, because the number of images (the number
of positive plus negative images chosen by the user)
VISAPP 2007 - International Conference on Computer Vision Theory and Applications
168