visited in the past p
i
= (p
i1
, p
i2
, . . . , p
iD
) (particle
best) and a velocity v
i
= (v
i1
, v
i2
, . . . , v
iD
). Usually
the position is initialized uniformly distributed over S
and the velocity is also uniformly distributed depend-
ing on the size of S. The movement of each particle
takes place in discrete steps using an update function.
In order to calculate the update of a particle we need
a supplementary vector g = (g
1
, g
2
, . . . , g
D
) (global
best), the best position of a particle in its neighbor-
hood. The update function, called inertia weight, con-
sists of two parts. The new velocity of a particle P
i
is
calculated for each dimension d = 1, 2, . . . , D:
v
new
id
= w ·v
id
+ c
1
ε
1d
(p
id
−x
id
) + c
2
ε
2d
(g
d
−x
id
)
(1)
then the position is updated: x
new
id
= x
id
+ v
new
id
. The
new velocity depends on the global best (g
d
), particle
best (p
id
) and the old velocity (v
id
) which is weighted
by the inertia weight w. The parameters c
1
and c
2
pro-
vide the possibility to determine how strong a particle
is attracted by the global and the particle best. The
random vectors ε
1
and ε
2
are uniformly distributed
over [0, 1)
D
and produce the random movements of
the swarm.
2.2 Algorithm Configuration Problem
The general problem of configuring algorithms (algo-
rithm configuration problem) is defined by Hutter et
al. (Hutter et al., 2007) as finding the best tuple θ
out of all possible configurations Θ (θ ∈ Θ). θ rep-
resents a tuple with a concrete assignment of values
for the parameter of an algorithm. Applied to meta-
heurisitcs the configuration of the algorithm parame-
ters for a specific problem influences the behavior of
the optimization process. Different parameter settings
exhibit different performances at solving a problem.
The problem to configure metaheuristics is a super
ordinate problem and is analyzed for different kinds
of metaheuristics. In PSO the convergence of the
optimization depending on different parameter set-
tings and different functions are analyzed by (Trelea,
2003), (Shi and Eberhart, 1998) and (van den Bergh
and Engelbrecht, 2002). But these approaches focus
only on the convergence of the PSO but not on func-
tion characteristics and the relationship between the
parameter configuration and the function landscape.
Different approaches to solve this algorithm con-
figuration problem on metaheurisitcs are introduced:
One approach is to find sets of adequate parameters
which performs a good optimization on most different
types of objective functions. This “standard param-
eters” are evaluated on a preset of functions to find
a parameter set which leads to global good behavior
of the metaheuristic. In PSO standard parameter sets
are presented by (Clerc and Kennedy, 2002) and (Shi
and Eberhart, 1998). Some approaches do not present
a preset of parameters but change the values of the
parameters during the runtime to get a better perfor-
mance (Pant et al., 2007).
Another approach is introduced by Leyton-Brown
et al. They try to create features which describe the
underlying problem (Leyton-Brown et al., 2002) and
generate a model predicting the right parameters de-
pending on the classification. They introduce several
features which are grouped into nine groups. The fea-
tures include, among others, problem size statistics,
e.g. number of clauses and variables, and measures
based on different graphical representations. This
analysis is based on discrete search spaces because
on continuous search spaces it is not possible to set
adequate discrete values for the parameter configura-
tion which is needed by their appraoch.
Our problem is to configure an algorithm working
on continuous search spaces and offers infinite pos-
sibilities of parameter sets. To solve this challenge
we try, similar to Leyton-Brown et al., to train a clas-
sifier with features of the fitness function landscape
computed by observing swarm behavior. These fea-
tures are computed and combined with the best found
parameter set on the function to a training instance
(see figure 1). With a trained classifier at hands we
compute the features of the objective function prior
to the start of the optimization process. The classifier
– in our case a decision tree – classifies the function
and selects the specific parameter set that is expected
to perform better in the optimization process than us-
ing the standard parameters. In our first experiments,
which we understand as proof of concept, we choose
only a few functions which do not represent any spe-
cific types of function. We want to show that our
technique is able to identify functions based on the
swarm behavior provided features and thereby, select
the specific parameter configuration. In order to learn
the classifier which suggests the parameter configura-
tion, different function features are computed. These
features are the basis of our training instances.
3 COMPUTATION OF FUNCTION
FEATURES
Our computed features can be divided into three
groups. Each group implies a distinct way of collect-
ing information about the fitness topology of the ob-
jective function from particles. The first group Ran-
dom Probing describes features which are calculated
based on a random selection of fitness values and
provides a general overview of the fitness topology.
EXTRACTION OF FUNCTION FEATURES FOR AN AUTOMATIC CONFIGURATION OF PARTICLE SWARM
OPTIMIZATION
53