If
t
tt
TxyP ,,≤
κ
, set
tt
yx =
+1
and
tt
yfxf =
+1
; otherwise, set
tt
xx =
+1
and
tt
xfxf =
+1
.
Step 4.
If the prescribed termination condition is
satisfied, then stop; otherwise, update the value of
the temperature by means of the temperature
updating function, and then go back to Step 2.
Thus, by applying the generation mechanism and
the Metropolis acceptance criterion, the SA
algorithm produces two sequences of random points.
These are the sequence
0, ≥ty
t
of trial points
generated by (4) and the sequence
0, ≥tx
t
of
iteration points determined by applying the
Metropolis acceptance criterion as described in Step
3. These two sequences of random variables are all
dependent on the temperature sequence
}
0, ≥tT
t
determined by the temperature updating function,
the state neighbouring sequence
{}
0, ≥t
t
, and the
approach of random vector generation.
The sequence
{}
0, ≥t
t
of positive numbers
specified in Step 1 of the above SA algorithm is
used to impose a lower bound on the random vector,
generated at the each iteration, for obtaining the
random trial point. This lower bound should be
small enough and monotonically decreasing as the
annealing proceeds. Since the temperature-
dependent generation probability density function is
used to generate random trial points and since only
one trial point is generated at each temperature value
the SA algorithm considered is characterized by a
nonhomogeneous continuous-state Markov chain.
The convergence conditions of the SA were
studied by Yang (Yang, 2000) and several updating
functions for the method parameters were given,
which ensure convergence of the method. We
applied the next updating functions in testing our
approach.
Let
n
r ℜ∈
, with component
ii
Dyx
i
yxr −=
∈,
max
,
ni ≤≤1
,
1>d
,
1>u
,
u
0
,
i
ni
r
≤≤
<<
1
0
min0
,
nu
t
t
⋅
−
⋅=
λ
ρρ
0
for all
1≥t
, where
{}
0, ≥t
t
is the sequence used to impose lower
bounds on the random vectors generated in the SA
algorithm. Let the temperature-dependent generation
probability density function
()
t
Tp ,⋅
be given by
.,1log1
2
)1(
),(
1
n
d
t
i
n
i
t
i
t
t
z
T
z
T
z
T
a
Tzp ℜ∈
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
+
⎟
⎟
⎠
⎞
⎜
⎜
⎝
⎛
+⋅
−
=
∏
=
Then, for any initial point
Dx ∈
0
, the sequence
0);( ≥txf
t
of objective function values converges
in probability to the global minimum
*
f
, if the
temperature sequence
}
0, ≥tT
t
determined by the
temperature updating function satisfies the following
condition:
⎟
⎟
⎟
⎠
⎞
⎜
⎜
⎜
⎝
⎛
⋅−⋅=
⋅nd
t
tlTT
1
0
exp
,
...,,,i 21=
where
0
0
>T
is the initial temperature value and
0>l
is a given real number (Yang, 2000).
Typically a different form of the temperature
updating function has to be used with respect to a
different kind of the generation probability density
function in order to ensure the global convergence of
the corresponding SA algorithm. Furthermore, the
flatter is the tail of the generation probability
function, the faster is the decrement of the
temperature sequence determined by the temperature
updating function.
4 SVM CLASSIFICATION
Data classification is a common problem in science
and engineering. Support Vector Machines (SVMs)
are powerful tools for classifying data that are often
used in data mining operations.
In the standard binary classification problem, a
set of training data
ii
y,u , … ,
mm
y,u is
observed, where the input set of points is
ni
Uu ℜ⊂∈ , the
i
y is either +1 or −1, indicating
the class to which the point
i
u belongs,
}
11 −+∈ ,y
i
. The learning task is to create the
classification rule
{}
11 −+→ ,U:f that will be
used to predict the labels for new inputs. The basic
idea of SVMs classification is to find a maximal
margin separating hyperplane between two classes.
It was first described by Cortes and Vapnik (Cortes
& Vapnik, 1995). The standard binary SVM
classification problem is shown visually in Figure 1.
ICORES 2012 - 1st International Conference on Operations Research and Enterprise Systems
152