4 STEEPEST DESCENT METHOD
In this section, we summarize the main ideas behind
the so-called steepest descent method (Richard and
Faires 2001) that provides a deterministic means for
solving a system of nonlinear equations, and then we
employ the steepest descent method to solve the
equations in (10) to form our algorithm.
We expand the expression in (10) as
()
12 1 2
1212
;
;1,,
0;
1, , ; 1, ,
qj qiiijij
iiii n
faa
qQjm
ϕ
≤=
==
=∀=
∑
a
"
""
., (12)
and then define our optimal goal function when
combining the constraint in (2) as
() () ()
222
1
,
1, ,
Q
jqjjj
q
gf S
jm
γ
=
+
∀=
∑
aaa
"
. (13)
Here,
is added as a homogeneous factor,
which is applied to make the different square items
in (13) well-proportioned, and in our algorithm we
set
0.1
= . Notice that we don’t think it as a
penalty term for imposing the constraint for it just
adjust the constraint in (2) and (12) to have the same
level of function values. To satisfy the constraint (2)
we normalize
j
a in each iterative step to unit vector.
The direction of greatest decrease in the value
of
()
j
g a at
k
j
a with
k
-th iteration is the
direction given by its minus gradient
)
g−∇ a of
()
g a . The gradient is expressed as
() ()
()
T
2
jj
g∇=aJaFx. (14)
Here,
() () () ()
()
T
1
,, ,
Q
ffS
γ
=Fx x x x" , and
()
j
Ja is its Jacobian matrix. The objective is to
reduce
()
g a to its minimal value of zero, and an
appropriate choice for
()
g a is
( ) () ()
()
1kk k
jj j
g
α
+
=−∇aa a, (15)
where
()
()
1
0
arg min
k
j
g
α
α
+
= a is the critical point.
We can apply any single-variable function optimal
method to find the minimum value of
()
)
1k
j
g
+
a
by an appropriate choice for the value
. In our
algorithm, we use Newton’s forward divided-
difference interpolating polynomial, detailed in
Richard and Faires 2001.
We employ the initial solutions as equal
distributed vectors in the super space of
j
a . To
guarantee that all the local minimums of the
proposed algorithm are obtained, we can use 8 or
more initial solutions equal distributed in the super
space, and then find the correct solutions by
clustering method. For simplicity, we decide the
four correct solutions by their minimum distances
between each other.
5 SIMULATIONS
We adopt three mixtures of four speech signals the
same example as in Laar et al. 2008. For
convenience, we name our algorithm as MIBI
Steepest Descent and the algorithm in Laar et al.
2008 as MIBI Homotopy.
The speech signals are sampled as 8kHz,
consist of 10,000 samples with 1,250ms length, and
are normalized to unit variance
1
s
= . The signal
sequences are partitioned into five disjoint blocks
consisting of 2000 samples, and for each block, the
one-dimensional sensor correlation functions are
computed for lags 1, 2, 3, 4 and 5. Hence, in total
for each sensor correlation functions 25 values are
estimated and employed, i.e., the employed noise-
free ROS in the domain of block-lag pairs is given
by
)
)
)()
}
1, 1 , , 1, 5 , 2,1 , , 5, 5Ω= "",
where the first index in each pair represents the
block index and the second the lag index. The
sensor signals are obtained from (1) with
34
mixing matrix,
0.6749 0.4082 0.8083 0.1690
0.5808 0.8165 0.1155 0.5071
0.4552 0.4082 0.5774 0.8452
−
⎤
⎥
=− −
⎥
⎥
−−
⎦
A
.
The noise signals are mutually statistically
independent white Gaussian noise sequences with
variances
2
1
ν
σ
. The signal-to-noise ratio (SNR)
is -1.23dB, which is quite bad. We set the maximum
iterative number is 30, and stop the iteration if the
correction of the estimated is smaller than a certain
tolerance 10
-3
.
Let
i
be the included angle between the
-th
column of
A and its estimate. The estimated
mixing matrix is
SIGMAP 2009 - International Conference on Signal Processing and Multimedia Applications
120