and
m
k
= ( ¯u
k
, m
T
1,k
)
T
(28)
with m
1,k
= ( ¯u
k−1
, . . . , ¯u
k−nd
, ¯y
k
, . . . , ¯y
k−n f+1
, u
k−l+1
,
. . . , u
k−l−nr+2
, y
k−l+2
, . . . , y
k−l+ns+2
)
T
. The filtered
output and input signals are obtained as ¯y
k
=
A
∗
(q
−1
)y
k
, ¯u
k
= A
∗
(q
−1
)u
k
.
The corresponding diophantine equation and Be-
zout identity are
A(q
−1
)[r
0
+q
−1
R(q
−1
)]+q
−1
B(q
−1
)S(q
−1
)=r
0
A
∗
(q
−1
),
(29)
A(q
−1
)D(q
−1
) + B(q
−1
)F(q
−1
) = r
0
q
−l+2
, (30)
where the polynomial degrees are: nr = na − 1, ns =
na−κ−1, l = na+nb, nd = nb− 2, nf = na−1, and
κ is the number of possible integrators in the system.
It can be shown that the filtered output ¯y
k
can be
represented in the following regressor form
¯y
k
= p
T
m
k−1
+ v
k
(31)
For estimation of parameters p
(note that parameters
p
0
are included into p
) the Kalman filter algorithm
(13)-(16) can again be used where
ˆ
θ
k
should be re-
placed by ˆp
k
, s
k
should be replaced by m
k
, ε
k+1
should
be calculated as ε
k+1
= ¯y
k+1
−m
T
k
ˆp
k
, and the variance
σ
2
w
should be replaced by the variance σ
2
v
which can
be evaluated from (29), (30), (1).
4 SIMULATION TESTS
Performance of the described control methods is illus-
trated through the example of a second-order system
with the following true values: a
1
= −1.8, a
2
= 0.9,
b
1
= 1.0, b
2
= 0.5, where the Kalman filter algorithm
(13)-(16) was applied for estimation. The initial pa-
rameter estimates were taken half their true values
with P
0
= 10I. The reference signal was a square
wave ±3, and then the minimal value of constraint
α ensuring the tracking is α
min
= 3
|A(1)|
|B(1)|
= 0.2. Fig.
1 shows the reference, output and input signals dur-
ing tracking process under the constraint α = 1 for all
control policies.
For the control policy Π
MIDC
the constant learn-
ing weight was λ
k
= λ = 0.98. The policy Π
PP
was
simulated for third order polynomial A
∗
(q
−1
) having
poles at 0.2± i0.1, −0.1, and for the probing weight
η = 0.2. The control policy Π
CE
can easily be ob-
tained from MIDC by taking p
b
1
,k
= 0, p
T
b
1
θ
∗
,k
= 0.
Next, the simulated performance index
¯
J =
N−1
∑
k=0
(y
k+1
− r
k+1
)
2
was considered. The plots of
¯
J versus the constraint
α are shown in Figs. 2, 3 for σ
2
w
= 0.05, 0.1, re-
spectively, and N = 1000. The control u
TSDSC,qp
k
was
obtained solving the minimization of quadratic form
(20) using MATLAB function quadprog. The perfor-
mance of this control is not included in plots of Figs.
5, 6, because it performs surprisingly essentially in-
ferior with respect to u
TSDSC,co
k
. In the latter case,
a short-term behaviour phenomenon (G.P. Chen and
Hope, 1993) can be observed in Figs. 2, 3. This
means that when the cutoff method is used then the
range of constraint α can be found where for increas-
ing α the performance index is also increasing.
5 CONCLUSIONS
This paper presents various approaches toward a sub-
optimal solution to the discrete-time dual control
problem under the amplitude-constrained control sig-
nal. A simulation example of second-order system is
given and the performance of the presented control
policies is compared by means of the simulated per-
formance index.
The MIDC method seems to be a good suboptimal
dual control approach, however it has been found that
the MIDC control is quite sensitive to the value of the
learning weight λ. In (Kr
´
olikowski, 2000) it has been
found that this method often performs very close to
the IPS algorithm (Bayard, 1991).
Performance of all control policies except
Π
TSDSC,co
is comparable, however the differences be-
tween all methods are less noticeable when the con-
straint α gets tight, i.e. when α → α
min
. In all con-
sidered control policies except u
TSDSC,co
k
, the perfor-
mance index increases when the input amplitude con-
straint gets more tight. This means that for u
TSDSC,co
k
the effect of the short term behaviour phenomenon
discussed in (G.P. Chen and Hope, 1993) could ap-
pear.
REFERENCES
˚
Astr
¨
om, H. and Wittenmark, B. (1989). Adaptive Control.
Addison-Wesley.
Bayard, D. (1991). A forward method for optimal stochastic
nonlinear and adaptive control. IEEE Trans. Automat.
Contr., 9:1046–1053.
Bayard, D. and Eslami, M. (1985). Implicit dual control for
general stochastic systems. Opt. Contr. Appl.& Meth-
ods, 6:265–279.
Filatov, N. and Unbehauen, H. (2004). Dual Control.
Springer.
DUAL CONTROLLERS FOR DISCRETE-TIME STOCHASTIC AMPLITUDE-CONSTRAINED SYSTEMS
133