solution of LQG problem. It is convenient to take
the same value of f
(0)
as an initial iteration in (15).
It can be shown (Tovoinen, 1983) that there is a con-
stant a > 0 such that for every α
k
∈ (0, a) it holds
J
f
(f
(k+1)
) < J
f
(f
(k)
),
if (
∂J
f
∂f
)
(k)
) 6= 0. Thus, the proper choice of step α
k
assures the convergence of the algorithm.
4 CONTROL UNDER VARIANCE
CONSTRAINT
In the case of variance constraint given by the inequal-
ity (4) the associated Lagrangian is
L = J + λ(σ
2
u
− c
2
) (18)
or alternatively, the Lagrangian L can be rewritten
L = trQ
x
R
x
+ (q
u
+ λ)σ
2
u
, (19)
where λ ≥ 0 is the Lagrange multiplier. The Kuhn-
Tucker necessary conditions for the constrained min-
imum of L are
∂L
∂λ
≤ 0,
∂L
∂f
= 0. (20)
The optimal variance constrained control strategy can
be computed by solving the conditions (20). In prac-
tice, this is done iteratively, as it will be shown in Sec-
tion 5.
The controller to be designed is of the form
u
t
= f
T
ˆx
t
, (21)
where f
follows from appropriate Riccati equation
and ˆx
t
is the Kalman filter output. The minimiza-
tion of the Lagrangian (19) w.r.t. all admissible u
t
is
closely related to the minimization of the loss func-
tion J subject to the constraint (4). If u
∗
t
= f
∗T
ˆx
t
minimizes the Lagrangian (19), and the inequality
constraint (4) and complementary condition
λ(σ
2
u
− c
2
) = 0 (22)
are fulfilled at u
∗
t
, then u
∗
t
is also an optimal control
signal for variance-constrained control problem.
A major problem is the determination of appropriate
estimates for the Lagrange multiplier λ such that the
conditions (4) and (22) are satisfied for u
∗
t
. In prac-
tice this is done iteratively where each iteration step
k consists of solving a standard LQG problem, i.e.
of minimizing the Lagrangian (19) with λ = λ
(k)
and of updating the Lagrange multiplier according to
a suitable algorithm. A realization of this algorithm
needs the appropriate equations for R
ˆx
and σ
2
u
, (see
eqns.(25), (26)).
An iterative algorithm for updating the Lagrange mul-
tiplier λ
(k)
proposed in (M
¨
akil
¨
a, 1982, M
¨
akil
¨
a et at,
1984) can be combined with an algorithm described
in Section 3 to yield the algorithm given below.
5 SIMULTANEOUS AMPLITUDE
AND VARIANCE
CONSTRAINTS
First, it can be observed that the amplitude constraint
α (3) restricts itself the input variance because σ
2
u
≤
α
2
. Taking into account (4) and assuming c
2
= γσ
2
e
one obtains
γ ≤
α
2
σ
2
e
. (23)
This means that if for a given amplitude constraint α,
a given variance constraint has a form γ ≥
α
2
σ
2
e
then it
is automatically fulfilled and optimization of the feed-
back gain can only be performed wrt amplitude con-
straint as shown in Section 3. On the other hand, if
for a given α, a given variance constraint is such that
γ <
α
2
σ
2
e
then a problem may have an optimization
sense according to the problem formulated in Section
2. The proposed algorithm consists of the following
steps:
step 1: Take λ
(0)
> 0, h
0
= 1, 0 < α
0
< 1.
step 2: Calculate f
(k)
according to the method given
in Section 3 for
q
(k)
u
= q
u
+ λ
(k)
. (24)
step 3: Calculate R
(k)
ˆx
according to eqn. (11) taking
into account (12), (13), i.e.
R
(k)
ˆx
= F R
(k)
ˆx
F
T
+
+(F R
(k)
ˆx
f
(k)
g
T
+ g
f
T (k)
R
T (k)
ˆx
F
T
) ×
×g
2
(σ
(k)
) + g
g
T
f
T (k)
R
(k)
ˆx
f
(k)
g
1
(σ
(k)
) +
+k
k
T
σ
2
˜y
(25)
and
σ
2(k)
u
= f
T (k)
R
(k)
ˆx
f
(k)
g
1
(σ
(k)
), (26)
σ
2(k)
= f
T (k)
R
(k)
ˆx
f
(k)
. (27)
step 4: Check out the value (22), i.e.
ψ
(k)
= λ
(k)
(σ
2(k)
u
− c
2
). (28)
If ψ
(k)
is sufficiently close to zero, according to
some prescribed criterion then STOP, otherwise go
to step 5.
step 5: If k = 0, then go to step 6, otherwise update
h
k
(if positive) according to
h
k
= h
k−1
+
∆λ
(k)
+ h
k−1
∆ψ
(k)
∆ψ
(k)
, (29)
where ∆λ
(k)
= λ
(k)
− λ
(k−1)
, ∆ψ
(k)
= ψ
(k)
−
ψ
(k−1)
and ψ
(k)
is given by (27).