H
Measurement-feedback Tracking with Preview
Eli Gershon
Holon Institute of Technology, HIT, Holon, Israel
Keywords:
H
Control, Discrete-time Systems, Tracking Preview.
Abstract:
Finite-horizon H
output-feedback tracking control for linear discrete time-varying systems is explored along
with the stationary infinite-horizon case. We consider three tracking patterns depending on the nature of the
reference signal i.e : whether it is perfectly known in advance, measured on-line or previewed in a fixed time-
interval ahead. For each of the above three cases a solution is found where, given a specific reference signal,
the controller plays against nature which chooses the initial condition and the energy-bounded disturbances.
The problems are solved based on a specially devised bounded real lemma for systems with tracking signals.
The finite-horizon case is extended to the stationary one where similar results are achieved.
1 INTRODUCTION
In the present paper we address the problem of
H
output-feedback tracking control with preview of
discrete-time linear systems, in both the finite and the
infinite time settings.
The stability analysis and control design for sys-
tems with stochastic uncertainties have received much
attention in the past (see (Gershon and U. Shaked,
2005) and the references therein), where mainly
continuous-time non retarded systems were conside-
red. In the late 80’s, a renewed interest in the cont-
rol and estimation designs of these systems has been
encountered and solutions to the stochastic control
and filtering problems of both: continuous-time and
discrete-time systems, have been derived that ens-
ure a worst case performance bound in the H
sense
(see (Gershon and U. Shaked, 2005) for an exten-
sive review). Systems whose parameter uncertain-
ties are modeled as white noise processes in a li-
near setting have been treated in (Dragan and Mo-
rozan, 1997a) (Hinriechsen and Pritchard, 1998), for
the continuous-time case and in (Dragan and Moro-
zan, 1997b), (Bouhtouri et al., 1999) for the discrete-
time case. Such models of uncertainties are encoun-
tered in many areas of applications (see (Gershon and
U. Shaked, 2005) and the references therein) such as:
nuclear fission and heat transfer, population models
and immunology. In control theory such models are
encountered in gain scheduling when the scheduling
parameters are corrupted with measurement noise.
Tracking feedback-control has been a central pro-
blem in both the frequency domain and in the state-
space domain over the last decades. Numerous varia-
tions of this problem, by large, have been published in
both the continuous-time and in the discrete-time set-
tings (Yang and Zhang, 2008), (Wencheng Luo and
Ling, 2005),(Kim and Tao, 2002), (Wei-qian You,
2010) and (Gao and Chen, 2008),(Verriest and Flor-
chinger, 1995). The problem of tracking control with
preview has been solved in the deterministic case by
(Shaked and deSouza, 1995). In the latter work a pre-
viewed tracking signal appears in the system dyna-
mics allowing for three patterns of preview. The first
case deals with the inclusion of the tracking signal
where no preview is given [the simplest case] while
in the second case the preview signal is known for a
given fixed finite time interval in the future which is
smaller the system dynamic horizon. The third pat-
tern deals with the case where the previewed signal is
known along the full finite-horizon of the system dy-
namics. The latter problems were solved in (Shaked
and deSouza, 1995) by applying a game theoretical
approach where a saddle point strategy is adopted.
The solution of the finite-horizon, stochastic coun-
terpart, state-feedback tracking control with pre-
view was obtained in (Gershon and U. Shaked,
2005),(Gershon and U. Shaked, 2010), where three
preview patterns of the tracking signal were conside-
red. These include the simple case of on-line measu-
rement of the tracking signal and two patterns where
this signal is either previewed with a fixed time in-
terval ahead or perfectly known in advance along the
system horizon (Gershon and U. Shaked, 2010) (see
Gershon, E.
H
Measurement-feedback Tracking with Preview.
DOI: 10.5220/0006853903250331
In Proceedings of the 15th International Conference on Informatics in Control, Automation and Robotics (ICINCO 2018) - Volume 1, pages 325-331
ISBN: 978-989-758-321-6
Copyright © 2018 by SCITEPRESS Science and Technology Publications, Lda. All rights reserved
325
also (Gershon and U. Shaked, 2005) for further de-
tails). For all these patterns the solutions were obtai-
ned using a game theory approach where the con-
troller plays against nature which chooses the system
energy-bounded disturbance and the initial condition.
In the present paper, we extend the work of
(Gershon and U. Shaked, 2010) which deals with the
stochastic state-feedback tracking control [zeroing,
of course, the stochastic terms] to the case where
there is no full access to the state-vector and a dyn-
amic output-feedback strategy must be applied. Here,
we treat the case where correlated parameter uncer-
tainties appear in both the system dynamics and the
measurement matrices. An optimal output-feedback
tracking strategy is derived which minimizes the stan-
dard H
performance index, for the three tracking
patterns of the reference signal. In both, the finite-
horizon horizon case and the stationary one, a min-
max strategy is applied that yields an index of perfor-
mance that is less than or equal to a certain cost. We
address the problem via two approaches: In the finite-
horizon case we apply the Difference LMI (DLMI)
method (Gershon and U. Shaked, 2005) for the so-
lution of the Riccati inequality obtained, and in the
stationary case we apply a special Lyapunov function
which leads to an LMI based tractable solution.
Notation: Throughout the paper the superscript T
stands for matrix transposition, R
n
denotes the n di-
mensional Euclidean space, R
n×m
is the set of all
n × m real matrices, N is the set of natural num-
bers and the notation P > 0, (respectively, P 0) for
P R
n×n
means that P is symmetric and positive
definite (respectively, semi-definite). We denote by
L
2
(,R
n
) the space of square-integrable R
n
valued
functions on the probability space (,F ,P ), where
is the sample space, F is a σ algebra of a subset
of called events and P is the probability measure
on F . By (F
k
)
kN
we denote an increasing family
of σ-algebras F
k
F . We also denote by
˜
l
2
(N ; R
n
)
the n-dimensional space of nonanticipative stochas-
tic processes { f
k
}
kN
with respect to (F
k
)
kN
where
f
k
L
2
(,R
n
). On the latter space the following l
2
-
norm is defined:
||{ f
k
}||
2
˜
l
2
= E{
0
|| f
k
||
2
} =
0
E{|| f
k
||
2
} < ,
{ f
k
}
˜
l
2
(N ; R
n
),
(1)
where ||·|| is the standard Euclidean norm. We denote
by Tr{·} the trace of a matrix and by δ
i j
the Kronecker
delta function. Throughout the manuscript we refer
to the notation of exponential l
2
stability, or internal
stability, in the sense of (Bouhtouri et al., 1999) (see
Definition 2.1, page 927, there). By [Q
k
]
+
, [Q
k
]
we
denote the causal and non causal parts respecti-
vely, of a sequence {Q
i
, i = 1, 2, ..., N}.
2 PROBLEM FORMULATION
Given the following linear discrete time-varying sy-
stem:
x
k+1
= A
k
x
k
+B
2,k
u
k
+B
1,k
w
k
+ B
3,k
r
k
y
k
= C
2,k
x
k
+ D
21,k
n
k
,
(2a-b)
where x
k
R
n
is the state vector, y
k
R
z
is the mea-
surement vector, w
k
R
p
is a deterministic exogenous
disturbance, r
k
R
r
is deterministic reference signal
which can be measured on line or previewed, u
k
R
l
is the control input signal and x
0
is an unknown initial
state and where we denote
z
k
= C
k
x
k
+D
2,k
u
k
+D
3,k
r
k
, z
k
R
q
, k [0,N]
(3)
and we assume, for simplicity that:
[C
T
k
D
T
3,k
D
T
2,k
]D
2,k
= [0 0
˜
R
k
],
˜
R
k
> 0.
Our objective is to find a control law {u
k
} that mini-
mizes the energy of {z
k
} by using the available know-
ledge on the reference signal, for the worst-case of the
process disturbances {w
k
},{n
k
} and the initial condi-
tion x
0
. We, therefore, consider, for a given scalar
γ > 0, the following performance index:
˜
J
E
(r
k
,u
k
,w
k
,n
k
,x
0
)
=
kC
N
x
N
+D
3,N
r
N
k
2
+
||z
k
||
2
2
γ
2
[||w
k
||
2
2
+ ||n
k
||
2
2
]
γ
2
x
T
0
R
1
x
0
,
R
1
0.
(4)
Similarly to (Gershon and U. Shaked, 2010) we con-
sider three tracking problems differing on the infor-
mation pattern over {r
k
}:
1) H
-tracking with full preview of {r
k
}: The
tracking signal is perfectly known for the interval
k [0, N].
2) H
-tracking with no preview of {r
k
}: The
tracking signal measured at time k is known for i k.
3) H
-tracking with fixed-finite preview of {r
k
}:
At time k, r
i
is known for i min(N,k + h) where
h is the preview length.
In all the above three cases we seek a control law {u
k
}
of the form
u
k
= H
y
y
k
+ H
r
r
k
where H
y
is a causal operator and where the causa-
lity of H
r
depends on the information pattern of the
reference signal. The design objective is to minimize
max
˜
J
E
(r
k
,u
k
,w
k
,n
k
,x
0
) ∀{w
k
},{n
k
},
{u
k
} l
2
[0,N 1], x
o
R
n
,
where for all of the three tracking problems we derive
a controller {u
k
} which plays against it’s adversaries
{w
k
},{n
k
} and x
0
.
ICINCO 2018 - 15th International Conference on Informatics in Control, Automation and Robotics
326
3 FINITE-HORIZON
OUTPUT-FEEDBACK
TRACKING
We bring first the deterministic counterpart of the
state-feedback result of (Gershon and U. Shaked,
2010) which is based on a game theory approach and
which constitutes the first step in the solution of the
output-feedback control problem. We consider the sy-
stem of (2a) and (3) and obtain the following result:
Lemma 1: Consider the system of (2a), (3) and
˜
J
E
of (4) with the term of n
k
excluded. Given γ > 0,
the state-feedback tracking game possesses a saddle-
point equilibrium solution iff there exists Q
i
> 0,i
[0,N] that solves the following Riccati-type equation
Q
k
=A
T
k
M
k+1
A
k
+C
T
k
C
k
A
T
k
M
k+1
B
2,k
Φ
1
k
B
T
2,k
M
k+1
A
k
, Q(N)=C
T
N
C
N
,
M
k+1
= Q
k+1
[I γ
2
B
1,k
B
T
1,k
Q
k+1
]
1
,
Φ
k
= B
T
2,k
M
k+1
B
2,k
+
˜
R
k
.
(5a-c)
and satisfies
R
k+1
> 0, k [0 N 1], γ
2
R
1
Q
0
> 0,
where R
k+1
= γ
2
I B
T
1,k
Q
k+1
B
1,k
.
(6a-c)
When a solution exists, the saddle-point strategies are
given by:
x
0
= (γ
2
R
1
Q
0
)
1
θ
0
,
w
k
= R
1
k+1
B
T
1,k
[θ
k+1
+Q
k+1
(A
k
x
k
+B
2,k
u
k
+B
3,k
r
k
)]
u
k
= Φ
1
k
{B
T
2,k
M
k+1
[A
k
x
k
+ B
3,k
r
k
+ Q
1
k+1
θ
c
k+1
]}
(7a-c)
where the causal part of θ
k+1
is
θ
c
k+1
= [θ
k+1
]
+
(8)
and where θ
k
satisfies
θ
k
=
¯
A
T
k
θ
k+1
+
¯
B
k
r
k
, θ
N
= C
T
N
D
3,N
r
N
,
¯
A
k
= Q
1
k+1
S
1
k+1
A
k
, S
k+1
= M
1
k+1
+ B
2,k
T
1
k+1
B
T
2,k
,
¯
B
k
=
¯
A
T
k
Q
k+1
B
3,k
+C
T
k
D
3,k
, T
k+1
=
˜
R
k
,
(9)
The game value is then given by:
˜
J
E
(r
k
,u
k
,w
k
,x
0
)=kS
1
2
k+1
(Q
1
k+1
θ
k+1
+B
3,k
r
k
)k
2
2
−kQ
1
2
k+1
θ
k+1
k
2
2
+kD
3,k
r
k
k
2
2
+kD
3,N
r
N
k
2
+θ
T
0
(γ
2
R
1
Q
0
)
1
θ
0
.
(10)
Proof: (Gershon and U. Shaked, 2010) The proof is
based on adapting the standard completing to squares
arguments to the case. We bring in the Appendix
the Sufficiency part of the proof, which is needed
for the derivation of the BRL of the next section.
We note also, at this point, that the solution of the
output-feedback control problem is also based on
these derivations.
Remark 1: It is important to note that the signal of
θ
k
in (9) is admitted in the above derivation because
of the tracking signal which affects the dynamics of
(2a) (Gershon and U. Shaked, 2010). This signal
accounts for the nature of the tracking pattern, where
it’s causal part (i.e [θ
k+1
]
+
) appears in the structure of
the controller in accordance with the preview patterns.
Remark 2: Applying the result of Lemma 1 on the
specific pattern of the reference signal it is shown in
(Gershon and U. Shaked, 2010) that the saddle-point
controller strategy depends on the causal part of θ
k+1
(i.e [θ
k+1
]
+
), where θ
k+1
is given in (9). The latter
dependency on θ
k+1
appears also in the structure of
the control signal in the output-feedback case.
The solution of the output-feedback control pro-
blem involves 2 steps where the latter one is a filte-
ring problem of order n. A second Riccati equation
is thus achieved by applying the BRL to the dynamic
equation of the estimation error. The latter imposes
augmentation of the system to 2n order. This augmen-
ted system contains also a tracking signal component
and therefore one needs to apply a special BRL for
systems with tracking signal. We thus bring first the
following lemma:
3.1 BRL for Systems with Tracking
Signal
We consider the following system:
x
k+1
= A
k
x
k
+ B
1,k
w
k
+ B
3,k
r
k
z
k
= C
k
x
k
+ D
3,k
r
k
, z
k
R
q
, k [0, N]
(11a,b)
which is obtained from (2,a) and (3) by setting
B
2,k
0 and D
2,k
0. We consider the following in-
dex of performance:
J
B
(r
k
,w
k
,x
0
)
=
kC
N
x
N
+D
3,N
r
N
k
2
+||z
k
||
2
2
γ
2
[||w
k
||
2
2
]
γ
2
x
T
0
R
1
x
0
, R
1
0.
(12)
We arrive at the following theorem:
Theorem 1: Consider the system of (11a,b) and
J
B
of (12). Given γ > 0, J
B
of (12) satisfies J
B
˜
J(r,ε), ∀{w
k
} l
2
[0,N 1], x
o
R
n
, where
˜
J(r, ε) =
N1
k=0
θ
T
k+1
{B
1,k
R
1
k+1
B
T
1,k
}θ
k+1
+
N1
k=0
r
T
k
(D
T
3,k
D
3,k
)r
k
+||D
3,N
r
N
||
2
+ 2
N1
k=0
θ
T
k+1
˜
Q
1
k+1
(
¯
M
1
k+1
)
1
B
3,k
r
k
+ θ
T
0
ε
1
θ,
H
Measurement-feedback Tracking with Preview
327
if there exists
˜
Q
k
that solves the following Riccati-
type equation
˜
Q
k
=A
T
k
¯
M
k+1
A
k
+C
T
k
C
k
,
γ
2
I B
T
1,k
˜
Q
k+1
B
1,k
> 0,
˜
Q(0) = γ
2
R
1
εI,
(13a,c)
for some ε > 0 where
˜
θ
k
=
ˆ
A
T
k
˜
θ
k+1
+
ˆ
B
k
r
k
,
˜
θ
N
= C
T
N
D
3,N
r
N
,
ˆ
A
k
=
˜
Q
1
k+1
¯
M
k+1
A
k
,
ˆ
B
k
=
ˆ
A
T
k
˜
Q
k+1
B
3,k
+C
T
k
D
3,k
,
ˆ
B
k
=
ˆ
A
T
k
˜
Q
k+1
B
3,k
+C
T
k
D
3,k
,
¯
M
k+1
=
˜
Q
k+1
[I γ
2
B
1,k
B
T
1,k
˜
Q
k+1
]
1
.
(14a-e)
Proof: Unlike the state-feedback tracking control
in (Gershon and U. Shaked, 2010), the solution of
the BRL does not acquire saddle-point strategies
(Since the input signal u
k
is no longer an adversary).
It can, however, be readily derived based on the
first part of the sufficiency of Lemma 1 by setting
B
2,k
0 and D
2,k
0. In the Appendix we bring the
proof of the BRL as a derivation of the proof of the
state-feedback tracking control solution (which is not
included in (Gershon and U. Shaked, 2010). The
latter proof is also essential for the derivation of the
output-feedback tracking control solution.
Remark 3: The choice of ε > 0 in
˜
Q(0) of (13b)
reflects on both, the cost value (i.e
˜
J(r,ε)) of (13b)
and the minimum achievable γ. If one chooses
0 < ε << 1 then, the cost of
˜
J(r,ε) increases while
the solution of (13a) is easier to achieve, which
results in a smaller γ. The choice of large ε, on the
other hand, causes the reverse effect, which leads to a
larger γ.
3.2 The Output-feedback Tracking
Control
We consider the system of (2a,b) and (3). Like in the
state-feedback case (Gershon and U. Shaked, 2010)
we seek a control law {u
k
}, based on the informa-
tion of the reference signal {r
k
} that minimizes the
tracking error between the system output and the
tracking trajectory, for the worst case of the initial
condition x
0
, the process disturbances {w
k
} and {n
k
}.
We, therefore, consider the performance index of (4)
and we assume that (5a) has a solution Q
k+1
> 0 over
[0,N] where (6a,b) are satisfied. The solution of the
output-feedback problem is stated in the following
theorem, for the a priori case, where u
k
can use the
information on {y
i
, 0 i < k} :
Theorem 2: Consider the system of (2a,b), (3) and
˜
J
E
of (4). Given γ > 0, the output-feedback tracking
control problem, where {r
k
} is known a priori for all
k N (the full preview case) possesses a solution if
there exists
ˆ
P
k
R
2n×2n
> 0, K
o,k
R
n×z
, i [0,N]
that solves the following Difference LMI (DLMI):
ˆ
P
1
k
˜
A
T
k
0
˜
C
T
1,k
˜
A
k
ˆ
P
k+1
γ
1
˜
B
1,k
0
0 γ
1
˜
B
T
1,k
I 0
˜
C
1,k
0 0 I
0, (15a,b)
where P
0
=
ˆ
Q
1
0
= γ
2
"
R R
R R + εI
n
#
with a
forward iteration, starting from the above initial
condition of P
0
, where R is defined in (4), where
˜
A
k
,
˜
B
1,k
,
˜
C
k
are defined in (23) and where
ˆ
Q
k
=
ˆ
P
1
k
is given in (24).
Proof: Using the expression that is achieved in the
Appendix for
˜
J
E
(r
k
,u
k
,w
k
,x
0
) in the state-feedback
case, the index of performance is now given by:
˜
J
E
(r
k
,u
k
,w
k
,n
k
,x
0
) =
˜
J
E
(r
k
,u
k
,w
k
,x
0
)
γ
2
||n
k
||
2
2
= γ
2
|| ¯w
k
||
2
2
γ
2
||x
0
x
0
||
2
R
1
γ
1
Q
0
+
N1
k=0
[{|| ¯u
k
+
ˆ
C
1,k
x
k
||
2
}]
+
γ
2
||n
k
||
2
2
+
¯
J(r).
(16)
We note that in the full preview case [θ
k+1
]
+
= θ
k+1
.
Using the following definitions:
¯u
k
= Φ
1/2
k+1
u
k
+ Φ
1/2
k+1
B
T
2,k
M
k+1
(B
3,k
r
k
+ Q
1
k+1
θ
k+1
)
¯w
k
= γ
1
R
1/2
k+1
w
k
γ
1
R
1/2
k+1
B
T
1,k
[Q
k+1
(A
k
x
k
+ B
2,k
u
k
+B
3,k
r
k
) + θ
k+1
],
(17)
we note that
¯w
k
= γ
1
R
1/2
k+1
(w
k
w
k
), ¯u
k
= Φ
1/2
k+1
(u
k
u
k
),
where w
k
, u
k
are defined in (7b,c) respectively. Note
also that the terms that are not accessed by the con-
troller (i.e the terms with x
k
), are exclude from u
k
.
Considering the above ¯w
k
and ¯u
k
we seek a controller
of the form
¯u
k
=
ˆ
C
1,k
ˆx
k
.
We, therefore, re-formulate the state equation of (2a)
adding the additional terms to recover the original
equation of (2a). Considering the above, we obtain
the following new state equation:
x
k+1
=
ˆ
A
k
x
k
+
¯
B
1,k
¯w
k
+
¯
B
2,k
¯u
k
+
¯
B
3,k
r
k
+
¯
B
4,k
θ
k+1
,
(18)
ICINCO 2018 - 15th International Conference on Informatics in Control, Automation and Robotics
328
where
ˆ
A
k
= Q
1
k+1
M
k+1
A
k
,
¯
B
1,k
= γB
1,k
R
1/2
k+1
,
¯
B
2,k
= Q
1
k+1
M
k+1
B
2,k
Φ
1/2
k+1
,
¯
B
3,k
= B
3,k
+ B
1,k
R
1
k+1
B
T
1,k
Q
k+1
B
3,k
¯
B
2,k
Φ
1/2
k+1
B
T
2,k
M
k+1
B
3,k
,
¯
B
4,k
=
¯
B
1,k
R
1
k+1
¯
B
T
1,k
¯
B
2,k
Φ
1/2
k+1
B
T
2,k
M
k+1
Q
1
k+1
.
(19a-e)
Replacing for ¯w
k
and ¯u
k
we obtain:
x
k+1
= (Q
1
k+1
M
k+1
A
k
)x
k
+
¯
B
1,k
(γ
1
R
1/2
k+1
(w
k
w
k
)) +
¯
B
2,k
(Φ
1/2
k+1
(u
k
u
k
)) +
¯
B
3,k
r
k
+
¯
B
4,k
θ
k+1
.
(20)
We consider the following Luenberger-type state
observer:
ˆx
k+1
=
ˆ
A
k
ˆx
k
+ K
o,k
(y
k
C
2,k
ˆx
k
) + d
k
, ˆx
0
= 0,
ˆz
k
=
ˆ
C
1,k
ˆx
k
,
(21)
where
d
k
=
¯
B
2,k
¯u
k
+
¯
B
3,k
r
k
+
¯
B
4,k
θ
k+1
.
Denoting e
k
= x
k
ˆx
k
and using the latter we obtain:
e
k+1
= (
ˆ
A
k
K
o,k
C
2,k
)e
k
+
ˆ
B
1,k
ˆw
k
,
where we define
ˆw
k
= [ ¯w
T
k
n
T
k
]
T
,
ˆ
B
1,k
= [
¯
B
1,k
K
o,k
D
21,k
].
Defining also ξ
k
= [x
T
k
e
T
k
]
T
, ¯r
k
= [r
T
k
θ
T
k+1
]
T
, we
obtain
ξ
k+1
=
˜
A
k
ξ
k
+
˜
B
1,k
ˆw
k
+
ˆ
B
3,k
¯r
k
,
˜z
k
=
˜
C
1,k
ξ
k
,
(22a-b)
where
˜
A
k
=
"
˜
A
11,k
¯
B
2,k
ˆ
C
1,k
0
˜
A
22,k
#
,
˜
B
1,k
=
"
¯
B
1,k
0
¯
B
1,k
K
o,k
D
21,k
#
,
ˆ
B
3,k
=
"
¯
B
3,k
¯
B
4,k
0 0
#
,
˜
A
11,k
=
ˆ
A
k
¯
B
2,k
ˆ
C
1,k
,
˜
A
22,k
=
ˆ
A
k
K
o,k
C
2,k
ˆ
C
1,k
= Φ
1/2
k+1
B
T
2,k
M
k+1
A
k
,
˜
C
1,k
= [0
ˆ
C
1,k
].
(23a-g)
Applying the results of Theorem 2 to the system of
(22) we obtain the following Riccati-type equation:
ˆ
Q
k
=
˜
A
T
k
[
ˆ
Q
1
k+1
γ
2
˜
B
1,k
˜
B
T
1,k
]
1
˜
A
k
+
˜
C
T
k
˜
C
k
,
(24)
where
ˆ
Q
0
is given in (15b). Denoting
ˆ
P
k
=
ˆ
Q
1
k
and
using Schur complement we obtain the DLMI of
(15a). The latter DLMI is initiated with the initial
condition of (15b) which corresponds to the case
where a weighting γ
2
ε
1
I
n
is applied to ˆx
0
in order
to force nature to select ˆx
0
= 0 in the corresponding
differential game (see (Gershon and U. Shaked,
2005), Chapter 9, for details).
In the case where {r
k
} is measured on line, or with
preview h > 0, we note that nature strategy which is
not restricted to causality constraints, will be the same
as in the case of full preview of {r
k
}, meaning that ¯w
k
of (17) is unchanged. We obtain the following:
Lemma 2: H
Output-feedback Tracking with full
preview of {r
k
}: We obtain
¯u
k
= Φ
1/2
k+1
u
k
+Φ
1/2
k+1
B
T
2,k
M
k+1
(B
3,k
r
k
+Q
1
k+1
[θ
k+1
]
+
)
where we note that in this case [θ
k+1
]
+
= θ
k+1
. Sol-
ving (9) we obtain:
[θ
k+1
]
+
=
ˆ
Φ
k+1
θ
N
+
Nk1
j=1
Ψ
k+1, j
¯
B
N j
r
N j
where
ˆ
Φ
k+1
=
¯
A
T
k+1
¯
A
T
k+2
...
¯
A
T
N1
Ψ
k+1, j
=
¯
A
T
k+1
¯
A
T
k+2
...
¯
A
T
N j1
j < N k 1
I j = N k 1
(25a-b)
Proof: Considering (9) and taking k + 1 = N we
obtain:
θ
N1
=
¯
A
T
N1
θ
N
+
¯
B
N1
r
N1
,
where θ
N
is given in (9). Similarly we obtain for N 2
θ
N2
=
¯
A
T
N2
θ
N1
+
¯
B
N2
r
N2
=
¯
A
T
N2
[
¯
A
T
N1
θ
N
+
¯
B
N1
r
N1
] +
¯
B
N2
r
N2
=
¯
B
N2
r
N2
+
¯
A
T
N2
¯
A
T
N1
θ
N
+
¯
A
T
N2
¯
B
N1
r
N1
.
The above procedure is thus easily iterated to yield
(25a,b). Taking, for example N = 3 one obtains from
(9) the following equation for θ
1
:
θ
1
=
¯
A
T
1
¯
A
T
2
θ
3
+
¯
A
T
1
¯
B
2
r
2
+
¯
B
1
r
1
.
The same result is recovered by taking k = 0 in (25a,b)
where j = 1,2.
Lemma 3: H
Output-feedback Tracking with no
preview of {r
k
}: In this case [θ
k+1
]
+
= 0 since at time
k, r
i
is known only for i k. We obtain:
¯u
k
= Φ
1/2
k+1
u
k
+ Φ
1/2
k+1
B
T
2,k
M
k+1
B
3,k
r
k
.
Lemma 4: H
Output-feedback tracking with
fixed-finite preview of {r
k
}: In this case we obtain:
¯u
k
= Φ
1/2
k+1
u
k
+Φ
1/2
k+1
B
T
2,k
M
k+1
(B
3,k
r
k
+Q
1
k+1
[θ
k+1
]
+
)
and
d
k
=
¯
B
2,k
¯u +
¯
B
3,k
r
k
+
¯
B
4,k
[θ
k+1
]
+
.
where [θ
k+1
]
+
satisfies :
[θ
k+1
]
+
=
h
j=1
¯
Ψ
k+1, j
¯
B
k+h+1 j
r
k+h+1 j
k + h N 1
ˆ
Φ
k+1
θ
N
+
h1
j=1
¯
Ψ
k+1, j
¯
B
N j
r
N j
k + h = N
.
where
¯
Ψ
k+1, j
is obtained from (25b) by replacing N
by k + h + 1.
H
Measurement-feedback Tracking with Preview
329
4 STATIONARY
OUTPUT-FEEDBACK
TRACKING CONTROL
We treat the case where the matrices of the system in
(2) and (3) are all time-invariant, N tends to infinity
and the system of (2) is exponential l
2
stable. In this
case, the solution
˜
Q
k
of (13a,b), if it exists, will tend to
the mean square stabilizing solution of the following
equation:
˜
Q = A
T
¯
MA +C
T
C, γ
2
I B
T
1
˜
QB
1
,
˜
Q(0) = γ
2
R
1
εI.
(26a,c)
We Introduce the following Lyapunov function:
V
k
= ξ
T
k
˜
Qξ
k
, with
˜
Q =
"
Q α
ˆ
Q
α
ˆ
Q
ˆ
Q
#
, (27)
where ξ
k
is the state vector of (22), Q and
ˆ
Q are n × n
matrices and α ia a scalar. Considering (26) and the
above result, we obtain the following theorem:
Theorem 3: Consider the system of (2), (3) where
the matrices A,B
1
,B
2
,B
3
,C
2
,D
21
,C,D
2
and D
3
are
all constant and T . Given γ > 0, there exists
a controller that minimizes max
˜
J
E
of (4) if there exist
Q = Q
T
R
n×n
,
ˆ
Q =
ˆ
Q
T
R
n×n
, Y R
n×q
and a
tuning scalar parameter α that satisfy
˜
ϒ > 0 where
˜
ϒ
is the following LMI:
Q α
ˆ
Q ϒ(1,3) ϒ(1,4) 0 0 0
ˆ
Q ϒ(2,3)
˜
ϒ(2,4) 0 0
ˆ
C
T
1
Q α
ˆ
Q ϒ(3,5) ϒ(3,6) 0
ˆ
Q ϒ(4,5) 0 0
I 0 0
I 0
I
(28)
where
Y
= K
T
o
ˆ
Q,ϒ(1,3) =
ˆ
A
T
Q
ˆ
C
T
1
¯
B
T
2
Q,
ϒ(1,4) = (
ˆ
A
T
ˆ
C
T
1
¯
B
T
2
)α
ˆ
Q,
ϒ(2,3) =
ˆ
C
T
1
¯
B
T
2
Q + α
ˆ
A
T
ˆ
Q αC
T
2
Y,
ϒ(2,4) = (α
ˆ
C
T
1
¯
B
T
2
+
ˆ
A
T
)
ˆ
Q C
T
2
Y,
ϒ(3,5) = γ
1
(Q + α
ˆ
Q)
¯
B
1
, ϒ(3,6) = γ
1
αY
T
D
21
,
ϒ(4,5) = γ
1
ˆ
Q
¯
B
1
(1 + α).
Proof: The proof outline for the above stationary
case resembles the one of the finite-horizon case.
Considering the stationary version of (2), (3) the sta-
tionary state-feedback control problem is solved to
obtain the optimal stationary strategies of both w
s,k
and u
s,k
(Gershon and U. Shaked, 2010). Thus we
obtain:
w
s,k
= R
1
k+1
B
T
1
[θ
k+1
+P(Ax
k
+B
2
u
k
+B
3
r
k
)],
u
s,k
= Φ
1
k
{B
T
2
ˆ
M[Ax
k
+ B
3
r
k
+ Pθ
c
k+1
]},
ˆ
M
= P
1
[I γ
2
B
1
B
T
1
P
1
]
1
,
where P
1
is the stationry version of the solution Q
k
of (5).
Using the above optimal strategies we transform
the problem to an estimation one, thus arriving to
the stationary counterpart of the augmented system of
(22). Applying the result of (26) to the latter system
the algebraic counterpart of (24) is obtained which,
similarly to the finite-horizon horizon case, becomes
the following stationary version of (15):
˜
P
1
˜
A
T
0
˜
C
T
1
˜
A
˜
P γ
1
˜
B
1
0
0 γ
1
˜
B
T
1
I 0
˜
C
1
0 0 I
0. (29)
Multiplying the above LMI by diag{I
n
,
˜
P
1
,I
2p
,I
l
}
from the left and the right, denoting
˜
Q =
˜
P
1
and
using the matrix partition of
˜
Q of (27), the result of
(28) is obtained.
5 CONCLUSIONS
In this paper we solve the problem of deterministic
output-feedback tracking control with preview. Un-
like the state-feedback case the solution is not obtai-
ned by applying a game theory approach, (where a
saddle-point tracking strategy is derived) but rather
as a min-max optimization problem. The output-
feedback tracking problem is solved by applying
a special form of the BRL to a filtering problem
which is formulated once the state-feedback solution
is obtained. The parameters of the a priori type state
observer used in our proof are recovered by iterative
solution of the DLMI of Theorem 2 which can be ea-
sily implemented. The result of the finite horizon case
were extended to the stationary case, where a simple
LMI condition is formulated for all the three preview
patterns. The theory is demonstrated by a numerical
example.
REFERENCES
Bouhtouri, A. E., Hinriechsen, D., and Pritchard, A. (1999).
H
type control for discrete-time stochasic systems.
ICINCO 2018 - 15th International Conference on Informatics in Control, Automation and Robotics
330
In Int. J. Robust Nonlinear Control, vol. 9, pp. 923-
948.
Dragan, V. and Morozan, T. (1997a). Mixed input-output
optimization for time-varying ito systems with state
dependent noise. In Dynamics of Continuous, Discrete
and Impulsive Systems,vol. 3, pp. 317-333.
Dragan, V. and Morozan, T. (1997b). Mixed input-output
optimization for time-varying ito systems with state
dependent noise. In Dynamics of Continuous, Discrete
and Impulsive Systems, vol. 3, pp. 317-333.
Gao, H. and Chen, T. (2008). Network-based H
output
tracking control. In IEEE transactions on automatic
control, vol. 53 (3), pp. 655 - 667.
Gershon, E. and U. Shaked, I. Y. (2005). H
control and
estimation of state-multiplicative linear systems. In
Lecture Notes in Control and Information sciences,
LNCIS, Springer, vol. 318.
Gershon, E. and U. Shaked, N. B. (2010). H
preview
tracking control for retarded stochastic systems. In
Conference on Control and Fault Tollerant Systems
(SysTol10), Nice, France.
Hinriechsen, D. and Pritchard, A. (1998). Stochastic H
˙In
SIAM Journal of Control Optimization, vol. 36(5), pp.
1504-1538.
Kim, B. S. and Tao, T. C. (2002). An integrated feedforward
robust repetitive control design for tracking near peri-
odic time varying signals. In Japan-USA on Flexible
Automation, Japan.
Shaked, U. and deSouza, C. E. (1995). Continuous-time
tracking problems in an H
setting: a game theory ap-
proach. In IEEE Transactions on Automatic Control,
vol. 40(5).
Verriest, E. I. and Florchinger, P. (1995). Stability of sto-
chastic systems with uncertain time delays. In Systems
and Control Letters, vol. 24(1), pp. 41-47.
Wei-qian You, Huai-hai Chen, X.-d. H. (2010). Tracking
control research of high-order flexible structures on
the h-infinity control method. In 2nd International
Conference on Advanced Computer Control (ICACC).
Wencheng Luo, Y.-C. C. and Ling, K.-V. (2005). H
inverse
optimal attitude-tracking control of rigid spacecraft.
In Journal of Guidance, Control, and Dynamics, Vol.
28 (3), pp. 481-494.
Yang, G. H. and Zhang, X. N. (2008). Reliable H
flight
tracking control via state feedback. In American Con-
trol Conference Westin Seattle Hotel, pp. 1800-1805.
Appendix
Proof Sketch of Theorem 1: The proof of Theorem
1 is based on the proof of Lemma 1 which concerns
the state-feedback control problem (for a detailed
proof see (Gershon and U. Shaked, 2010), (Gershon
and U. Shaked, 2005)). We first bring in Part I, a
proof sketch of Lemma 1 which is also needed for
the solution of the output-feedback control. We then
bring, in Part II, the derivation of the sufficiency part
of Theorem 1, which is derived from Lemma 1, and
the necessity part of the proof of Theorem 1.
Part I: The proof of Lemma 1 follows the standard
line of applying a Lyapunov-type quadratic function
in order to comply with the index of performance.
This is usually done by using two successive com-
pleting to squares operations however, since the
reference signal of r
k
is introduced in the dynamics
of (2a), we apply a third completing to squares
operation with the aid of the fictitious signal of θ
k+1
.
This latter signal finally affects the controller design
through it’s causal part [θ
k+1
]
+
(for a detailed proof
see (Gershon and U. Shaked, 2010), (Gershon and
U. Shaked, 2005)).
Part II: The sufficient part of the proof of Theorem 1
stems from the above proof of Lemma 1 where B
2,k
=
0, D
2,k
= 0. Analogously to the proof of Lemma 1,
we obtain the following: J
B
(r
k
,u
k
,w
k
,x
0
) =
N1
k=0
|| ˆw
k
R
1
k+1
B
T
1,k
˜
θ
k+1
||
2
R
k+1
γ
2
||x
0
(γ
2
R
1
˜
Q
0
)
1
˜
θ
0
||
2
R
1
γ
2
˜
Q
0
.
where we replace θ
k+1
, Q
k
by
˜
θ
k+1
and
˜
Q
k
, respecti-
vely and where
˜
P
0
= [R
1
γ
2
˜
Q
0
]
1
, x
0
= γ
2
˜
P
0
˜
θ
0
= [γ
2
R
1
˜
Q
0
]
1
˜
θ
0
The necessity follows from the fact that for r
k
0, one
gets
˜
J(r,ε) = 0 (noting that in this case
˜
θ
k
0 in (14)
and therefore the last 3 terms in
˜
J(r,ε) of Theorem
1 are set to zero) and J
B
< 0. Thus the existence of
˜
Q > 0 that solves (13) is the necessary condition in
the BRL.
H
Measurement-feedback Tracking with Preview
331