progress made in methods that are based on op-
timization analogies from nature, it became pos-
sible to achieve a remedy to some of the above
mentioned disadvantages through the use of global
methods of optimization. These include stochas-
tic methods, such as simulated annealing (Laarhoven
and Aarts, 1989), (Kirkpatrick and Vecchi, 1983)
and evolutionary computation methods (Fogel, 1998),
(Schwefel, 1995) such as genetic algorithms (GAs)
(Michalewicz, 1992), see also (Z. Michalewicz and
Krawczyk, 1992) for an interesting treatment of the
linear discrete-time problem.
Genetic algorithms provide a powerful mechanism
towards a global search for the optimum, but in many
cases, the convergence is very slow. However, as will
be shown in this paper, if the GA is supplemented
by problem specific heuristics, the convergence can
be accelerated significantly. It is well known that
GAs are based on a guided random search through
the genetic operators and evolution by artificial se-
lection. This process is inherently very slow, be-
cause the search space is very large and evolution pro-
gresses step by step, exploring many regions with so-
lutions of low fitness. However, it is often possible to
guide the search further, by incorporating qualitative
knowledge about potential good solutions. In many
problems, this might involve simple heuristics, which
when combined with the genetic search, provide a
powerful tool for finding the optimum very quickly.
The purpose of the present work is to incorporate
problem specific heuristic arguments, which when
combined with a modified hybrid GA, can solve
the discrete-time optimal control problem very eas-
ily. There are significant advantages to this approach.
First, the need to solve a difficult two-point boundary
value problem (TPBVP) is completely avoided. In-
stead, only initial value problems (IVP) need to be
solved. Second, after finding an optimal solution,
we verify that it approximately satisfies the first-order
necessary conditions for a stationary solution, so the
mathematical soundness of the traditional necessary
conditions is retained. Furthermore, after obtaining
a solution by direct genetic search, the static and dy-
namic Lagrange multipliers, i.e., the adjoint variables,
can be computed and compared with the results from
a gradient method. All this is achieved without di-
rectly solving the TPBVP. There is a price to be paid,
however, since, in the process, we are solving many
initial value problems (IVPs). This might present a
challenge in more advanced and difficult problems,
where the dynamics are described by higher order
systems of ordinary differential equations, or when
the equations are difficult to integrate over the re-
quired time interval and special methods of numer-
ical integration are required. On the other hand, if
the system is described by discrete-time difference
equations that are relatively well behaved and easy
to iterate, the need to solve the initial value problem
many times does not represent a serious problem. For
instance, the example problem presented here , the
discrete velocity programming problem (DVDP) with
the combined effects of gravity, thrust and drag, to-
gether with a terminal constraint (Bryson, 1999), runs
on a 1.6 GHz pentium 4 processor in less than one
minute CPU time.
In the next section, a mathematical formulation of
the discrete time optimal control problem is given.
This formulation is used to study a specific example
of a discrete time problem, namely the velocity di-
rection programming of a body moving in a viscous
fluid. Details of this problem are given in Section 3.
The evolutionary computation approach to the solu-
tion is then described in Section 4 where results are
presented and compared with the results of an indi-
rect gradient method developed by Bryson (Bryson,
1999). In Section 5, a mathematical formulation of
the continuous time optimal control problem for non-
linear dynamical systems is presented. A specific il-
lustrative example of a continuous time optimal con-
trol problem is described in Section 6, where we study
the Goddard’s problem of rocket dynamics using the
proposed evolutionary computation method. Finally
conclusions are summarized in Section 7.
2 OPTIMAL CONTROL OF
DISCRETE TIME NONLINEAR
SYSTEMS
In this section, a formulation is developed for the
nonlinear discrete-time optimal control problem sub-
ject to terminal constraints. Consider the nonlinear
discrete-time dynamical system described by differ-
ence equations with initial conditions
x(i + 1) = f[x(i), u(i), i] (2.1)
x(0) = x
0
(2.2)
where x ∈ R
n
is the vector of state variables, u ∈
R
p
, p < n is the vector of control variables and i ∈
[0, N − 1] is a discrete time counter. The function f
is a nonlinear function of the state vector, the control
vector and the discrete time i, i.e., f : R
n
x R
p
x
R 7→ R
n
. Next, define a performance index
J[x(i), u(i), i, N] = φ[x(N)]+Σ
M
i=0
L[x(i), u(i), i]
(2.3)
where M = N − 1, φ : R
n
7→ R, L : R
n
x R
p
x
R 7→ R
Here L is the Lagrangian function and φ[x(N)] is
a function of the terminal value of the state vector
ICINCO 2005 - INTELLIGENT CONTROL SYSTEMS AND OPTIMIZATION
46