quasi-Newton method. Details can be found at
http://www.mathworks.com/help/toolbox/optim/. We
needed to set L = 40 and used a spline interpolation
method to get reasonable results at all.
Figure 2 show the surface of the combined, opti-
mal control α
∗
−β
∗
at time step k = 98. The calcula-
tion time for one time step was 450 seconds.
In this example, our FB algorithm is 170 times
faster than the DP method. This is due to 1) the
smaller amount of function evaluations and 2) the dif-
ferent interpolation method needed.
Moreover, the α
∗
−β
∗
surface of the FB method
in Figure 2 is smooth, while the surface of the DP
method in Figure 3 already has become rough at time
step k = 98, indicating instability. One reason may be
that in the DP method, the optimization is performed
over the highly nonlinear value function V. In our
FB algorithm, the optimization step depends on the
functions Y
t
and Z
t
only.
0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
2
2.2
A
B
−0.3
−0.2
−0.1
0
0.1
0.2
0.3
Figure 3: Optimal control α
∗
−β
∗
at time t
k
= 1.96 using
the DP method.
5.3 Comparison of Computational Cost
We briefly compare our FB algorithm with the stan-
dard dynamic programming approach. Let M
n
= |I
l
|;
that is, M is the number of grid points in each dimen-
sion of space where we have n dimensions. Also,
let L be the number of calculated transition proba-
bilities (or, alternatively, the number of quantization
points/simulations if we use a Quantization/Monte
Carlo method). N is the number of time-steps and
let m be the number of iterations for the Newton-
Raphson method to reach a given level of accuracy
ε
h
. Assuming that for a given level of accuracy,
the same time grid and space grid may be used for
both algorithms, the computational cost for the dy-
namic programming approach is NM
n
Lm(1 + 2r
2
)
while the computational cost for the FB algorithm is
NM
n
[L(1+ n+ nd) + mr]. Assuming a large number
of simulations L are needed for each evaluation, the
FB algorithm is superior if:
2mr
2
> n(d + 1) + 1. (43)
It is clear that the FB algorithm has a significantly
lower computational cost if a nontrivialnumber of op-
timization iterations are required. For example, in one
dimension the FB algorithm has
3
2m
the computational
cost of the dynamic programming approach. The ad-
vantage of the FB algorithm is that we do not need
to optimize over the entire value function, which re-
quires one to recalculate the expectation in the value
function for at least m times. This is very computa-
tionally expensive. Instead, one only has to optimize
the Hamiltonian, which is a much simpler procedure.
6 CONCLUSIONS
We have proposed a complete numerical algorithm to
solve optimal control problems through the associated
FBSDE system. By complete we mean that the algo-
rithm explicitly includes the optimization step. Our
numerical approach is an alternative to standard dy-
namic programming methods. A comparison of com-
putational cost between the dynamic programming
method and the FBSDE method illustrate the advan-
tages of the FBSDE approach.
We included results of a numerical example that
commonly appears in finance and economics. These
results confirm the advantage in accuracy and compu-
tational efficiency of the FB algorithm compared to
the dynamic programming method for certain prob-
lem classes.
A next step would be to analyze the convergence
speed and the convergenceerror in theory and practice
in detail.
REFERENCES
Davis, M. and Norman, A. (1990). Portfolio selection with
transaction costs. In Mathematics of Operation Re-
search, Vol. 15, No. 4. The lnstitute of Management
Sciences/Operations Research Society of America.
Delarue, F. and Menozzi, S. (2006). A forward-backward
stochastic algorithm for quasi-linear pdes. In The An-
nals of Applied Probability, Vol. 16, No. 1, 140184.
Universite Paris.
Kushner, H. and Dupuis, P. (1992). Numerical Methods
for Stochastic Control Problems in Continuous Time.
Springer, London, 1st edition.
Nocedal, J. and Wright, S. (2006). Numerical Optimization.
Springer Series in Operation Research and Financial
Engineering, London, 2nd edition.
Peng, S. (1990). A general stochastic maximum principle
for optimal control problems. In Journal of Control
and Optimization, Vol. 28, No. 4, pp.966-979. SIAM.
Pham, H. (2009). Continuous-time stochastic control and
optimization with financial applications. In Stochas-
tic Modeling and Applied Probability 61. Springer-
Verlag Berlin Heidelberg.
A FORWARD-BACKWARD ALGORITHM FOR STOCHASTIC CONTROL PROBLEMS - Using the Stochastic
Maximum Principle as an Alternative to Dynamic Programming
89