Beruflich Dokumente
Kultur Dokumente
Abstract
A general maximum principle (necessary and sufficient conditions)
for an optimal control problem governed by a stochastic differential
equation driven by an infinite dimensional martingale is established.
The solution of this equation takes its values in a separable Hilbert space
and the control domain need not be convex. The result is obtained by
using the adjoint backward stochastic differential equation.
Introduction
(1.1)
This work is supported by King Abdulaziz City for Science and Technology (KACST),
Riyadh, Saudi Arabia.
AbdulRahman Al-Hussein
seen from equation (5.2) in Section 5. It is quite important to realize that the
adjoint equations of such SDEs are in general BSDEs driven by martingales.
This happens also even if the martingale M, which is appearing in equation
(1.1), is a Brownian motion with respect to a right continuous filtration being larger than its natural filtration. There is a discussion on this issue in
Bensoussans lecture note [10, Section 4.4], and in [1] and its erratum [5].
The paper is organized as follows. Section 2 is devoted to some preliminary notation. In Section 3 we present our main stochastic control problems.
Then in Section 4 we establish many of our necessary estimates, which will be
needed to derive the maximum principle for the control problem of (1.1). The
maximum principle in the sense of Pontryagin for the above control problem
is derived in Section 5. In Section 6 we establish some sufficient conditions for
optimality for this control problem, and present some examples as well.
Preliminary notation
AbdulRahman Al-Hussein
We refer the reader to Example 6.2 for a precise computation of this process
Q().
For fixed (t, ), we denote by LQ(t,) (K) to the set of all linear operators
: Q1/2 (t, )(K) K and satisfy Q1/2 (t, ) L2 (K), where L2 (K) is the
space of all Hilbert-Schmidt operators from K into
itself.
The inner product
and norm in L2 (K)Rwill be denoted respectively by , 2 and || ||2. Then the
stochastic integral 0 (s)dM(s) is defined for mappings such that for each
(t, ), (t, ) LQ(t,) (K), Q1/2 (t, )(h) h K is predictable, and
Z T
E[
||(Q1/2 )(t)||22 dt ] < .
0
Such integrands
a Hilbert space with respect to the scalar product
R T
form
1/2
(1 , 2 ) 7 E [ 0 1 Q (t) , 2 Q1/2 (t) dt ]. Simple processes taking values
in L(K; K) are examples of such integrands. By letting 2 (K; P, M) be the
closure of the set of simple processes in this Hilbert space, it becomes a Hilbert
subspace. We have also the following isometry property:
i
hZ T
h Z T
i
2
= E
E |
(2.1)
(t)dM(t)|
||(t)Q1/2 (t)||22 ds
0
for mappings 2 (K; P, M). For more details and proofs we refer the reader
to [25].
On the other hand, we emphasize that the process Q() will be play an
important role in deriving the adjoint equation of the SDE (1.1) as it can be
seen from equations (5.1), (5.2) in Section 5. This is due to the fact that the
integrand is not necessarily bounded. More precisely, it is needed in order
for the mapping x H, which appear in both equations, to be defined on the
space L2 (K), since the process Z u() there need not be bounded. This always
has to be considered when working with BSDEs or BSPDEs driven by infinite
dimensional martingales.
Next let us introduce the following space:
Z T
2
LF (0, T ; E) := { : [0, T ] E, predictable and E [
|(t)|2 dt ] < },
0
Since Q(t) Q for all t [0, T ] a.s., it follows from [3, Proposition 2.2] that
if L2F (0, T ; LQ (K)) (where as above LQ (K)) = L2 (Q1/2 (K); K)), the space
of all Hilbert-Schmidt operators from Q1/2 (K) into K), then 2 (K; P, M)
and
Z T
Z T
1/2
2
E[
||(t)Q (t)||2 dt ] E [
||(t)||2LQ(K) dt ].
(2.2)
0
(3.2)
AbdulRahman Al-Hussein
Estimates
Let (X , u()) be the given optimal pair. Let 0 t0 < T be fixed and
0 < < T t0 . Let v be a random variable taking its values in U, Ft0 measurable and sup |v()| < . Consider the following spike variation of the
control u ():
u (t) =
u (t)
v
if t [0, T ]\[t0 , t0 + ]
if t [t0 , t0 + ].
(4.1)
Lemma 4.1 Let (E1) hold. Assume that {p(t), t0 t T } is the solution of
the following linear equation:
dp(t) = Fx (X (t), u (t)) p(t) dt + Gx (X (t)) p(t) dM(t), t0 < t T,
(4.2)
p(t0 ) = F (X (t0 ), v) F (X (t0 ), u (t0 )).
Then
sup E [ |p(t)|2 ] < C
t[t0 ,T ]
X (t) X (t) =
[F (X (s), v) F (X (s), v)] ds
t0
Z t
+
[F (X (s), v) F (X (s), u(s))] ds
t
Z 0t
[G(X (s)) G(X (s))]dM(s),
+
(4.3)
t0
or, in particular,
|X (t) X (t)| 3 (t t0 )
t0
AbdulRahman Al-Hussein
(4.6)
and
G(X (s)) G(X (s))
Z 1
=
Gx (X (s) + (X (s) X (s))) (X (s) X (s)) d
0
2
E[|
G(X (s)) G(X (s)) dM(s)| ] = E [ |
(s)dM(s)|2 ]
t0
t0
Z t
= E[
||(s)Q1/2 (s)||22 ds ]
t
Z 0t
E[
||(s)||2LQ(K) ds ]
t0
Z t Z 1
= E[
||
Gx (X (s) + (X (s) X (s))) (X (s) X (s)) d||2LQ(K) ds ]
t0
0
Z tZ 1
E[
||Gx (X (s) + (X (s) X (s))) (X (s) X (s))||2LQ(K) d ds ]
t0 0
Z t
C2 E [
|X (s) X (s)|2 ds ].
(4.8)
t0
Therefore, from (4.4), (4.5), (4.6), (E1) and (4.8), it follows evidently that
Z
t
2
E [ | X(s) X (s) |2 ] ds
E [ |X (t) X (t)| ] 3 C1 (t t0 ) + C2
t0
Z t
E [ | v u (s) |2 ] ds,
+ 3 (t t0 ) C3
t0
t0 +
E [ |v u (s)|2 ] ds,
t0
(4.9)
E [ |v u (s)|2 ] ds.
(4.10)
t0
It follows then from (4.4), (4.9), standard martingale inequalities, (4.8) and
(4.10) that
E[
sup
t0 tt0 +
|X (t) X (t)|2 ]
3 (C1 +C2 )
3 C3 [ 3 (C1 + 4C2 ) e
+ 1]
t0 +
2
2 C4 ()
E [ |v u (s)|2 ] ds
E[
| X (t) X (t) | dt ] 9 C3 e
t0 +
t0
and
E[
sup
t0 +tT
|X (t) X (t)|2 ] 27 C3 eC4 () [ 1 + (T t0 ) C1 + 4 C2 ]
t0 +
E [ |v u (s)|2 ] ds,
(4.13)
t0
where C4 () = [3 2 + 3 (T t0 )2 ] C1 + (T t0 + 2 ) C2.
Now (4.11) and (4.13) imply that
Z
t0 +
2
E [ |v u (s)|2 ] ds,
E [ sup |X (t) X (t)| ] C5 () + C6 ()
t0 tT
t0
10
AbdulRahman Al-Hussein
and
C6 () = 27 C3 eC4 () [ 1 + (T t0 ) C1 + 4 C2 ].
(4.14)
s
RT
But since 0 E [ |u(t) u (s)|2 ] dt < (for fixed s), then, as it is well-known
from measure theory (e.g. [13]), there exists a subset O of [0, T ] such that
P(O) = 1 and the mapping O t 7 E [ |u(t) u (s)|2 ] is continuous. Thus,
if s O, this function is continuous in a neighborhood of s, and so we have
Z
1 s+
E [ |u(t) u (s)|2 ] dt 0, as 0,
s
which by (4.15) implies (4.14) for = F.
We will choose t0 such that (4.14) holds for = F, . This assumption will
be considered until the end of Section 5.
Lemma 4.4 Assume (E1). Let
(t) =
1
(X (t) X (t)) p(t), t [t0 , T ].
Then
lim E [ |(T )|2 ] = 0.
11
1
+ [ G(X (t)) G(X (t)) Gx (X (t)) p(t) ]dM(t),
(t0 ) = F (X (t0 ), v) F (X (t0 ), u (t0 )) .
d (t) =
Thus
1
(t0 + ) =
t0 +
[ F (X (s), v) F (X (s), v) ] ds
t0
Z
1 t0 +
+
[ F (X (s), v) F (X (t0 ), v) ] ds
t0
Z
1 t0 +
[ F (X (t0 ), u (t0 )) F (X (s), u(s)) ] ds
+
t0
Z
1 t0 +
[ G(X (s)) G(X (s)) ]dM(s)
+
t0
Z t0 +
Fx (X (s), u (s))p(s) ds
t
Z 0t0 +
Gx (X (s))p(s)dM(s).
t0
sup
t0 tt0 +
Z t0 +
sup
t0 tt0 +
|X (t) X (t)|2 ]
6
E [ |F (X (s), u (s)) F (X (t0 ), u(t0 ))|2 ] ds
t0
6 C2
E [ sup |X (t) X (t)|2 ]
+
t0 tt0 +
Z t0 +
+ 6 (C1 + C2 ) E [
|p(s)|2 ds ].
+
t0
(4.16)
12
AbdulRahman Al-Hussein
t0 tt0 +
3 (C1 +C2 )
3 C3 [ 3 (C1 + 4C2 ) e
+ 1]
t0 +
E [ |v u (s)|2 ] ds
t0
as 0. (4.17)
as 0.
(4.18)
t0
Thus, by applying Lemma 4.2, (4.18), (4.17), (4.14) and Lemma 4.1 in
(4.16), we deduce
E [ | (t0 + ) |2 ] 0
as 0.
(4.19)
1
+ [ G(X (t)) G(X (t)) Gx (X (t)) p(t) ]dM(t),
(s) =
[ Gx (X (s) + (X (s) X (s))) Gx (X (s)) ] p(s) d,
0
we get
Z
(t) = (t0 + ) +
Fx (X (s) + (X (s) X (s)), u (s)) (s) d ds
t0 + 0
Z t Z 1
+
Gx (X (s) + (X (s) X (s))) (s) d dM(s)
t0 + 0
Z t Z 1
+
[ Fx (X (s) + (X (s) X (s)), u(s))
t0 +
t0 +
13
for all t [t0 + , T ]. Hence by making use of the isometry property (2.1) it
holds t [t0 + , T ],
2
E [ | (s) |2 ] ds
t0 +
+5E
hZ
|
t0
i2
Fx (X (s), u (s)) p(s) d ds |
Z T
(s) Q1/2 (s)||2 ds ].
+5E[
||
(4.20)
2
t0
But as done for the second equality and first inequality in (4.8) we can derive
easily that
E[
=
Gx (X (s)) ] p(s)||2LQ(K) d ds ].
(4.21)
14
AbdulRahman Al-Hussein
Lemma 4.5 Assume (E1) and (E2). Let be the solution of the equation:
d(t) = x (X (t), u(t))p(t)dt, t0 < t T,
(t0 ) = (X (t0 ), v) (X (t0 ), u(t0 )).
Then
h 1 Z T
2 i
(X (t), u (t)) (X (t), u (t)) dt (t) = 0.
lim E
0
t0
Proof. Let
1
(t) =
t
t0
(X (t), u (t)) (X (t), u (t)) dt (T ),
for t [t0 , T ]. Then (t0 ) = (X (t0 ), v)(X (t0 ), u(t0 )) . So one can proceed easily as done in the proof of Lemma 4.4 to show that E [ | (T ) |2 ] 0,
though this case is rather simpler.
Let us now for a C 1 mapping : K R denote by to the gradient
of , which is defined, by using the directional
derivative
D(x)(k) of at a
point x K in the direction of k K, as (x), k = D(x)(k) (= x (k)).
We shall sometimes write x for (x).
Corollary 4.6 Under the assumptions of Lemma 4.5
d
J(u ())=0 = E [ h(X (T )), p(T ) + (T ) ].
d
(4.22)
Proof. Note that from the definition of the cost functional in (3.2) we see
that
1 h
1
Z T
i
+
(X (s), u (s)) (X (s), u (s)) ds
t0
hZ 1
(X (T ) X (T ))
=E
d
hx (X (T ) + (X (T ) X (T )))
0
Z T
i
1
(X (s), u (s)) (X (s), u(s)) ds .
+
t0
Now let 0 and use the properties of h in (E1), Lemma 4.4 and Lemma 4.5
to get (4.22).
15
Maximum principle
The maximum principle is a good tool for studying the optimality of controlled
SDEs like (3.1) since in fact the dynamic programming approach for similar
optimal control problems require usually a Markov property to be satisfied by
the solution of (3.1), cf. for instance [36, Chapter 4]. But this property does
not hold in general especially when the driving noise is a martingale.
Let us recall the SDE (3.1) and the mappings in (3.2), and define the
Hamiltonian H : [0, T ] K O K L2 (K) R for (t, , x, u, y, z)
[0, T ] K O K L2 (K) by
H(t, , x, u, y, z) := (x, u) + F (x, u) , y + G(x)Q1/2 (t, ) , z 2 .
(5.1)
dY u() (t) = x H(t, X u() (t), u(t), Y u() (t), Z u() (t)Q1/2 (t)) dt
Z u() (t) dM(t) dN u() (t), t0 t < T, (5.2)
u()
(t) = +
Z
Z
t
Theorem 5.1 Assume that (E1)(E2) hold. Then there exists a unique solution (Y u() , Z u() , N u() ) of the BSDE (5.2).
For the proof of this theorem one can see [2].
We shall denote briefly the solution of (5.2), which corresponds to the
optimal control u () by (Y , Z , N ).
In the following lemma we shall try to compute E [ Y (T ), p(T ) ].
16
AbdulRahman Al-Hussein
Lemma 5.2
T
E [ Y (T ), p(T ) ] = E
x (X (s), u(s))p(s) ds
t0
+ E Y (t0 ), F (X (t0 ), v) F (X (t0 ), u(t0 ) . (5.3)
Proof. Use Itos formula together to compute d Y (t), p(t) for t [t0 , T ],
and use the facts that
Z T
p(s) , x H(s, X (s), u (s), Y (s), Z (s)Q1/2 (s)) ds
t0
t0
i
x (X (s), u (s))p(s) + Fx (X (s), u (s))p(s) , Y (s) ds
T
t0
ds,
(5.4)
E [ Y (T ), p(T ) + (T ) ] 0.
(5.5)
On other hand by applying (5.5) and Lemma 5.2 one sees that
Z T
0 E[
x (X (s), u (s))p(s) ds ]
t0
+ E [ Y (t0 ), F (X (t0 ), v) F (X (t0 ), u (t0 ) + (T ) ]. (5.6)
But
(T ) = (t0 ) +
t0
x (X (s), u (s))p(s) ds
17
and
H(t0 , X (t0 ), v, Y (t0 ), Z (t0 )Q1/2 (t0 ))
H(t0 , X (t0 ), u (t0 ), Y (t0 ), Z (t0 )Q1/2 (t0 ))
= (t0) + Y (t0 ), F (X (t0 ), v) F (X (t0 ), u (t0 )) .
(5.7)
Now varying t0 as in (4.14) shows that (5.7) holds for a.e. t., and so by
arguing for instance as in [10, P. 19] we obtain easily (5.4).
Remark 5.4 Let us assume for example that the space K in Theorem 5.1 is
the real space R and M is the martingale given by the formula
Z t
M(t) =
(s)dB(s),
t [0, T ],
0
L2F (0, T ; R)
for some
and a one dimensional Brownian motion B. If
(s) > 0 for each s, then Ft (M) = Ft (B) for each t, where
Ft (R) = {R(s), 0 s t}
for R = M, B. Consequently, by applying the unique representation property for
martingales with respect to {Ft (M), t 0} or larger filtration in [2, Theorem
2.2] or [5] and the Brownian martingale representation theorem as e.g. in [14,
Theorem 3.4, P. 200], we deduce that the martingale N u() in (5.2) vanishes
almost surely if the filtration furnished for the SDE (3.1) is {Ft (M), 0 t
T }. This result follows from the construction of the solution of the BSDE (5.2).
More details on this matter can be found in [2, Section 3]. As a result, in this
particular case BSDE (5.2) fits well with those BSDEs studied by Pardoux &
Peng in [27], but with the variable Z replacing Z there.
Thus in particular we conclude that many of the applications of BSDEs,
which were studied in the literature, to both stochastic optimal control and
finance (e.g. [37] and the references therein) can be applied directly or after
slight modification to work here for BSDEs driven by martingales. For example
we refer the reader to [23] for financial application. Another interesting case
can be found in [9].
On the other hand, in this respect we shall present an example (see Example 6.4) in Section 6, by modifying an interesting example due to Bensoussan
[10].
18
AbdulRahman Al-Hussein
and
I2 := E [ h(X (T )) h(X u() (T )) ].
Then readily
J(u ()) J(u()) = I1 + I2 .
Let us define
I3 := E
I4 := E
I5 := E
H(t, X u() (t), u(t), Y (t), Z (t) Q1/2 (t)) dt ,
T
F (X (t), u (t)) F (X u() (t), u(t)) , Y (t) dt ,
G(X u
()
(t)) G(X u() (t)) Q1/2 (t) , Z (t)Q1/2 (t) 2 dt ,
(6.1)
19
and
h Z
I6 := E
i
X (t) X u() (t) dt .
(6.2)
h(X (T )) , X (T ) X u() (T )
a.s.,
I2 E [ Y (T ) , X (T ) X u (T ) ].
(6.3)
I2 I4 + I5 I6 .
(6.4)
Next by applying Itos formula to compute d Y (t) , X (t) X u() (t) and
using equations (5.2) and (3.1) we find with the help of (6.3) that
(6.5)
H(t, X (t), u (t), Y (t), Z (t)Q1/2 (t))
Z
Z
H(t, X u()(t), u(t), Y (t), Z (t)Q1/2 (t)) dt
T
0
X (t) X u() (t) dt
u H(t, X (t), u(t), Y (t), Z (t)Q1/2 (t)) , u (t) u(t) O dt.
20
AbdulRahman Al-Hussein
As a result
I3 I6 + I7 ,
(6.6)
where
I7 = E
h Z
u H(t, X (t), u (t), Y (t), Z (t)Q1/2 (t)) , u(t) u(t) O 0.
Therefore I7 0, which by (6.6) implies that I3 I6 0. So (6.5) becomes
J(u ()) J(u()) 0.
Now since u() Uad is arbitrary, this inequality proves that (X , u ()) is
an optimal pair for the control problem (3.1)-(3.3) as required.
Example 6.2 Let m be a continuous square integrable
R t one dimensional martingale with respect to {Ft }t such that < m >t = 0 (s)ds 0 t T for
Rt
some continuous : [0, T ] (0, ). Consider M(t) = m(t)(= 0 dm(s)),
with 6= 0 being a fixed element of K. Then M M2,c (K) and << M >>t
Rt
^
equals
0 (s)ds, where ^
is the identification of in L1 (K),
Rt
^
that is (
)(k) = , k , k K. Also < M >t = ||2 0 (s) ds. Now
Rt
^
letting Q(t) =
(t) yields that << M >>t = 0 Q(s) ds. This process
^
Q() is bounded since Q(t) Q t, where Q =
max (t). It is also
0tT
,k
easy to see that Q1/2 (t)(k) = || 1/2 (t). In particular Q1/2 (t)(K).
Let K = L2 (Rn ). Let M be the above martingale. Suppose that O = K.
LQ (K) or even a bounded linear operator from K into itself,
Assume that G
and F is a bounded linear operator from O into K. Let us consider the SDE:
dM(t), t [0, T ],
dX(t) = F u(t) dt + X(t) , G
X(0) = x0 K.
For a given fixed element c of K we assume that the cost functional is given
by the formula:
Z T
J(u()) = E [
|u(t)|2 dt ] + E [ c , X(T ) ],
0
21
(x, u) = |u|2, and h(x) = c , x ,
F (x, u) = F u, G(x) = x , G,
where (x, u) K O.
The Hamiltonian then becomes the mapping
H : [0, T ] K O K L2 (K) R,
Q1/2 (t) , z ,
H(t, x, u, y, z) = |u|2 + F u , y + x , G
2
(t, x, u, y, z) K O K L2 (K).
It is obvious that H(, ,
y, z) is convex
with respect to (x, u) for each y and
1/2
Q1/2 (t) , Z(t) ] dt Z(t) dM(t) dN(t),
dY (t) = [ G
2
Y (T ) = c.
This BSDE attains an explicit solution Y (t) = c , since c is non-random. But
this implies that Z(t) = 0 and N(t) = 0 for each t [0, T ].
On the other hand, we note that the function O u 7 H(t, x, u, y, z) R
attains its minimum at u = 21 F y, for fixed (x, y, z). So we choose our
candidate for an optimal control as
u (t, ) =
1
1
F Y (t, ) = F c ( U := O),
2
2
.
With this choice all the requirements in Theorem 6.1 are verified. Consequently u () is an optimal control of this control problem with an optimal
given by the solution of the following closed loop equation:
solution X
, G
dM(t),
dX(t)
= 21 F F Y (t) dt + X(t)
X(0)
= x0 K.
The value function takes the following value:
J =
1 2
) ].
|F c| T + E [ c , X(T
4
22
AbdulRahman Al-Hussein
Remark 6.3 It would be possible if we take h(x) = |x|2 , x K, in the preceding example and proceeds as above. However if a result of existence and
uniqueness os solutions to what we may call forward-backward stochastic differential equations with martingale noise holds, it should certainly be very
useful to deal with both this particular case and similar problems.
Example 6.4 Let O = K. We are interested in the following linear quadratic
example, which is gleaned from Bensoussan [10, P. 33]. Namely, we consider
the SDE:
dX(t) = (A(t)X(t) + C(t)u(t) + f (t)) dt + (B(t)X(t) + D(t)) dM(t),
(6.7)
X(0) = x0 ,
P (t)X u() (t) , X u() (t) +
Q(t)u(t) , u(t) dt
J(u()) = E
2
2
0
i
1
(6.9)
+ D(t),
G(t, x) = (t) , x G(t)
1
(t, x, u) =
P (t)x , x +
Q(t)u , u ,
2
2
1
P1 x , x .
h(x) =
2
23
H(t, x, u, y, z) = (t, x, u) + F (t, x, u) , y + G(t, x)Q1/2 (t) , z 2
1
P (t)x , x +
Q(t)u , u
=
2
2
+ A(t)x + C(t)u + f (t) , y
+ D(t))Q1/2 (t) , z .
+ ( (t) , x G(t)
2
1/2
u()
dY
(t)
=
A (t)Y u() (t) + P (t)X u() (t)
1/2
u()
1/2
Y u() (T ) = P1 X u() (T ).
Now the maximum principle theorems (Theorem 5.3, Theorem 6.1) in this
case hold readily if we consider Remark 3.1 again, and yield eventually
C (t)Y (t) + Q(t)u (t) = 0.
References
[1] A. Al-Hussein, Backward stochastic partial differential equations in infinite dimensions, Random Oper. and Stoch. Equ., 14, 1 (2006), 122.
[2] A. Al-Hussein, Backward stochastic partial differential equations driven
by infinite dimensional martingales and applications, Stochastics, 81, 6
(2009), 601626.
[3] A. Al-Hussein, Maximum principle for controlled stochastic evolution
equations, Int. Journal of Math. Analysis, 4, 30 (2010), 14471464.
24
AbdulRahman Al-Hussein
25
26
AbdulRahman Al-Hussein