Volume 2013, Issue 1 791786
Research Article
Open Access

Infinite Horizon Optimal Control of Stochastic Delay Evolution Equations in Hilbert Spaces

Xueping Zhu

Xueping Zhu

School of Astronautics, Northwestern Polytechnical University, Xi′an, Shaanxi 710072, China nwsuaf.edu.cn

Search for more papers by this author
Jianjun Zhou

Corresponding Author

Jianjun Zhou

College of Science, Northwest A&F University, Yangling, Shaanxi 712100, China nwsuaf.edu.cn

Search for more papers by this author
First published: 24 February 2013
Academic Editor: Juan J. Trujillo

Abstract

The aim of the present paper is to study an infinite horizon optimal control problem in which the controlled state dynamics is governed by a stochastic delay evolution equation in Hilbert spaces. The existence and uniqueness of the optimal control are obtained by means of associated infinite horizon backward stochastic differential equations without assuming the Gâteaux differentiability of the drift coefficient and the diffusion coefficient. An optimal control problem of stochastic delay partial differential equations is also given as an example to illustrate our results.

1. Introduction

In this paper, we consider a controlled stochastic evolution equation of the following form:
()
where
()
u is the control process in a measurable space   (U, 𝒰),and W is a cylindrical Wiener process in a Hilbert space Ξ.  A is the generator of a strongly continuous semigroup of bounded linear operator in another Hilbert space H, and the coefficients F and G, defined on [0, ) × C  ([−τ, 0], H), are assumed to satisfy Lipschitz conditions with respect to appropriate norms. We introduce the cost function
()
Here, g is a given real function, λ is large enough, and the control problem is understood in the weak sense. We wish to minimize the cost function over all admissible controls.

The particular form of the control system is essential for our results, but it covers numerous interesting cases. For example, in the particular cases U = H and R(t, x, u) = u, the term u(t)dt + dW(t) in the state equation can be considered as a control affected by noise.

The stochastic optimal control problem was considered in 1977 by Bismut [1]. The optimal control problem for stochastic partial differential equations in the framework of a compact control state space has been studied in [25]. Buckdahn and Raşcanu [6] considered an optimal control problem for a semilinear parabolic stochastic differential equation with a nonlinear diffusion coefficient, and the existence of a quasioptimal (nonrelaxed) control is showed without assuming convexity of the coefficients. In [711], the authors provided a direct (classical or mild) solution of the Hamilton-Jacobi-Bellman equation for the value function, which is then used to prove that the optimal control is related to the corresponding optimal trajectory by a feedback law. In Gozzi [10, 11], the existence and uniqueness of a mild solution of the associated Hamilton-Jacobi-Bellman equation are proved, when the diffusion term only satisfies weak nondegeneracy conditions. The proofs are based on the corresponding regularity properties of the transition semigroup of the associated Ornstein-Uhlenbeck process.

The main tools for the control problem are techniques from the theory of backward stochastic differential equations (BSDEs) in the sense of Pardoux and Peng, first considered in the nonlinear case in [12]; see [13, 14] as general references. BSDEs have been successfully applied to control problems; see, for example, [15, 16] and we also refer the reader to [1720]. Fuhrman and Tessiture [19] considered the optimal control problem for stochastic differential equation in the strong form, assuming Lipschitz conditions and allowing degeneracy of the diffusion coefficient, under some structural constraint on the state equation. Existence of an optimal control for stochastic systems in infinite dimensional spaces also has been obtained in [2127]. In [21], Fuhrman and Tessitore showed the regularity with respect to parameters and the regularity in the Malliavin spaces for the solution of the backward-forward system and defined the feedback law by Malliavin calculus. Finally, the optimal control is obtained by the feedback. Appealing to the Malliavin calculus, compared with Fuhrman et al. [23], the existence of optimal control for stochastic differential equations with delay is proved by the feedback law. Fuhrman and Tessiture [24] dealt with an infinite horizon optimal control problem for the stochastic evolution equation in Hilbert space, and the optimal control is showed by means of infinite horizon backward stochastic differential equation in infinite dimensional spaces and Malliavin calculus. In Masiero [25], the infinite horizon optimal control problem for stochastic evolution equation is also studied by means of the Hamilton-Jacobi-Bellman equation. In Fuhrman [26], a class of optimal control problems governed by stochastic evolution equations in Hilbert spaces which includes state constraints is considered, and the optimal control is obtained by the Fleming logarithmic transformation. Masiero [27] studied stochastic evolution equations evolving in a Banach space where G is a constant and characterized the optimal control via a feedback law by avoiding use of Malliavin calculus. Since there is a lack of regularity of F and G, Malliavin calculus is not available in this case; the method in [27] also cannot be used as G is not a constant, but we can prove a theorem similar to [26, Proposition 3.2], which will be used to define the feedback law.

In the present paper, we study the infinite horizon optimal control problem for stochastic delay evolution equations in Hilbert spaces, and by using Theorem 10, the optimal control is obtained. Since we do not relate the optimal feedback law with the gradient of the value function and do not consider the associated Hamilton-Jacobi-Bellman equation, we can drop the Gâteaux differentiability of the drift term and the diffusion term.

The plan of the paper is as follows. In the next section, some notations are fixed, and the stochastic delay evolution equations are considered with an infinite horizon; in particular, continuous dependence on initial value (t, x) is proved. In Section 3, we give the proof of Theorem 10, which is the key of many subsequent results. The addressed optimal control problem is considered, and the fundamental relation between the optimal control problem and BSDEs is established in Section 4. Section 5 is devoted to proving the existence and uniqueness of optimal control in the weak sense. Finally, an application is given in Section 6.

2. Preliminaries

We list some notations that are used in this paper. We use the symbol |·| to denote the norm in a Banach space F, with a subscript if necessary. Let Ξ, H, and K denote real separable Hilbert spaces, with scalar products (·, ·) Ξ, (·, ·) H, and (·, ·) K, respectively. For fixed τ > 0, 𝒞 = C([−τ, 0], H) denotes the space of continuous functions from [−τ, 0] to H, endowed with the usual norm |f|C = sup θ∈[−τ,0] | f(θ)|H. Let Ξ* denote the dual space of Ξ, with scalar product , and let L(Ξ, H) denote the space of all bounded linear operators from Ξ into H; the subspace of Hilbert-Schmidt operators, with the Hilbert-Schmidt norm, is denoted by L2(Ξ, H).

Let (Ω, , P) be a complete space with a filtration {t} t≥0 which satisfies the usual condition. By a cylindrical Wiener process with values in a Hilbert space Ξ, defined on (Ω, , P), we mean a family {W(t), t ≥ 0} of linear mappings Ξ → L2(Ω) such that for every ξ,   η ∈ Ξ, {W(t)ξ, t ≥ 0} is a real Wiener process and E(W(t)ξ · W(t)η) = (ξ, η) Ξ. In the following, {W(t), t ≥ 0} is a cylindrical Wiener process adapted to the filtration {t} t≥0.

In this section and the next section, {t} t≥0 will denote the natural filtration of W, augmented with the family 𝒩 of P-null of . The filtration {t, t ≥ 0} satisfies the usual conditions. For [a, b],   [a, ) ⊂ [0, ), we also use the following notations:
()
By 𝒫 we denote the predictable σ-algebra, and by (Λ) we denote the Borel σ-algebra of any topological space Λ.
Similar to [24], we define several classes of stochastic processes with values in a Banach space F as follows.
  • (i)

    denotes the space of equivalence classes of processes YL2(Ω × [t, ); F), admitting a predictable version. is endowed with the norm

    ()

  • (ii)

    , defined for βR and p, q ∈ [1, ), denotes the space of equivalence classes of processes {Y(s), st}, with values in F, such that the norm

    ()

  • is finite and Y admits a predictable version.

  • (iii)

    denotes the space . The norm of an element is |(Y, Z)| = |Y | +|Z|. Here, F is a Hilbert space.

  • (iv)

    , defined for T > t ≥ 0 and p ∈ [1, ), denotes the space of predictable processes {Y(s), s ∈ [t, T]} with continuous paths in F, such that the norm

    ()

is finite. Elements of are identified up to indistinguishability.
  • (v)

    , defined for ηR and q ∈ [1, ), denotes the space of predictable processes {Y(s), st} with continuous paths in F, such that the norm

    ()
    is finite. Elements of are identified up to indistinguishability.

  • (vi)

    Finally, for ηR and q ∈ [1, ), we defined as the space , endowed with the norm

    ()

For simplicity, we denote , , , and by , , , and , respectively.
Now, for every fixed t ≥ 0, we consider the following stochastic delay evolution equation:
()
We make the following assumptions.

Hypothesis 1. (i) The operator A is the generator of a strongly continuous semigroup {etA, t ≥ 0} of bounded linear operators in the Hilbert space H. We denote by M and ω two constants such that |etA | ≤ Meωt, for t ≥ 0.

(ii) The mapping F: [0, ) × 𝒞H is measurable and satisfies, for some constant L > 0 and 0 ≤ θ < 1,

()

(iii) G is a mapping [0, ) × 𝒞L(Ξ, H) such that for every v ∈ Ξ, the map Gv: [0, ) × 𝒞H is measurable, esAG(t, x) ∈ L2(Ξ, H) for every s > 0, t ∈ [0, ) and x𝒞, and

()
for some constants L > 0 and γ ∈ [0, 1/2).

We say that X is a mild solution of (10) if it is a continuous, {t} t≥0-predictable process with values in H, and it satisfies P-a.s.,

()
To stress dependence on initial data, we denote the solution by X(s, t, x). Note that X(s, t, x) is [t,s] measurable, hence, independent of t.

We first recall a well-known result on solvability of (10) on bounded interval.

Theorem 1. Assume that Hypothesis 1 holds. Then, for all q ∈ [2, ) and T > 0, there exists a unique process as mild solution of (10). Moreover,

()

for some constant C depending only on q, γ, θ, T, τ, L, ω, and M.

By Theorem 1 and the arbitrariness of  T in its statement, the solution is defined for every st. We have the following result.

Theorem 2. Assume that Hypothesis 1 holds and the process X(·, t, x) is mild solution of (10) with initial value (t, x)∈[0, ) × 𝒞. Then, for every q ∈ [1, ), there exists a constant η(q) such that the process . Moreover, for a suitable constant C > 0, one has

()
with the constant η(q) depending only on q,   γ,   θ,   τ,   L,   ω, and M.

Proof. We define a mapping Φ from to by the formula

()
We are going to show that, provided η is suitably chosen, Φ(·, t, x) is well defined and that it is a contraction in ; that is, there exists c < 1 such that
()
For simplicity, we set t = 0, and we treat only the case F = 0, the general case being handled in a similar way. We will use the so called factorization method; see [28, Theorem 5.2.5]. Let us take q > 1 and α ∈ (0,1) such that .

By the stochastic Fubini theorem,

()
where
()
Since sup τl≤0 | e(s+l)Ax(0)| ≤ Meωs | x|C, the process e(s+·)Ax(0), s ≥ 0, belongs to provided ω + η < 0. Next, we estimate Φ(X·), where
()
setting q = q/(q − 1), so that
()
Applying the Young inequality for convolutions, we have
()
and we conclude that
()
If we start again from (20) and apply the Hölder inequality, we obtain
()
So, we conclude that
()
On the other hand, by the Burkholder-Davis-Gundy inequalities, for some constant cq depending only on q, we have
()
which implies that
()
so that
()
for suitable constants C1,   C2. Applying the Young inequality for convolutions, we obtain
()
This shows that is finite provided we assume that η < 0 and ω + η < 0; so, the map is well defined.

If are processes belonging to and Y1,   Y2 are defined accordingly, the entirely analogous passages show that

()
Recalling the inequalities (23) and (25) and noting that the map Y → Φ(X·) is linear, we obtain an explicit expression for the constant c in (17), and it is immediate to verify that c < 1 provided η < 0 is chosen sufficiently large. We fix such a value of η(q). The first result is a consequence of the contraction principle. The estimate (15) also follows from the contraction property of Φ(·, t, x).

For investigating the dependence of the solution X(s, t, x) on the initial data x and t, we reformulate (13) as an equation on [0, ). We set
()
and we consider the equation
()
Under the assumptions of Hypothesis 1, by Theorem 2, it is easy to prove that equation (32) has a unique solution X and for every q ∈ [2, ). It clearly satisfies X(s) = x((st)∨(−τ)) for s ∈ [−τ, t), and its restriction to the time internal [t, ) is the unique mild solution of (10). From now on, we denote by X(s, t, x),   s ∈ [0, ), the solution of (32).

We need the following parameter-depending contraction principle, which is stated in the following lemma and proved in [29, Theorems 10.1 and 10.2].

Lemma 3 (Parameter Depending Contraction Principle). Let B,   D denote Banach spaces. Let h : B × DB be a continuous mapping satisfying

()
for some α ∈ [0,1) and every x1, x2B,    y ∈ D. Let ϕ(y) denote the unique fixed point of the mapping h(·, y) : BB. Then, ϕ : DB is continuous.

Theorem 4. Assume that Hypothesis 1 holds true. Then, for every q ∈ [1, ), the map (t, x) → X·(t, x) is continuous from [0, ) × 𝒞 to .

Proof. Clearly, it is enough to prove the claim for q large. Let us consider the map Φ defined in the proof of Theorem 2. In our present notation, Φ can be seen as a mapping from to as follows:

()
By the arguments of the proof of Theorem 2, Φ(·, t, x) is a contraction in uniformly with respect to t, x. The process X·(t, x) is the unique fixed point of Φ(·, t, x). So, by the parameter-depending contraction principle (Lemma 3), it suffices to show that Φ is continuous from to . From the contraction property of Φ(·, t, x) mentioned earlier, we have that Φ(·, t, x) is continuous, uniformly in t, x. Moreover, for fixed X·, it is easy to verify that Φ(X·, ·, ·) is continuous from [0, ) × 𝒞 to . The proof is finished.

Remark 5. By similar passages, we can show that, for fixed t, Theorem 4 still holds true for q large enough if the spaces [0, ) × 𝒞 and are replaced by the spaces Lq(Ω, 𝒞, t) and respectively, where Lq(Ω, 𝒞, t) denotes that the space of t-measurable function with value in 𝒞, such that the norm

()
is finite.

3. The Backward-Forward System

In this section, we consider the system of stochastic differential equations, P-a.s.,
()
for s varying on the time interval [t, )⊂[0, ). As in Section 2, we extend the domain of the solution setting X(s, t, x) = x((st)∨(−τ)) for s ∈ [−τ, t).

We make the following assumptions.

Hypothesis 2. The mapping ψ : [0, ) × 𝒞 × K × L2(Ξ, K) → K is Borel measurable such that, for all t ∈ [0, ), ψ(t, ·) : 𝒞 × K × L2(Ξ, K) → K is continuous, and for some Ly, Lz > 0, μR, and m ≥ 1,

()
for every s ∈ [0, ),    x𝒞, y,   y1,  y2K, z,   z1,   and  z2L2(Ξ, K).

We note that the third inequality in (37) follows from the first one taking μ = −Ly but that the third inequality may also hold for different values of μ.

Firstly, we consider the backward stochastic differential equation

()
K is a Hilbert space, the mapping ψ : [0, ) × 𝒞 × K × L2(Ξ, K) → K is a given measurable function, X· is a predictable process with values in another Banach space 𝒞, and λ is a real number.

Theorem 6. Assume that Hypothesis 2 holds. Let p > 2 and δ < 0 be given, and choose

()
Then, the following hold.
  • (i)

    For and , (38) has a unique solution in that will be denoted by (Y(X·)(s), Z(X·)(s)),   s ≥ 0.

  • (ii)

    The estimate

    ()

  • holds for a suitable constant c. In particular, .

  • (iii)

    The map X· → (Y(X·), Z(X·)) is continuous from to , and X·Y(X·) is continuous from to .

  • (iv)

    The statements of points (i), (ii), and (iii) still hold true if the space is replaced by the space .

Proof. The theorem is very similar to Proposition 3.11 in [24]. The only minor difference is that the mapping ψ : [0, ) × 𝒞 × K × L2(Ξ, K) → K is a given measurable function, while in [24], the measurable function ψ is from H × K × L2(Ξ, K) to K; however, the same arguments apply.

Theorem 7. Assume that Hypothesis 1 holds and that Hypothesis 2 holds true in the particular case K = R. Then, for every p > 2,   q,   δ < 0 satisfying (39) with η = η(q), and for every , there exists a unique solution in of (36) that will be denoted by (X(·, t, x), Y(·, t, x), Z(·, t, x)). Moreover, . The map (t, x)→(Y(·, t, x), Z(·, t, x)) is continuous from [0, ) × 𝒞 to , and the map (t, x) → Y(·, t, x) is continuous from [0, ) × 𝒞 to .

Proof. We first notice that the system is decoupled; the first does not contain the solution (Y, Z) of the second one. Therefore, under the assumption of Hypothesis 1 by Theorem 2, there exists a unique solution X(·, t, x) and of the first equation. Moreover, from Theorem 4, it follows that the map (t, x) → X·(t, x) is continuous from [0, ) × 𝒞 to .

Let K = R; from Theorem 6, we have that there exists a unique solution of the second equation, and the map X· → (Y(X·), Z(X·)) is continuous from to while X· → (Y(X·)) is continuous from to . We have proved that is the unique solution of (36), and the other assertions follow from composition.

Remark 8. From Remark 5, by similar passages, we can show that for fixed t and for q large enough, under the assumptions of Theorem 7, the map x → (Y(·, t, x), Z(·, t, x)) is continuous from Lq(Ω, 𝒞, t) to .

We also remark that the process X(·, t, x) is [t,) measurable, since 𝒞 is separable Banach space, we have that X·(t, x) is [t,) measurable; So that Y(t) is measurable with respect to both [t,) and t, it follows that Y(t) is deterministic.

For later use, we notice three useful identities; for ts < , the equality, P-a.s.,

()
is a consequence of the uniqueness of the solution of (13). Since the solution of the backward equation is uniquely determined on an interval [s, ) by the values of the process X· on the same interval, for ts < , we have, P-a.s.,
()

Lemma 9 (see [30].)Let E be a metric space with metric d, and let f : Ω → E be strongly measurable. Then, there exists a sequence fn,    nN, of simple E-valued functions (i.e., fn is /(E) measurable and takes only a finite number of values) such that for arbitrary ω ∈ Ω, the sequence d(fn(ω), f(ω)),    nN, is monotonically decreasing to zero.

Let now fLq(Ω, 𝒞). By Lemma 9 we get the existence of a sequence of simple function fn,  nN, such that
()
Hence, fnf in by Lebesgue′s dominated convergence theorem.

We are now in a position of showing the main result in this section.

Theorem 10. Assume that Hypothesis 1 holds true and that Hypothesis 2 holds in the particular case K = R. Then, there exist two Borel measurable deterministic functions υ : [t, ) × 𝒞R and ζ : [t, ) × 𝒞 → Ξ* = L(Ξ, R) = L2(Ξ, R), such that for t ∈ [0, ) and x𝒞, the solution (X(t, x), Y(t, x), Z(t, x)) of (36) satisfies

()

Proof. We apply the techniques introduced in [26, Proposition 3.2]. Let {ei} be a basis of Ξ*, and let us define . Then, for every 0 ≤ t1 < t2 < ,    Δ > 0, and x1, x2𝒞, we have that

()
From Theorem 7, we have that the map is continuous from [0, ) × 𝒞 to R. By Remark 8, we also have that, for fixed t, the map is continuous from Lq(Ω, 𝒞, t) to R for q large enough. Let us define
()
It is clear that ζi,N : [0, ) × 𝒞R is a Borel function.

We fix x and 0 ≤ ts < . For l ∈ [s, ), we denote , the random variable obtained by composing Xs(t, x) with the map yE[Zi,N(l, s, y)].

By Lemma 9, there exists a sequence of 𝒞-valued s-measurable simple functions

()
where are pairwise distinct and , such that
()
For any Bs, we have
()
and we get that
()
Fix t and x. Recalling that |Zi,N | ≤ N, by the Lebesgue theorem on differentiation, it follows that P-a.s.
()
By the boundedness of Zi,N, applying the dominated convergence theorem, we get that
()
Now, we have proved that for every t, x,
()
for every i, N. Let C ⊂ [0, ) × 𝒞 denote the set of pairs (t, x) such that lim Nζi,N(t, x) exists and the series converges in Ξ*. We define
()
Since Z satisfies
()
for every ω,   s,   t,   x. From (53), it follows that for every t, x, we have (s, Xs(ω, t, x)) ∈ C, P-a.s., for almost all s ∈ [t, ), and Z(s, t, x) = ζ(s, Xs(t, x))  P-a.s., for a.a. s ∈ [t, ).

We define υ(t, x) = Y(t, t, x); since Y(t, t, x) is deterministic, so the map (t, x) → υ(t, x) can be written as a composition υ(t, x) = Γ32(t, Γ1(t, x))) with

()
From Theorem 7, it follows that Γ1 is continuous. By
()
we have that Γ2 is continuous. It is clear that Γ3 is continuous. Then, the map (t, x) → υ(t, x) is continuous from [0, ) × 𝒞 to R; therefore, υ(t, x)is a Borel measurable function. From uniqueness of the solution of (36), it follows that Y(s, t, x) = υ(s, Xs(t, x)), P-a.s., for a.a. s ∈ [t, ).

4. The Fundamental Relation

Let (Ω, , P) be a given complete probability space with a filtration {t} t≥0 satisfying the usual conditions. {W(t), t ≥ 0} is a cylindrical Wiener process in Ξ with respect to {t} t≥0. We will say that an {} t≥0-predictable process u with values in a given measurable space (U, 𝒰) is an admissible control. The function R :  [0, ) × 𝒞 × U → Ξ is measurable and bounded. We consider the following controlled state equation:
()
Here, we assume that there exists a mild solution of (58) which will be denoted by Xu(s, t, x) or simply by Xu(s). We consider a cost function of the form:
()
Here, g is function on [0, ) × 𝒞 × U with real values. Our purpose is to minimize the function J over all admissible controls.
We define in a classical way the Hamiltonian function relative to the previous problem; for all t ∈ [0, ),   x𝒞, and  z ∈ Ξ*,
()
and the corresponding, possibly empty, set of minimizers
()

We are now ready to formulate the assumptions we need.

Hypothesis 3. (i) A, F, and G verify Hypothesis 1.

(ii) (U, 𝒰) is a measurable space. The map g : [0, ) × 𝒞 × UR is continuous and satisfies for suitable constants Kg > 0, mg > 0 and all x𝒞,uU. The map R : [0, ) × 𝒞 × U → Ξ is measurable, and |R(t, s, u)| ≤ LR for a suitable constant KR > 0 and all x𝒞,uU, andz ∈ Ξ*.

(iii) The Hamiltonian ψ defined in (60) satisfies the requirements of Hypothesis 2 (with K = R).

(iv) We fix here p > 2, q and δ < 0 satisfying (39) with η = η(q) and such that q > mg.

We are in a position to prove the main result of this section.

Theorem 11. Assume that Hypothesis 3 holds, and suppose that λ verifies

()
Let υ,  ζ denote the function in the statement of Theorem 10. Then, for every admissible control u and for the corresponding trajectory X starting at (t, x), one has
()

Proof. Consider (58) in the probability space (Ω, , P) with filtration {t} t≥0 and with an {t} t≥0-cylindrical Wiener process {W(t),   t ≥ 0}. Let us define

()
Let Pu be the unique probability on [0,) such that
()
We notice that under Pu, the process Wu is a Wiener process. Let us denote by the filtration generated by Wu, and completed in the usual way. Relatively to Wu (58) can be rewritten as
()
In the space , we consider the following system of forward-backward equations:
()
Applying the Itô formula to eλsYu(s) and writing the backward equation in (67) with respect to the process W, we get
()
Recalling that R is bounded, we get, for all r ≥ 1 and some constant C,
()
It follows that
()
We conclude that the stochastic integral in (68) has zero expectation. If we set s = t in (68) and we take expectation with respect to P, we obtain
()
By Theorem 7, , so that
()
By the Hölder inequality, we have that for suitable constant C > 0,
()
From Theorem 2, we obtain ; by the similar process, we get that
()
for suitable constant C > 0 and
()
Since Yu(t, t, x) = υ(t, x) and -a.s., for a.a. s ∈ [t, ), we have that
()
Thus, adding and subtracting and letting T, we conclude that
()
The proof is finished.

We immediately deduce the following consequences.

Theorem 12. Let t ∈ [0, ) and x𝒞 be fixed, assume that the set-valued map Γ has nonempty values and it admits a measurable selection Γ0 : [0, ) × 𝒞 × Ξ*U, and assume that a control u(·) satisfies

()
Then, J(t, x, u) = υ(t, x), and the pair (u(·), X) is optimal for the control problem starting from x at time t.

Such a control can be shown to exist if there exists a solution for the so-called closed-loop equation as follows:

()
since in this case, we can define an optimal control setting
()
However, under the present assumptions, we cannot guarantee that the closed-loop equation has a solution in the mild sense. To circumvent this difficulty, we will revert to a weak formulation of the optimal control problem.

5. Existence of Optimal Control

We formulate the optimal control problem in the weak sense following the approach of [31]. The main advantage is that we will be able to solve the closed-loop equation in a weak sense, and, hence, to find an optimal control, even if the feedback law is nonsmooth.

We call (Ω, , {t} t≥0, P, W) an admissible setup, if (Ω, , {t} t≥0, P) is a filtered probability space satisfying the usual conditions, and W is a cylindrical P-Wiener process with values in Ξ, with respect to the filtration {t} t≥0.

By an admissible control system, we mean (Ω, , {t} t≥0, P, W, u, Xu), where (Ω, , {t} t≥0, P,   W) is an admissible setup, u is an t-predictable process with values in U, and Xu is a mild solution of (58). An admissible control system will be briefly denoted by (W, u, Xu) in the following. Our purpose is to minimize the cost functional
()
over all the admissible control system.
Our main result in this section is based on the solvability of the closed-loop equation
()
In the following sense, we say that X is a weak solution of (82) if there exists an admissible setup (Ω, , {t} t≥0, P, W) and an t-adapted continuous process X(t) with values in H, which solves the equation in the mild sense; namely, P-a.s.,
()
()

Theorem 13. Assume that Hypothesis 3 holds. Then, there exists a weak solution of the closed-loop equation (82) which is unique in law.

Proof (uniqueness). Let X be a weak solution of (82) in an admissible setup (Ω, , {t} t≥0,   P, W). We define

()
Since R is bounded, the Girsanov theorem ensures that there exists a probability measure P0 such that the process
()
is a P0-Wiener process and
()
Let us denote by the filtration generated by W0 and completed in the usual way. In , X is a mild solution of
()
By Hypothesis 3, the joint law of X and W0 is uniquely determined by A,   F,   G, and x. Taking into account the last displayed formula, we conclude that the joint law of X and ρ(T) under P0 is also uniquely determined, and consequently so is the law of X under P. This completes the proof of the uniqueness part.

Proof (existence). Let (Ω, , P) be a given complete probability space. {W(t), t ≥ 0} is a cylindrical Wiener process on (Ω, , P) with values in Ξ, and {t} t≥0 is the natural filtration of {W(t), t ≥ 0}, augmented with the family of P-null sets. Let X(·) be the mild solution of

()
and by the Girsanov theorem, let P1 be the probability on Ω under which
()
is a Wiener process (notice that R is bounded). Then, X is the weak solution of (82) relatively to the probability P1 and the Wiener process W1.

Now, we can state the main result of this section.

Corollary 14. Assume that Hypothesis 3 holds true and λ verifies (62) Also, assume that the set-valued map Γ has nonempty values and it admits a measurable selection Γ0 : [0, ) × 𝒞 × Ξ*U. Then, for every t ∈ [0, ) and x ∈ 𝒞 and for all admissible control system (W, u, Xu), one has

()
and the equality holds if
()

Moreover, from Theorem 13, it follows that the closed-loop equation (82) admits a weak solution (Ω, , {t} t≥0, P, W, X) which is unique in law, and setting
()
we obtain an optimal admissible control system (W, u, X).

6. Applications

In this section, we present a simple application of the previous results. We consider the stochastic delay partial differential equation in the bounded domain BRn with smooth boundary B as follows:
()
Here, W = (W1, W2, …, Wd) is a standard Wiener process in Rd, and the functions f : [0, +) × C([−1,0], R) → R and gi : [0, +) × C([−1,0], R) → R are Lipschitz continuous and bounded. Setting U as a bounded subset of Rd, Ξ = Rd, H = L2(B), and xC([−1,0], H). We define F and G as following:
()
and let A denote the Laplace operator Δ in L2(B); with domain then, (94) has the form (58) and Hypothesis 1 holds.
Let us consider the optimal control problem associated with the cost
()
where λ verifies (62) and σ : C([−1,0], R) × U → [0, ) is a bounded measurable function. Define g : C([−1,0], H) × U → [0, ) and R : C([−1,0], H) × U → Ξ by g(y, u) = ∫Bσ(t, y(ξ), u)dξ + u2 and R(y, u) = (∫Br1(ξ)u1dξ, ∫Br2(ξ)u2dξ, …, ∫Brd(ξ)uddξ) for yC([−1,0], H), u = (u1, u2, …, ud) ∈ U, respectively. It can be easily verified that Hypothesis 3 holds true, and the set-valued map Γ has nonempty values and it admits a measurable selection Γ0 : [0, ) × 𝒞 × Ξ*U. Then, the closed-loop equation (82) admits a weak solution (Ω, , {t} t≥0, P, W, u, z·(·)), and setting
()
we obtain an optimal admissible control system (W, u, z(·)).

    The full text of this article hosted at iucr.org is unavailable due to technical difficulties.