Infinite Horizon Optimal Control of Stochastic Delay Evolution Equations in Hilbert Spaces
Abstract
The aim of the present paper is to study an infinite horizon optimal control problem in which the controlled state dynamics is governed by a stochastic delay evolution equation in Hilbert spaces. The existence and uniqueness of the optimal control are obtained by means of associated infinite horizon backward stochastic differential equations without assuming the Gâteaux differentiability of the drift coefficient and the diffusion coefficient. An optimal control problem of stochastic delay partial differential equations is also given as an example to illustrate our results.
1. Introduction
The particular form of the control system is essential for our results, but it covers numerous interesting cases. For example, in the particular cases U = H and R(t, x, u) = u, the term u(t)dt + dW(t) in the state equation can be considered as a control affected by noise.
The stochastic optimal control problem was considered in 1977 by Bismut [1]. The optimal control problem for stochastic partial differential equations in the framework of a compact control state space has been studied in [2–5]. Buckdahn and Raşcanu [6] considered an optimal control problem for a semilinear parabolic stochastic differential equation with a nonlinear diffusion coefficient, and the existence of a quasioptimal (nonrelaxed) control is showed without assuming convexity of the coefficients. In [7–11], the authors provided a direct (classical or mild) solution of the Hamilton-Jacobi-Bellman equation for the value function, which is then used to prove that the optimal control is related to the corresponding optimal trajectory by a feedback law. In Gozzi [10, 11], the existence and uniqueness of a mild solution of the associated Hamilton-Jacobi-Bellman equation are proved, when the diffusion term only satisfies weak nondegeneracy conditions. The proofs are based on the corresponding regularity properties of the transition semigroup of the associated Ornstein-Uhlenbeck process.
The main tools for the control problem are techniques from the theory of backward stochastic differential equations (BSDEs) in the sense of Pardoux and Peng, first considered in the nonlinear case in [12]; see [13, 14] as general references. BSDEs have been successfully applied to control problems; see, for example, [15, 16] and we also refer the reader to [17–20]. Fuhrman and Tessiture [19] considered the optimal control problem for stochastic differential equation in the strong form, assuming Lipschitz conditions and allowing degeneracy of the diffusion coefficient, under some structural constraint on the state equation. Existence of an optimal control for stochastic systems in infinite dimensional spaces also has been obtained in [21–27]. In [21], Fuhrman and Tessitore showed the regularity with respect to parameters and the regularity in the Malliavin spaces for the solution of the backward-forward system and defined the feedback law by Malliavin calculus. Finally, the optimal control is obtained by the feedback. Appealing to the Malliavin calculus, compared with Fuhrman et al. [23], the existence of optimal control for stochastic differential equations with delay is proved by the feedback law. Fuhrman and Tessiture [24] dealt with an infinite horizon optimal control problem for the stochastic evolution equation in Hilbert space, and the optimal control is showed by means of infinite horizon backward stochastic differential equation in infinite dimensional spaces and Malliavin calculus. In Masiero [25], the infinite horizon optimal control problem for stochastic evolution equation is also studied by means of the Hamilton-Jacobi-Bellman equation. In Fuhrman [26], a class of optimal control problems governed by stochastic evolution equations in Hilbert spaces which includes state constraints is considered, and the optimal control is obtained by the Fleming logarithmic transformation. Masiero [27] studied stochastic evolution equations evolving in a Banach space where G is a constant and characterized the optimal control via a feedback law by avoiding use of Malliavin calculus. Since there is a lack of regularity of F and G, Malliavin calculus is not available in this case; the method in [27] also cannot be used as G is not a constant, but we can prove a theorem similar to [26, Proposition 3.2], which will be used to define the feedback law.
In the present paper, we study the infinite horizon optimal control problem for stochastic delay evolution equations in Hilbert spaces, and by using Theorem 10, the optimal control is obtained. Since we do not relate the optimal feedback law with the gradient of the value function and do not consider the associated Hamilton-Jacobi-Bellman equation, we can drop the Gâteaux differentiability of the drift term and the diffusion term.
The plan of the paper is as follows. In the next section, some notations are fixed, and the stochastic delay evolution equations are considered with an infinite horizon; in particular, continuous dependence on initial value (t, x) is proved. In Section 3, we give the proof of Theorem 10, which is the key of many subsequent results. The addressed optimal control problem is considered, and the fundamental relation between the optimal control problem and BSDEs is established in Section 4. Section 5 is devoted to proving the existence and uniqueness of optimal control in the weak sense. Finally, an application is given in Section 6.
2. Preliminaries
We list some notations that are used in this paper. We use the symbol |·| to denote the norm in a Banach space F, with a subscript if necessary. Let Ξ, H, and K denote real separable Hilbert spaces, with scalar products (·, ·) Ξ, (·, ·) H, and (·, ·) K, respectively. For fixed τ > 0, 𝒞 = C([−τ, 0], H) denotes the space of continuous functions from [−τ, 0] to H, endowed with the usual norm |f|C = sup θ∈[−τ,0] | f(θ)|H. Let Ξ* denote the dual space of Ξ, with scalar product , and let L(Ξ, H) denote the space of all bounded linear operators from Ξ into H; the subspace of Hilbert-Schmidt operators, with the Hilbert-Schmidt norm, is denoted by L2(Ξ, H).
Let (Ω, ℱ, P) be a complete space with a filtration {ℱt} t≥0 which satisfies the usual condition. By a cylindrical Wiener process with values in a Hilbert space Ξ, defined on (Ω, ℱ, P), we mean a family {W(t), t ≥ 0} of linear mappings Ξ → L2(Ω) such that for every ξ, η ∈ Ξ, {W(t)ξ, t ≥ 0} is a real Wiener process and E(W(t)ξ · W(t)η) = (ξ, η) Ξ. In the following, {W(t), t ≥ 0} is a cylindrical Wiener process adapted to the filtration {ℱt} t≥0.
- (i)
denotes the space of equivalence classes of processes Y ∈ L2(Ω × [t, ∞); F), admitting a predictable version. is endowed with the norm
() - (ii)
, defined for β ∈ R and p, q ∈ [1, ∞), denotes the space of equivalence classes of processes {Y(s), s ≥ t}, with values in F, such that the norm
() -
is finite and Y admits a predictable version.
- (iii)
denotes the space . The norm of an element is |(Y, Z)| = |Y | +|Z|. Here, F is a Hilbert space.
- (iv)
, defined for T > t ≥ 0 and p ∈ [1, ∞), denotes the space of predictable processes {Y(s), s ∈ [t, T]} with continuous paths in F, such that the norm
()
- (v)
, defined for η ∈ R and q ∈ [1, ∞), denotes the space of predictable processes {Y(s), s ≥ t} with continuous paths in F, such that the norm
()is finite. Elements of are identified up to indistinguishability. - (vi)
Finally, for η ∈ R and q ∈ [1, ∞), we defined as the space , endowed with the norm
()
Hypothesis 1. (i) The operator A is the generator of a strongly continuous semigroup {etA, t ≥ 0} of bounded linear operators in the Hilbert space H. We denote by M and ω two constants such that |etA | ≤ Meωt, for t ≥ 0.
(ii) The mapping F: [0, ∞) × 𝒞 → H is measurable and satisfies, for some constant L > 0 and 0 ≤ θ < 1,
(iii) G is a mapping [0, ∞) × 𝒞 → L(Ξ, H) such that for every v ∈ Ξ, the map Gv: [0, ∞) × 𝒞 → H is measurable, esAG(t, x) ∈ L2(Ξ, H) for every s > 0, t ∈ [0, ∞) and x ∈ 𝒞, and
We say that X is a mild solution of (10) if it is a continuous, {ℱt} t≥0-predictable process with values in H, and it satisfies P-a.s.,
We first recall a well-known result on solvability of (10) on bounded interval.
Theorem 1. Assume that Hypothesis 1 holds. Then, for all q ∈ [2, ∞) and T > 0, there exists a unique process as mild solution of (10). Moreover,
for some constant C depending only on q, γ, θ, T, τ, L, ω, and M.
By Theorem 1 and the arbitrariness of T in its statement, the solution is defined for every s ≥ t. We have the following result.
Theorem 2. Assume that Hypothesis 1 holds and the process X(·, t, x) is mild solution of (10) with initial value (t, x)∈[0, ∞) × 𝒞. Then, for every q ∈ [1, ∞), there exists a constant η(q) such that the process . Moreover, for a suitable constant C > 0, one has
Proof. We define a mapping Φ from to by the formula
By the stochastic Fubini theorem,
If are processes belonging to and Y1, Y2 are defined accordingly, the entirely analogous passages show that
We need the following parameter-depending contraction principle, which is stated in the following lemma and proved in [29, Theorems 10.1 and 10.2].
Lemma 3 (Parameter Depending Contraction Principle). Let B, D denote Banach spaces. Let h : B × D → B be a continuous mapping satisfying
Theorem 4. Assume that Hypothesis 1 holds true. Then, for every q ∈ [1, ∞), the map (t, x) → X·(t, x) is continuous from [0, ∞) × 𝒞 to .
Proof. Clearly, it is enough to prove the claim for q large. Let us consider the map Φ defined in the proof of Theorem 2. In our present notation, Φ can be seen as a mapping from to as follows:
Remark 5. By similar passages, we can show that, for fixed t, Theorem 4 still holds true for q large enough if the spaces [0, ∞) × 𝒞 and are replaced by the spaces Lq(Ω, 𝒞, ℱt) and respectively, where Lq(Ω, 𝒞, ℱt) denotes that the space of ℱt-measurable function with value in 𝒞, such that the norm
3. The Backward-Forward System
We make the following assumptions.
Hypothesis 2. The mapping ψ : [0, ∞) × 𝒞 × K × L2(Ξ, K) → K is Borel measurable such that, for all t ∈ [0, ∞), ψ(t, ·) : 𝒞 × K × L2(Ξ, K) → K is continuous, and for some Ly, Lz > 0, μ ∈ R, and m ≥ 1,
We note that the third inequality in (37) follows from the first one taking μ = −Ly but that the third inequality may also hold for different values of μ.
Firstly, we consider the backward stochastic differential equation
Theorem 6. Assume that Hypothesis 2 holds. Let p > 2 and δ < 0 be given, and choose
- (i)
For and , (38) has a unique solution in that will be denoted by (Y(X·)(s), Z(X·)(s)), s ≥ 0.
- (ii)
The estimate
() -
holds for a suitable constant c. In particular, .
- (iii)
The map X· → (Y(X·), Z(X·)) is continuous from to , and X· → Y(X·) is continuous from to .
- (iv)
The statements of points (i), (ii), and (iii) still hold true if the space is replaced by the space .
Proof. The theorem is very similar to Proposition 3.11 in [24]. The only minor difference is that the mapping ψ : [0, ∞) × 𝒞 × K × L2(Ξ, K) → K is a given measurable function, while in [24], the measurable function ψ is from H × K × L2(Ξ, K) to K; however, the same arguments apply.
Theorem 7. Assume that Hypothesis 1 holds and that Hypothesis 2 holds true in the particular case K = R. Then, for every p > 2, q, δ < 0 satisfying (39) with η = η(q), and for every , there exists a unique solution in of (36) that will be denoted by (X(·, t, x), Y(·, t, x), Z(·, t, x)). Moreover, . The map (t, x)→(Y(·, t, x), Z(·, t, x)) is continuous from [0, ∞) × 𝒞 to , and the map (t, x) → Y(·, t, x) is continuous from [0, ∞) × 𝒞 to .
Proof. We first notice that the system is decoupled; the first does not contain the solution (Y, Z) of the second one. Therefore, under the assumption of Hypothesis 1 by Theorem 2, there exists a unique solution X(·, t, x) and of the first equation. Moreover, from Theorem 4, it follows that the map (t, x) → X·(t, x) is continuous from [0, ∞) × 𝒞 to .
Let K = R; from Theorem 6, we have that there exists a unique solution of the second equation, and the map X· → (Y(X·), Z(X·)) is continuous from to while X· → (Y(X·)) is continuous from to . We have proved that is the unique solution of (36), and the other assertions follow from composition.
Remark 8. From Remark 5, by similar passages, we can show that for fixed t and for q large enough, under the assumptions of Theorem 7, the map x → (Y(·, t, x), Z(·, t, x)) is continuous from Lq(Ω, 𝒞, ℱt) to .
We also remark that the process X(·, t, x) is ℱ[t,∞) measurable, since 𝒞 is separable Banach space, we have that X·(t, x) is ℱ[t,∞) measurable; So that Y(t) is measurable with respect to both ℱ[t,∞) and ℱt, it follows that Y(t) is deterministic.
For later use, we notice three useful identities; for t ≤ s < ∞, the equality, P-a.s.,
Lemma 9 (see [30].)Let E be a metric space with metric d, and let f : Ω → E be strongly measurable. Then, there exists a sequence fn, n ∈ N, of simple E-valued functions (i.e., fn is ℱ/ℬ(E) measurable and takes only a finite number of values) such that for arbitrary ω ∈ Ω, the sequence d(fn(ω), f(ω)), n ∈ N, is monotonically decreasing to zero.
We are now in a position of showing the main result in this section.
Theorem 10. Assume that Hypothesis 1 holds true and that Hypothesis 2 holds in the particular case K = R. Then, there exist two Borel measurable deterministic functions υ : [t, ∞) × 𝒞 → R and ζ : [t, ∞) × 𝒞 → Ξ* = L(Ξ, R) = L2(Ξ, R), such that for t ∈ [0, ∞) and x ∈ 𝒞, the solution (X(t, x), Y(t, x), Z(t, x)) of (36) satisfies
Proof. We apply the techniques introduced in [26, Proposition 3.2]. Let {ei} be a basis of Ξ*, and let us define . Then, for every 0 ≤ t1 < t2 < ∞, Δ > 0, and x1, x2 ∈ 𝒞, we have that
We fix x and 0 ≤ t ≤ s < ∞. For l ∈ [s, ∞), we denote , the random variable obtained by composing Xs(t, x) with the map y → E[Zi,N(l, s, y)].
By Lemma 9, there exists a sequence of 𝒞-valued ℱs-measurable simple functions
We define υ(t, x) = Y(t, t, x); since Y(t, t, x) is deterministic, so the map (t, x) → υ(t, x) can be written as a composition υ(t, x) = Γ3(Γ2(t, Γ1(t, x))) with
4. The Fundamental Relation
We are now ready to formulate the assumptions we need.
Hypothesis 3. (i) A, F, and G verify Hypothesis 1.
(ii) (U, 𝒰) is a measurable space. The map g : [0, ∞) × 𝒞 × U → R is continuous and satisfies for suitable constants Kg > 0, mg > 0 and all x ∈ 𝒞,u ∈ U. The map R : [0, ∞) × 𝒞 × U → Ξ is measurable, and |R(t, s, u)| ≤ LR for a suitable constant KR > 0 and all x ∈ 𝒞,u ∈ U, andz ∈ Ξ*.
(iii) The Hamiltonian ψ defined in (60) satisfies the requirements of Hypothesis 2 (with K = R).
(iv) We fix here p > 2, q and δ < 0 satisfying (39) with η = η(q) and such that q > mg.
We are in a position to prove the main result of this section.
Theorem 11. Assume that Hypothesis 3 holds, and suppose that λ verifies
Proof. Consider (58) in the probability space (Ω, ℱ, P) with filtration {ℱt} t≥0 and with an {ℱt} t≥0-cylindrical Wiener process {W(t), t ≥ 0}. Let us define
We immediately deduce the following consequences.
Theorem 12. Let t ∈ [0, ∞) and x ∈ 𝒞 be fixed, assume that the set-valued map Γ has nonempty values and it admits a measurable selection Γ0 : [0, ∞) × 𝒞 × Ξ* → U, and assume that a control u(·) satisfies
Such a control can be shown to exist if there exists a solution for the so-called closed-loop equation as follows:
5. Existence of Optimal Control
We formulate the optimal control problem in the weak sense following the approach of [31]. The main advantage is that we will be able to solve the closed-loop equation in a weak sense, and, hence, to find an optimal control, even if the feedback law is nonsmooth.
We call (Ω, ℱ, {ℱt} t≥0, P, W) an admissible setup, if (Ω, ℱ, {ℱt} t≥0, P) is a filtered probability space satisfying the usual conditions, and W is a cylindrical P-Wiener process with values in Ξ, with respect to the filtration {ℱt} t≥0.
Theorem 13. Assume that Hypothesis 3 holds. Then, there exists a weak solution of the closed-loop equation (82) which is unique in law.
Proof (uniqueness). Let X be a weak solution of (82) in an admissible setup (Ω, ℱ, {ℱt} t≥0, P, W). We define
Proof (existence). Let (Ω, ℱ, P) be a given complete probability space. {W(t), t ≥ 0} is a cylindrical Wiener process on (Ω, ℱ, P) with values in Ξ, and {ℱt} t≥0 is the natural filtration of {W(t), t ≥ 0}, augmented with the family of P-null sets. Let X(·) be the mild solution of
Now, we can state the main result of this section.
Corollary 14. Assume that Hypothesis 3 holds true and λ verifies (62) Also, assume that the set-valued map Γ has nonempty values and it admits a measurable selection Γ0 : [0, ∞) × 𝒞 × Ξ* → U. Then, for every t ∈ [0, ∞) and x ∈ 𝒞 and for all admissible control system (W, u, Xu), one has