Strong Convergence of a Projected Gradient Method
Abstract
The projected-gradient method is a powerful tool for solving constrained convex optimization problems and has extensively been studied. In the present paper, a projected-gradient method is presented for solving the minimization problem, and the strong convergence analysis of the suggested gradient projection method is given.
1. Introduction
Motivated by the above works, in this paper we will further construct a new projected gradient method for solving the minimization problem (1.1). It should be pointed out that our method also has strong convergence under some mild conditions.
2. Preliminaries
Proposition 2.1 (See [9]). (1) The composite of finitely many averaged mappings is averaged. That is, if each of the mappings is averaged, then so is the composite T1, …, TN. In particular, if T1 is α1-averaged and T2 is α2-averaged, where α1, α2 ∈ (0,1), then the composite T1T2 is α-averaged, where α = α1α2 − α1α2.
(2) T is ν-ism, then for γ > 0, γT is (ν/γ)-ism.
- (i)
xn → x means that xn converges strongly to x;
- (ii)
xn⇀x means that xn converges weakly to x;
- (iii)
Fix (T): = {x : Tx = x} is the fixed points set of T.
Lemma 2.2 (See [26]). Let {xn} and {yn} be bounded sequences in a Banach space X and let {βn} be a sequence in [0,1] with
Lemma 2.3 (See [27] (demiclosedness principle)). Let C be a closed and convex subset of a Hilbert space H and let T : C → C be a nonexpansive mapping with Fix (T) ≠ ∅. If {xn} is a sequence in C weakly converging to x and if {(I − T)xn} converges strongly to y, then
Lemma 2.4 (See [28]). Assume {an} is a sequence of nonnegative real numbers such that
- (1)
;
- (2)
limsup n→∞δn/γn ≤ 0 or .
3. Main Results
Let C be a closed convex subset of a real Hilbert space H. Let f : C → R be a real-valued Frechet differentiable convex function with the gradient ∇f. Let A : C → H be a ρ-contraction. Let B : H → H be a self-adjoint, strongly positive bounded linear operator with coefficient α > 0. First, we present our algorithm for solving (1.1). Throughout, we assume S ≠ ∅.
Algorithm 3.1. For given x0 ∈ C, compute the sequence {xn} iteratively by
Remark 3.2. In (3.1), we use two projections. Now, it is well-known that the advantage of projections, which makes them successful in real-word applications, is computational.
Next, we show the convergence analysis of this Algorithm 3.1.
Theorem 3.3. Assume that the gradient ∇f is L-Lipschitzian and σρ < α. Let {xn} be a sequence generated by (3.1), where γ ∈ (0,2/L) is a constant and the sequence {θn} satisfies the conditions: (i) lim n→∞θn = 0 and (ii) . Then {xn} converges to a minimizer of (1.1) which solves the following variational inequality:
By Algorithm 3.1 involved in the projection, we will use the properties of the metric projection for proving Theorem 3.3. For convenience, we list the properties of the projection as follows.
Proposition 3.4. It is well known that the metric projection PC of H onto C has the following basic properties:
- (i)
∥PC(x) − PC(y)∥≤∥x − y∥, for all x, y ∈ H;
- (ii)
〈x − y, PC(x) − PC(y)〉≥∥PC(x) − PC(y)∥2, for every x, y ∈ H;
- (iii)
〈x − PC(x), y − PC(x)〉≤0, for all x ∈ H, y ∈ C.
The Proof of Theorem 3.3 Let x* ∈ S. First, from (2.8), we note that PC(I − γ∇f)x* = x*. By (3.1), we have
Note that the Lipschitz condition implies that the gradient ∇f is (1/L)-inverse strongly monotone (ism), which then implies that γ∇f is (1/γL)-ism. So, I − γ∇f is (γL/2)-averaged. Now since the projection PC is (1/2)-averaged, we see that PC(I − γ∇f) is ((2 + γL)/4)-averaged. Hence we have that
Indeed, we can choose a subsequence of {xn} such that
In (3.1), if we take A = 0 and B = I, then (3.1) reduces to the following.
Algorithm 3.5. For given x0 ∈ C, compute the sequence {xn} iteratively by
From Theorem 3.3, we have the following result.
Theorem 3.6. Assume that the gradient ∇f is L-Lipschitzian and σρ < α. Let {xn} be a sequence generated by (3.20), where γ ∈ (0, 2/L) is a constant and the sequences {θn} satisfies the conditions: (i) lim n→∞θn = 0 and (ii) . Then {xn} converges to a minimizer of (1.1) which is the minimum norm element in S.
Acknowledgment
Y. Yao was supported in part by Colleges and Universities Science and Technology Development Foundation (20091003) of Tianjin, NSFC 11071279 and NSFC 71161001-G0105.