A New Implementable Prediction-Correction Method for Monotone Variational Inequalities with Separable Structure
Abstract
The monotone variational inequalities capture various concrete applications arising in many areas. In this paper, we develop a new prediction-correction method for monotone variational inequalities with separable structure. The new method can be easily implementable, and the main computational effort in each iteration of the method is to evaluate the proximal mappings of the involved operators. At each iteration, the algorithm also allows the involved subvariational inequalities to be solved in parallel. We establish the global convergence of the proposed method. Preliminary numerical results show that the new method can be competitive with Chen′s proximal-based decomposition method in Chen and Teboulle (1994).
1. Introduction
The PADM (11)–(13) is often easy to implement under the assumption that the decomposed subproblems have closed-form solutions or can be efficiently solved up to a high precision. However, in some cases, matrixes A and B are not identity matrices, and the two subproblems in PADM (11)-(12) are difficult to solve because the evaluation of (ATA + (1/β)f) −1(Aυ) and (BTB + (1/β)g) −1(Bυ) could be costly. To overcome this difficulty, we propose a new implementable prediction-correction method for the SVI. At each iteration, we first decompose the problem to two small problems with respect to x and y, respectively. The two subproblems are all easy to solve under the assumption that the resolvent operators of f and g are easy to evaluate, where the resolvent operator of mapping T is defined as (I + λT) −1(υ). Then, we update the Lagrange multipliers and make a correction step to ensure the algorithm′s convergence.
The SVI has been studied extensively both in the theoretical frameworks and applications. Recently, Han [13] proposed a hybrid entropic proximal decomposition method for the SVI. Han′s method is based on logarithmic-quadratic functions and combined with self-adaptive strategy. He [14] presented a parallel splitting augmented Lagrangian method which can be extended to solve the system of equilibrium problems with three separable operators. Xu et al. [15] proposed two classes of correction methods for the SVI in which the mapping F does not have an explicit form. Besides, Xu and Wu [16] also studied a class of linearized proximal alternating direction methods and showed that the relaxation factor can have the same restriction region as for the general ADM. Yuan and Li [17] developed a logarithmic-quadratic-proximal- (LQP-) based decomposition method by applying the LQP terms to regularize the ADM subproblems; then Bnouhachem et al. [18] studied a new inexact LQP alternating direction method by solving a series of related systems of nonlinear equations.
The rest of this paper is organized as follows. In Section 2, we review some preliminaries which are useful for further analysis. In Section 3, we present the new implementable prediction-correction method for SVI, and the global convergence result is established. Numerical experiments and some conclusions are addressed in Sections 4 and 5, respectively.
2. Preliminaries
In this section, we make some standard assumptions and summarize some basic properties of VI which will be used in the subsequent discussions.
- (A1)
𝒳, 𝒴 are simple closed convex sets.
-
A set which is said to be simple means that the projection onto the set is easy to compute, where the projection of a point υ onto the closed convex set Ω, denoted by PΩ(υ), is defined as the nearest point u ∈ Ω to υ; that is,
() - (A2)
The mapping F is point-to-point, monotone, and continuous.
-
A mapping F : ℜn → ℜn is said to be monotone on Ω if
() - (A3)
The solution set of SVI (Ω, F), denoted by Ω*, is nonempty.
Properties. Let G be a symmetric positive definite matrix; the G-norm of the vector u is denoted by . In particular, when G = I, is the Euclidean norm of u. For a matrix A, ∥A∥ denotes its norm ∥A∥ : = max {∥Ax∥ : ∥x∥ ≤ 1}.
The following well-known properties of the projection operator will be used in the coming analysis.
Lemma 1. Let Ω ∈ ℜn be a nonempty closed convex set; let PΩ(·) be the projection operator onto Ω under the G-norm. Then
Lemma 2. u* is a solution of the SVI (Ω, F) if and only if e(u*, β) = 0 for any given positive constant β (see [2, page 267]).
Lemma 3. Solving SVI (Ω, F) (7) is equivalent to find a zero point of the mapping
3. The New Algorithm
Obviously, H is a symmetric positive definite matrix whenever r > 0, s > 0, and β > 0, and we also have Q = HM.
3.1. Description of the Algorithm
Algorithm 4. It is a prediction-correction-based algorithm for the SVI (Ω, F).
Phase 4 (convergence verification). If ∥uk − uk+1∥ ≤ ϵ, stop; otherwise set k : = k + 1; go to Phase 2.
Remark 5. Note that (22) does not involve and that (23) is independent on the generated by (22). Hence the two projections (22) and (23) are eligible for parallel computation.
Remark 6. It is easy to check that is a solution of SVI (Ω, F) if and only if , and . Thus, it is reasonable to take the magnitude of ∥uk − uk+1∥ ≤ ϵ as the stopping criterion.
Remark 7. The strategy of choosing the step size αk in the correction step which coincides with the strategy in He′s papers, see, for example, [19], will be explained in detail in the following section.
3.2. Contractive Properties
Now, we start to prove some properties of the sequence . The first lemma quantifies the discrepancy between the point and a solution point of SVI (Ω, F).
Proof. Note that generated by (22)–(24) are actually solutions of the following VIs:
The assertion (27) is thus proved.
The following lemma plays a key role in proving the convergence of the algorithm.
Lemma 10. Let matrixes Q, H be defined in (20), if the parameters r > 0, s > 0, and β > 0 in (22)–(24) satisfy
Proof. For any , we have
Lemma 11. Suppose that u* = (x*, y*, λ*) ∈ Ω is a solution point of (9) and the sequences {uk+1} are corrected by an undeterminate step size denoted by α instead of (26); that is,
Proof. One can see that
Remark 12. Note that φk(α) is a quadratic function of α and it reaches its maximum at
Now, we prove the Fejér monotonicity of the iterative sequence {uk} generated by the algorithm.
Theorem 13. Suppose that u* = (x*, y*, λ*) ∈ Ω is a solution point of (9) and the sequences {uk} are generated by the algorithm. Then
Proof. According to Lemma 11,
Based on the earlier results, we are now ready to prove the global convergence of the algorithm.
Theorem 14. The sequence {uk} generated by the proposed algorithm converges to a solution of SVI (Ω, F).
Proof. We prove the convergence of the proposed algorithm by following the standard analytic framework of contraction-type methods. It follows from (46) that {uk} is bounded, and we have that
On the other hand, by taking limits over the subsequences in (52) and using , we have that, for any k > kj, it follows from (46) that
4. Numerical Experiments
In this section, we present some numerical experiments results to show the effectiveness of the proposed algorithm. The codes are run on a notebook computer with Inter(R) Core(TM) 2 CPU 2.0 GHZ and RAM 2.00 GM under MATLAB Version 2009b.
- (i)
P and Q were generated randomly with eigenvalues in [5,10] according to the following MATLAB scripts:
-
P = rand (n); [Q1, R1] = qr (P),
-
S = 5 + 5*rand (n, 1); P = Q1*diag (S)*Q1′,
-
Q = rand (p); [Q2, R2] = qr (Q),
-
S = 5 + 5*rand (m, 1); P = Q2*diag (S)*Q2′.
-
- (ii)
A and B were generated randomly with singular values in [0,3], and the maximum singular value is 3 according to the following MATLAB scripts:
-
A = rand (m, n); [U, S, V] = svd (A),
-
S = S/S (1,1)*3; A = U*S*V′,
-
B = rand (m, p); [U, S, V] = svd (B),
-
S = S/S(1,1)*3; B = U*S*V′.
-
- (iii)
b is generated randomly with b = rand (m, 1)*10.
The computational results are given in Table 1 for different choices of m, n, and p. We reported the number of iterations (Iter.) and the computing time in seconds (CPU(s)) when the mentioned stopping criterion is achieved.
m | n | p | PDM | New algorithm | ||||
---|---|---|---|---|---|---|---|---|
Iter. | CPU (s) | Error | Iter. | CPU (s) | Error | |||
10 | 10 | 10 | 237 | 0.075 | 9.956 × 10−5 | 237 | 0.075 | 9.895 × 10−5 |
10 | 15 | 15 | 250 | 0.143 | 9.758 × 10−5 | 250 | 0.086 | 9.815 × 10−5 |
20 | 20 | 20 | 314 | 0.115 | 9.669 × 10−5 | 314 | 0.112 | 9.728 × 10−5 |
20 | 30 | 30 | 372 | 0.175 | 9.586 × 10−5 | 372 | 0.178 | 9.597 × 10−5 |
40 | 50 | 50 | 561 | 3.443 | 9.631 × 10−5 | 561 | 1.340 | 9.625 × 10−5 |
50 | 80 | 80 | 714 | 3.534 | 9.990 × 10−5 | 715 | 1.963 | 9.892 × 10−5 |
60 | 100 | 100 | 842 | 8.107 | 9.982 × 10−5 | 842 | 7.274 | 9.996 × 10−5 |
100 | 120 | 120 | 1065 | 9.773 | 9.926 × 10−5 | 1065 | 11.786 | 9.938 × 10−5 |
150 | 200 | 200 | 1661 | 24.451 | 9.942 × 10−5 | 1661 | 21.366 | 9.947 × 10−5 |
200 | 250 | 250 | 2055 | 38.037 | 9.907 × 10−5 | 2055 | 35.020 | 9.911 × 10−5 |
200 | 300 | 300 | 2445 | 66.520 | 9.964 × 10−5 | 2445 | 61.673 | 9.970 × 10−5 |
The data in Table 1 indicates clearly that the proposed method is efficient compared with the classical PDM in [21, 22]. We can observe that the iteration numbers and the CPU time of the two algorithms are almost the same.
5. Conclusions
In this paper, we proposed a new implementable algorithm for solving the monotone variational inequalities with separable structure. At each iteration, the algorithm performs two easily implementable projections parallelly to produce a predictor and then makes a simple correction to generate the new iterate. Under some mild conditions, we proved the global convergence of the new method. We also give some numerical experiments to show that the proposed method is applicable and valid.