Volume 2013, Issue 1 830698
Research Article
Open Access

A Smoothing Method with Appropriate Parameter Control Based on Fischer-Burmeister Function for Second-Order Cone Complementarity Problems

Yasushi Narushima

Corresponding Author

Yasushi Narushima

Department of Management System Science, Yokohama National University, 79-4 Tokiwadai, Hodogaya-ku, Yokohama 240-8501, Japan ynu.ac.jp

Search for more papers by this author
Hideho Ogasawara

Hideho Ogasawara

Department of Mathematical Information Science, Tokyo University of Science, 1-3 Kagurazaka, Shinjuku-ku, Tokyo 162-8601, Japan tus.ac.jp

Search for more papers by this author
Shunsuke Hayashi

Shunsuke Hayashi

Graduate School of Information Sciences, Tohoku University, 6-3-09 Aramaki-aza Aoba, Aoba-ku, Sendai 980-8579, Japan tohoku.ac.jp

Search for more papers by this author
First published: 01 July 2013
Citations: 3
Academic Editor: Jein-Shan Chen

Abstract

We deal with complementarity problems over second-order cones. The complementarity problem is an important class of problems in the real world and involves many optimization problems. The complementarity problem can be reformulated as a nonsmooth system of equations. Based on the smoothed Fischer-Burmeister function, we construct a smoothing Newton method for solving such a nonsmooth system. The proposed method controls a smoothing parameter appropriately. We show the global and quadratic convergence of the method. Finally, some numerical results are given.

1. Introduction

In this paper, we consider the second-order cone complementarity problem (SOCCP) of the following form:
()
where 〈·, ·〉 denotes the Euclidean inner product, ≥ 0, n ≥ 1, F : Rn × Rn × RRn+ is a continuously differentiable function, and 𝒦 denotes the Cartesian product of several second-order cones (SOCs), that is, with m, n1, …, nm ≥ 1, n = n1 + ⋯+nm, and
()

The SOCCP is a wide class of complementarity problems. For example, it involves the mixed complementarity problem (MCP) and the nonlinear complementarity problem (NCP) [1] as subclasses, since when ni = 1 for each i = 1, …, m  (=n). Moreover, the second-order cone programming (SOCP) problem can be reformulated as an SOCCP by using the Karush-Kuhn-Tucker (KKT) conditions. Apart from them, some practical problems in the game theory [2, 3] and the architecture [4] can be reformulated as the SOCCP.

Much theoretical and algorithmic research has been made so far for solving the SOCCP. Fukushima et al. [5] showed that the natural residual function, also called the min function, and the Fischer-Burmeister function for the NCP can be extended to the SOCCP by using the Jordan algebra. They further constructed the smoothing functions for those SOC complementarity (C-) functions and analyzed the properties of their Jacobian matrices. Hayashi et al. [6] proposed a smoothing method based on the natural residual and showed its global and quadratic convergence. On the other hand, Chen et al. [7] proposed another smoothing method with the natural residual in which the smoothing parameter is treated as a variable in contrast to [6]. Moreover, they showed the global and quadratic convergence of their method. Similar to Chen et al., Narushima et al. [8] proposed a smoothing method treating a smoothing parameter as a variable. They used the Fischer-Burmeister function instead of the natural residual function and also provided the global and quadratic convergence of the method.

In the present paper, we propose a smoothing method with the Fischer-Burmeister function for solving SOCCP (1). The main difference from the existing methods is twofold.
  • (i)

    We do not assume the special structure on the function F in SOCCP (1). In [6, 7, 9, 10], the authors focused on the following type of SOCCP:

    ()
    which is a special case of SOCCP (1) with
    ()
    for some continuously differentiable function f : RnRn. In [1113], the authors studied the following type of SOCCP:
    ()
    which is obtained by letting
    ()
    where f and g : RnRn are continuously differentiable functions. However, we assume neither (4) nor (6). Therefore, our method is applicable to a wider class of SOCCPs.

  • (ii)

    In contrast to [8], we do not incorporate the smoothing parameter into the decision variable. We control the smoothing parameter appropriately in each iteration.

This paper is organized as follows. In Section 2, we give some preliminaries, which will be used in the subsequent analysis. In Section 3, we review the SOC C-function. In particular, we recall the property of the (smoothed) Fischer-Burmeister function. In Section 4, we propose an algorithm for solving the SOCCP and discuss its global and local convergence properties. In Section 5, we report some preliminary numerical results.

Throughout the paper, we use the following notations. Let R+ and R++ be the sets of nonnegative and positive reals. For a symmetric matrix A, we write AO (resp., A≻O) if A is positive semidefinite (resp., positive definite). For any x, yRn, we write xy (resp., xy) if xy𝒦n (resp., xy ∈ int 𝒦n), and we denote by 〈x, y〉 the Euclidean inner product, that is, 〈x, y〉: = xy. We use the symbol ∥·∥ to denote the usual 2-norm of a vector or the corresponding induced matrix norm. We often write x = (x1, x2) ∈ R × Rn−1 (possibly x2 vacuous), instead of . In addition, we often regard Rp × Rq as Rp+q. We sometimes divide a vector xRn according to the Cartesian structure of 𝒦, that is, with . For any Fréchet-differentiable function G : RnRm, we denote its transposed Jacobian matrix at xRn by ∇G(x) ∈ Rn×m. For a given set SRn, int S, bd S, and conv S mean the interior, the boundary, and the convex hull of S in Rn, respectively.

2. Some Preliminaries

In this section, we recall some background materials and preliminary results used in the subsequent sections.

First, we review the Jordan algebra associated with SOCs. For any x = (x1, x2) ∈ R × Rn−1 and y = (y1, y2) ∈ R × Rn−1  (n ≥ 1), the Jordan product associated with 𝒦n is defined as
()
When n = 1, that is, the second components x2 and y2 are vacuous, we interpret that the second component in (7) is also vacuous. We will write x2 to mean xx and write x + y to mean the usual componentwise addition of vectors x and y. For the Jordan product, the identity element eRn is defined by e : = (1,0, …, 0) . It is easily seen that xe = ex = x for any xRn. For any x𝒦n, we define x1/2 as
()
Note that (x1/2) 2 = x1/2x1/2 = x and that for n = 1, xy = x1y1, e = 1, and . Although the Jordan product is not associative, associativity holds under the inner product in the sense that
()
In addition, it follows readily from the definition of 𝒦n that 〈x, y〉≥0 (resp., 〈x, y〉>0) for any x, y≽0 (resp., x, y≻0).
For each x = (x1, x2) ∈ R × Rn−1  (n ≥ 1), we define the symmetric matrix Lx by
()
which can be viewed as a linear mapping having the following properties.

Property 1. There holds that

  • (a)

    Lxy = xy = yx = Lyx and Lx+y = Lx + Ly for any x, yRn;

  • (b)

    x≽0⇔LxO, and x≻0⇔LxO;

  • (c)

    Lx is invertible whenever x ∈ int 𝒦n with

    ()

where denotes the determinant of x.

An important character of the Jordan algebra is its spectral factorization. By the spectral factorization associated with SOC, any x = (x1, x2) ∈ R × Rn−1  (n ≥ 1) can be decomposed as
()
where λ1(x), λ2(x) and s{1}, s{2} are the spectral values and the associated spectral vectors of x given by
()
for j = 1,2, with being any vector in Rn−1 satisfying . If x2 ≠ 0, the decomposition (12) is unique. We note again that when n = 1 (viz., x = x1), we have λ1(x) = λ2(x) = x1, s{1} = s{2} = 1/2. The spectral factorization associated with SOC leads to a number of interesting properties, some of which are as follows.

Property 2. For any x = (x1, x2) ∈ R × Rn−1  (n ≥ 1), let λ1(x), λ2(x) and s{1}, s{2} be the spectral values and the associated spectral vectors at x. Then the following statements hold.

  • (a)

    x𝒦nλ1(x) ≥ 0, x ∈ int 𝒦nλ1(x) > 0, x ∈ bd 𝒦nλ1(x) = 0.

  • (b)

    x2 = λ1(x) 2s{1} + λ2(x) 2s{2}𝒦n.

  • (c)

    Let x𝒦n. Then . Moreover, x ∈ int 𝒦nx1/2 ∈ int 𝒦n, x ∈ bd  𝒦nx1/2 ∈ bd 𝒦n.

In what follows, we recall some definitions for functions and matrices. The semismoothness is a generalized concept of the smoothness, which was originally introduced by Mifflin [14] for functionals, and extended to vector-valued functions by Qi and Sun [15]. For vector-valued functions associated with SOC, see also the work of Chen et al. [16]. Now we give the definition of the Clarke subdifferential [17].

Definition 1. Let H : RnRm be a locally Lipschitzian function. The Clarke subdifferential of H at x is defined by

()
where 𝒟H is the set of points at which H is differentiable.

Note that if H is continuously differentiable at x, then H(x) = {∇H(x)}. We next give the definitions of the semismoothness and the strong semismoothness.

Definition 2. A directionally differentiable and locally Lipschitzian function H : RnRm is said to be semismooth at x if

()
for any sufficiently small dRn∖{0} and VH(x + d), where
()
is the directional derivative of H at x along the direction d. In particular, if o(∥d∥) can be replaced by O(∥d2), then function H is said to be strongly semismooth.

It is known that if H is (strongly) semismooth, then
()
holds (see [18], e.g.).

The definitions below for a function can be found in [10, 13, 19].

Definition 3 (see [10], [13].)A function F = (F1, …, Fm) with is said to have

  • (a)

    the Cartesian P0-property, if for any x = (x1, …, xm), with xy, there exists an index ν ∈ {1, …, m} such that xνyν and 〈xνyν, Fν(x) − Fν(y)〉≥0;

  • (b)

    the uniform Cartesian P-property, if there exists a constant ρ > 0 such that, for any x = (x1, …, xm), , there exists an index ν ∈ {1, …, m} such that 〈xνyν, Fν(x) − Fν(y)〉≥ρxy2.

By the definitions, it is clear that the Cartesian P-property implies the Cartesian P0-property. Definition 3 is associated with SOCCP (3), while the following definitions are associated with SOCCP (5).

Definition 4 (see [19].)Let F : RnRn and G : RnRn be functions such that F = (F1, …, Fm), G = (G1, …, Gm) with and . Then, F and G are said to have

  • (a)

    the joint uniform Jordan P-property, if there exists a constant ρ > 0 such that

    ()

  • (b)

    the joint Cartesian weak coerciveness, if there exists a vector such that

    ()

Next we recall the concept of linear growth of a function, which is weaker than the global Lipschitz continuity.

Definition 5 (see [19].)A function F : RnRn is said to have linear growth, if there exists a constant C > 0 such that ∥F(x)∥ ≤ ∥F(0)∥ + Cx∥ for any xRn.

The following definitions for a matrix are originally given in [8], which is a generalization of the mixed P0-property [1].

Definition 6 (see [8].)Let MR(n+)×(2n+) be a matrix partitioned as follows:

()
where M1, M2R(n+n and M3R(n+. Then, M is said to have
  • (a)

    the Cartesian mixed P0-property, if the following statements hold:

    • (1)

      M3 has full column rank;

    • (2)

      for any x = (x1, …, xm), with (x, y) ≠ 0 and pR such that M1x + M2y + M3p = 0, there exists an index ν ∈ {1, …, m} such that (xν, yν) ≠ 0 and 〈xν, yν〉≥0;

  • (b)

    the Cartesian mixed P-property, if (1) of (a) and the following statement hold:

    • (2) for any x = (x1, …, xm), with (x, y) ≠ 0 and pR such that M1x + M2y + M3p = 0, there exists an index ν ∈ {1, …, m} such that 〈xν, yν〉>0.

In the case = 0 (i.e., M3 is vacuous) and M = [M1  I], M has the Cartesian mixed P0 (P)-property if and only if M1 has the Cartesian P0 (P)-property (see [10, 13], e.g.). By the definitions, it is clear that the Cartesian mixed P-property implies the Cartesian mixed P0-property. Moreover, when n1 = ⋯ = nm = 1, the Cartesian mixed P0-property reduces to the mixed P0-property (see [1, page 1013]).

We now introduce the Cartesian mixed Jordan P0 (P)-property.

Definition 7. Let MR(n+)×(2n+) be a matrix partitioned as follows:

()
where M1, M2R(n+n and M3R(n+. Then, M is said to have
  • (a)

    the Cartesian mixed Jordan P0-property, if the following statements hold:

    • (1)

      M3 has full column rank;

    • (2)

      for any x = (x1, …, xm), with (x, y) ≠ 0 and pR such that M1x + M2y + M3p = 0, there exists an index ν ∈ {1, …, m} such that (xν, yν) ≠ 0 and xνyν≽0;

  • (b)

    the Cartesian mixed Jordan P-property, if (1) of (a) and the following statement hold:

    • (2) for any x = (x1, …, xm), with (x, y) ≠ 0 and pR such that M1x + M2y + M3p = 0, there exists an index ν ∈ {1, …, m} such that xνyν≻0.

Note that the relation xνyν≽0 (resp., xνyν≻0) can be rewritten as λ1(xνyν) ≥ 0 (resp., λ1(xνyν) > 0). By the definitions, it is clear that the Cartesian mixed Jordan P-property implies the Cartesian mixed Jordan P0-property, and that the Cartesian mixed Jordan P0 (P)-property implies the Cartesian mixed P0 (P)-property. Similar to the Cartesian mixed P0-property, in the case n1 = ⋯ = nm = 1, the Cartesian mixed Jordan P0-property also reduces to the mixed P0-property.

3. SOC C-Function and Its Smoothing Function

In this section, we introduce the SOC C-function and its smoothing function. In Section 3.1, we give the concept of the SOC C-function to transform the SOCCP into a system of equations. We focus on the Fischer-Burmeister function as an SOC C-function and review some properties of the smoothed Fischer-Burmeister function in Section 3.2.

3.1. SOC C-Function

First, we recall the concept of the SOC C-function.

Definition 8. A function ϕ : Rr × RrRr is said to be an SOC complementarity (C-) function, if the following holds:

()

Let be defined as

()
where x and yRn are divided as x = (x1, …, xm) and y = (y1, …, ym) with xi, , i = 1, …, m, and are SOC C-functions. Then it follows from (22) that
()
Accordingly, SOCCP (1) is reformulated as a system of equations , where is defined by
()
Moreover, we also give a merit function defined by
()
Note that for any (x, y, p) ∈ Rn × Rn × R, and that if and only if (x, y, p) is a solution of SOCCP (1).

There are many kinds of SOC C-functions. The natural residual function and the Fischer-Burmeister function are respectively defined by

()
()
where [z] + denotes the projection of z onto the SOC . Fukushima et al. [5] showed that (22) holds for functions and . Chen et al. [7] and Hayashi et al. [6] proposed methods for solving SOCCP based on the natural residual function (27), whereas Narushima et al. [8] proposed methods for solving SOCCP based on the Fischer-Burmeister function (28).

In what follows, functions ϕFB, HFB, and ΨFB denote , , and with , respectively. Also, functions ϕNR, HNR, and ΨNR denote , , and with , respectively.

Recently, Bi et al. [20] showed the following inequality:

()
We see from (29) that the level-boundedness of ΨFB is equivalent to that of ΨNR.

3.2. Smoothed FB Function and Its Properties

In this section, we consider the smoothing function associated with the Fischer-Burmeister function and give its properties and Jacobian matrix.

Since HFB is not differentiable in general, we cannot apply conventional methods such as Newton’s method or Newton-based methods. We therefore consider the smoothed Fischer-Burmeister function ϕt : R2nRn, which was originally proposed by Kanzow [21] for solving NCP and generalized by Fukushima et al. [5] to SOCCP. Let be defined by
()
for each i = 1, …, m, where t is a smoothing parameter and . Then, the smoothed Fischer-Burmeister function ϕt : R2nRn is defined as
()
Also, the smoothing function Ht : R2n+R2n+ and the merit function Ψt : R2n+R are defined as
()
()
respectively. Clearly, ϕ0(x, y) ≡ ϕFB(x, y), and so H0(x, y, p) ≡ HFB(x, y, p) and Ψ0(x, y, p) ≡ ΨFB(x, y, p). We note that
()
holds for any tR and (see [5] or [22]). From definition (32) of Ht and (34), it follows that
()
for any tR and (x, y, p) ∈ R2n+.
In what follows, we write for any vector . Moreover, for convenience, we use the following notation. For any xi, and any tR, we write and define the functions , by
()
Furthermore, we drop the subscript for t = 0 for simplicity, and thus,
()
Direct calculation yields
()
Note that is actually independent of t, so that hereafter we will write . We also easily get, for j = 1,2,
()
where , and s{1}, s{2} are the spectral values and the associated spectral vectors of , respectively, with if (wi) 2 ≠ 0, and otherwise, being any vector in satisfying .

Now we review some propositions needed to establish convergence properties of the smoothing Newton method. The following proposition gives explicit expression of the transposed Jacobian matrix with t ≠ 0.

Proposition 9 (see [5].)For any t ≠ 0, Ht is continuously differentiable on R2n+, and its transposed Jacobian matrix is given by

()
where denotes the block-diagonal matrix with block elements , and
()
with
()

In order to obtain the Newton step, the nonsingularity of ∇Ht is important. The next proposition establishes the nonsingularity of ∇Ht.

Proposition 10 (see [8].)Let t be an arbitrary nonzero number and let (x, y, p) ∈ R2n+ be an arbitrary triple such that the Jacobian matrix ∇F(x, y, p)  has the Cartesian mixed P0-property at (x, y, p), that is, ∇pF(x, y, p) satisfies

()
and
()
Then, the matrix ∇Ht(x, y, p) given by (40) is nonsingular.

The local Lipschitz continuity and the (strong) semismoothness of HFB play a significant role in establishing locally rapid convergence.

Proposition 11. The function HFB is locally Lipschitzian on R2n+ and, moreover, is semismooth on R2n+. In addition, if ∇F is locally Lipschitzian, then HFB is strongly semismooth on R2n+.

Proof. It follows from [23, Corollary 3.3] that is globally Lipschitzian and strongly semismooth. Since F is a continuously differentiable function, HFB is locally Lipschitzian on R2n+. Also, the (strong) semismoothness of HFB can be easily shown from the strong semismoothness of and the (local Lipschitz) continuity of ∇F.

We define function by . It is easily seen that Θi(xi, yi) = 0 if and only if (xi, yi) = (0,0). Now we partition as , where
()
In order to achieve locally rapid convergence of the method, we need to control the parameter t so that the distance between ∇Ht and HFB is sufficiently small. The following proposition is helpful to control the parameter t appropriately.

Proposition 12 (see [22].)Let (x, y, p) be any point in R2n+. Let θi(xi, yi) be any function such that Θi(xi, yi) ≤ θi(xi, yi). Let δ > 0 be given. Let be defined by

()
where
()
Then, for any tR such that ,
()
where dist (X, S) denotes min {∥XY∥∣YS}.

4. Smoothing Newton Method and Its Convergence Properties

In this section, we first propose an algorithm of the smoothing Newton method based on the Fischer-Burmeister function and its smoothing function. We then prove its global and Q-superlinear (Q-quadratic) convergence.

4.1. Algorithm

We provide the smoothing Newton algorithm based on the Fischer-Burmeister function. In what follows, we write v(k) = (x(k), y(k), p(k)) and z(k) = (x(k), y(k)) for simplicity.

Algorithm 13.   

   Step 0. Choose η, ρ ∈ (0,1), , σ ∈ (0,1/2), r > 1, κ > 0, and .

Choose v(0) = (x(0), y(0), p(0)) ∈ R2n+ and β0 ∈ (0, ). Let t0 : = ∥HFB(v(0))∥. Set k : = 0.

Step 1. If a stopping criterion, such as ∥HFB(v(k))∥ = 0, is satisfied, then stop.

Step 2.

Step 2.0. Set and j : = 0.

Step 2.1. Compute by solving

()

Step 2.2. If , then let and go to Step 3. Otherwise, go to Step 2.3.

Step 2.3. Let lj be the smallest nonnegative integer l satisfying

()

Let and .

Step 2.4. If

()
then let and go to Step 3. Otherwise, set j : = j + 1 and go back to Step 2.1.

Step 3. Update the parameters as follows:

()

Step 4. Set k : = k + 1. Go back to Step 1.

Note that the proposed algorithm consists of the outer iteration steps and the inner iteration steps. Step 2 is the inner iteration steps with the variable and the counter j, while the outer iteration steps have the variable v and the counter k.

From Step 3 of Algorithm 13 and (48), the following inequality holds:

()
Letting v* be a solution of SOCCP (1), we have HFB(v*) = 0. Therefore, from (53) and the local Lipschitz continuity of HFB, the following holds:
()

In the rest of this section, we consider convergence properties of Algorithm 13. In Section 4.2, we prove the global convergence of the algorithm, and in Section 4.3, we investigate local behavior of the algorithm. For this purpose, we make the following assumptions.

Assumption 1.

  • (A1)

    The solution set 𝒮 of SOCCP (1) is nonempty and bounded.

  • (A2)

    The function ΨFB is level-bounded, that is, for any , the level set is bounded.

  • (A3)

    For any t > 0 and vR2n+, ∇Ht(v) is nonsingular.

From Proposition 10, Assumption (A3) holds if ∇F(v)  has the Cartesian mixed P0-property for any vR2n+. The following remarks correspond to SOCCPs (3) and (5).

Remark 14. The case of SOCCP (3). If ∇f(x)  has the Cartesian P0-property, then ∇F(x, y, p)  with F in (4) has the Cartesian mixed P0-property and vice versa. Note that ∇f(x)  has the Cartesian P0-property at any xRn if f has the Cartesian P0-property (see [10, 13] for the definition of the Cartesian P0-property for a matrix).

Remark 15. The case of SOCCP (5). If ∇f(p) is nonsingular and (∇f(p) −1g(p))  has the Cartesian P0-property, then ∇F(x, y, p)  with F in (6) has the Cartesian mixed P0-property.

Note that Assumption (A2) is equivalent to the coerciveness in the sense that
()
We now give some sufficient conditions for Assumption (A2) in the case of SOCCP (3) or (5).

Lemma 16. Consider SOCCP (5). Let be a function such that for any pRn. Then is level-bounded if and only if is level-bounded.

Proof. We use below condition (55) equivalent to the level-boundedness. We first assume that is level-bounded and claim that ΨFB is level-bounded. Suppose the contrary. Then, we can find a sequence {v(k)} = {(x(k), y(k), p(k))} ⊂ R3n such that the sequence {ΨFB(v(k))} is bounded and ∥v(k)∥ → . If {∥p(k)∥} is bounded, then we must have ∥(x(k), y(k))∥ → . Thus, from the inequality

()
and from the continuity of f and g, we have ΨFB(v(k)) → . Since this is not possible, {∥p(k)∥} is unbounded. By taking a subsequence if necessary, we may assume that ∥p(k)∥ → as k. Since ϕFB is globally Lipschitzian by [23, Corollary 3.3], we have from (33) that, for any (x, y, p) ∈ R3n,
()
where L > 0 is a Lipschitz constant. Then, it follows from (57) and the level-boundedness of that lim kΨFB(v(k)) = +, contradicting the boundedness of {ΨFB(v(k))}. This proves the level-boundedness of ΨFB.

We next assume that ΨFB is level-bounded. Let {p(k)} ⊂ Rn be an arbitrary sequence such that ∥p(k)∥ → and let x(k) : = f(p(k)) and y(k) : = g(p(k)). Then we have

()
Thus, from ∥(x(k), y(k), p(k))∥ ≥ ∥p(k)∥ → and the level-boundedness of ΨFB, we have . Therefore, is level-bounded.

Remark 17. Consider SOCCP (3). Let be a function such that for any xRn. Then, in the same way as in Lemma 16, we can show that is level-bounded if and only if is level-bounded (we have only to consider g(p) ≡ p).

We now provide some sufficient conditions under which Assumption (A2) holds.

Proposition 18. Consider SOCCP (5). Assume that f and g have linear growth. Assume further that f and g satisfy one of the following statements:

  • (a)

    f and g have the joint uniform Jordan P-property;

  • (b)

    f and g have the joint Cartesian weak coerciveness.

Then ΨFB is level-bounded.

Proof. It follows from [19] that is level-bounded for each case. Thus from Lemma 16 and (29), we have desired results.

The following condition was given by Pan and Chen [13] to establish the level-boundedness property of the merit function defined in Remark 17.

Condition A. Consider SOCCP (3). For any sequence {ξk} ⊂ Rn   satisfying ∥ξk∥ → with , if there exists an index ν ∈ {1, …, m} such that and {λ1(fν(ξk))} are bounded below, and , then
()
Under Condition A, we have the following proposition, which corresponds to Proposition 5.2 of [13].

Proposition 19. Consider SOCCP (3). Assume that f has the uniform Cartesian P-property and satisfies Condition A. Then ΨFB is level-bounded.

4.2. Global Convergence

In this section, we show the global convergence of Algorithm 13. We first give the well-definedness of the algorithm.

Lemma 20. Suppose that Assumption (A3) holds. Let t be any fixed positive number. Every stationary point of Ψt satisfies .

Proof. For each stationary point of Ψt, holds. Since, from t > 0 and Assumption (A3), is nonsingular, we have , and thus, .

It follows from (35) that
()
for any , and hence we have, from Assumption (A2) and (55), that Ψt is level-bounded for any fixed t > 0. Therefore, there exists at least one stationary point of Ψt. Thus from Lemma 20, the system Ht(v) = 0 has at least one solution, and hence, there exists a point v satisfying in Step 2 at each iteration.

We are now ready to show the well-definedness of the algorithm.

Proposition 21. Suppose that Assumptions (A2) and (A3) hold. Then Algorithm 13 is well-defined.

Proof. To establish the well-definedness of Algorithm 13, we only need to prove the well-definedness and the finite termination property of Step 2 at each iteration. Now we fix k and tk > 0. Since is nonsingular for any by tk > 0 and Assumption (A3), is uniquely determined for any j ≥ 0. In addition, we have

()
If , then Step 2 terminates in Step 2.2. If , then integer l satisfying (50) can be found at Step 2.3, because and . Thus, Step 2 is well-defined at each iteration.

Next we prove the finite termination property of Step 2. To prove by contradiction, we assume that Step 2 never stops and then

()
holds for all j ≥ 0. We consider two cases.
  • (i)

    The case where there exists a subsequence J such that lim jJ,jτj = 0. From the boundedness of the level set of at and the line search rule (50), is bounded. In addition, from the continuous differentiability of and Assumption (A3), is also bounded. Thus, there exists a subsequence JJ such that

    ()
    Now τj ≠ 1 holds for all sufficiently large jJ, and hence, we have
    ()
    Passing to the limit j with jJ on the above inequality and taking (62) into account, we have
    ()
    On the other hand, it follows from (61) that , which contradicts (65).

  • (ii)

    The case where there exists such that for all j. It follows from (50) that

    ()
    which implies holds for sufficiently large j. This contradicts (62). Therefore, the proof is complete.

In order to show the global convergence of the proposed method, we recall the mountain pass theorem (see [24], e.g.), which is as follows.

Lemma 22. Let φ : RnR be a continuously differentiable and level-bounded function. Let 𝒞Rn be a nonempty and compact set and let be the minimum value of φ on bd 𝒞, that is,

()
Assume that there exist vectors ξ𝒞 and η𝒞 such that and . Then, there exists a vector ζRn such that ∇φ(ζ) = 0 and .

By using the mountain pass theorem, we can show the following global convergence property.

Theorem 23. Suppose that Assumptions (A1)–(A3) hold. Then, any accumulation point of the sequence {v(k)} generated by Algorithm 13 is bounded, and hence, at least one accumulation point exists, and any such point is a solution of SOCCP (1).

Proof. From the choices of tk and βk in Step 3 of Algorithm 13, tk and βk converge to zero. Since βk → 0, we have . Thus, it follows from tk → 0 and (35) that lim kHFB(v(k))∥ = 0. It implies from the continuity of HFB that any accumulation point of the sequence {v(k)} is a solution of HFB(v) = 0, and hence, it suffices to prove the boundedness of {v(k)}. To the contrary, we assume {v(k)} is unbounded. Then there exist an index set K and a subsequence {v(k)} kK such that lim kK,k  v(k)∥ = . Since, by Assumption (A1), the solution set 𝒮 is bounded, there exists a compact neighborhood 𝒞 of 𝒮 such that 𝒮 ⊂ int 𝒞. From the boundedness of 𝒞, v(k)𝒞 for all k sufficiently large kK. In addition, from 𝒮 ⊂ int 𝒞, we have

()
for otherwise, there would exist v ∈ bd 𝒞 with ΨFB(v) = 0, that is, v𝒮∩bd 𝒞, which contradicts 𝒮 ⊂ int 𝒞. Since tk is small enough for all sufficiently large k, it follows from (35) that
()
holds for any v𝒞. Now we take . Then (69) yields
()
Letting
()
we have from (68) and that . Therefore, it follows from (69) and (70) that, for all sufficiently large k,
()
On the other hand, since and βk → 0, we get
()
for all k sufficiently large.

Now we choose sufficiently large satisfying all the above arguments with and apply Lemma 22 with

()
Then there exists ζR2n+ satisfying
()
which contradicts Lemma 20, and therefore the proof is complete.

4.3. Local Q-Superlinear and Q-Quadratic Convergence

In Section 4.2, we have shown that the sequence {v(k)} is bounded, and any accumulation point of {v(k)} is a solution of SOCCP (1). In this section, we prove that {v(k)} is superlinearly convergent, or more strongly, quadratically convergent if ∇F is locally Lipschitzian. In order to establish the superlinear (quadratic) convergence of the algorithm, we need an assumption that every accumulation point of is nonsingular. We first consider a sufficient condition for this assumption to hold.

Let {v(k)} and {tk} be the sequences generated by Algorithm 13, and let v* = (x*, y*, p*) be any accumulation point of {v(k)}. Then, by Theorem 23, v* is a solution of SOCCP (1). We call the following condition nondegeneracy of a solution of the SOCCP (see also [13, 25]).

Definition 24. Let v* = (x*, y*, p*) ∈ R2n+ be a solution of SOCCP (1) with x* = (x*1, …, x*m), . Then we say that v* is nondegenerate if x* + y*≻0, or equivalently, x*i + y*i≻0 for all i = 1, …, m.

For a nondegenerate solution, we have the next lemma.

Lemma 25. Let v* = (x*, y*, p*) ∈ R2n+ be a nondegenerate solution of SOCCP (1), and put zi = (xi, yi), for i = 1, …, m. Let tR be a nonzero number. Then, for each i, the following holds:

()
where
()
with
()
Here w*i and u*i are defined by (37) with x*i and y*i. We also write , and set if (w*i) 2 ≠ 0, and otherwise, set to any vector in satisfying .

Proof. Since v* is a solution of SOCCP (1), it follows from [5, Propsition 4.2] that x*i + y*i − ((x*i) 2 + (y*i) 2) 1/2 = 0 for all i = 1, …, m. Hence, from the nondegeneracy of v*, we have

()
By Property 1(c), this implies that is nonsingular. In order to prove this lemma, it suffices to show
()
Since (36) yields , (80) follows from Property 1(c).

The following proposition gives a sufficient condition for the nonsingularity of accumulation points of .

Proposition 26. Suppose that Assumptions (A1)–(A3) hold. Let {v(k)} be a sequence generated by Algorithm 13 and let v* be any accumulation point of it. Moreover, assume that v* is nondegenerate and the Jacobian matrix ∇F(v*)  has the Cartesian mixed Jordan P-property, that is, rank   ∇pF(v*) = and

()
Then, every accumulation point of is nonsingular.

Proof. By Theorem 23, the sequence {v(k)} is bounded and has at least one accumulation point v*. Hence, we may assume that {v(k)} converges to v* without loss of generality. It follows from Lemma 25 and tk → 0 that any accumulation point of , say J0, is given in the following form:

()
In order to prove that J0 is nonsingular, suppose that , where (ξ, η, φ) ∈ R2n+. We will show that (ξ, η, φ) = 0. It follows from (82) that
()
()
where ξ = (ξ1, …, ξm), . Multiplying both sides of the above equations by from the left-hand side, we get
()
for all i = 1, …, m, where the second equality uses the fact u*i = x*i + y*i (see (79)). Suppose on the contrary that (ξ, η) ≠ 0. Then from (83) and the assumption (81), we have that
()
for some ν ∈ {1, …, m}. Since x*ν, y*ν≽0, by Property 1(b), we have and . By using Property 1(a), (85) can be rewritten equivalently as
()
Multiplying both sides of the first equation in (87) by (ξν)  from the left, we have
()
Since is positive semidefinite, we have 〈x*ν, ξνην〉≤0. Similarly, multiplying both sides of the second equation in (87) by (ην)  from the left, we have 〈y*ν, ξνην〉≤0. Adding these two inequalities yields
()
On the other hand, by the nondegeneracy of v*, we have x*ν + y*ν≻0. This together with (86) yields
()
which contradicts (89), and hence, we must have (ξ, η) = 0. Then, since the matrix ∇pF(v*) R(n+ has full column rank, we also have from (83) that φ = 0. Therefore, J0 is nonsingular.

Next, we show the local convergence properties of the sequence {v(k)} generated by Algorithm 13. The following lemma plays a key role in proving such properties.

Lemma 27. Suppose that Assumptions (A1)–(A3) hold. Let {v(k)} be a sequence generated by Algorithm 13 and let v* be any accumulation point of it. In addition, assume that every accumulation point of is nonsingular. Then, for v(k) sufficiently close to v*,

()
holds, where d(k) is defined by . Moreover, if ∇F is locally Lipschitzian and r ≥ 2, then
()
holds.

Proof. From Theorem 23, {v(k)} is bounded, and hence, there exists at least one accumulation point, and any such point v* satisfies HFB(v*) = 0. Since every accumulation point of is nonsingular, we have, from Assumption (A3), that there exists a constant c1 > 0 such that

()
for all k. Let VkHFB(v(k)) such that
()
Note that Vk exists for all k, because HFB(v(k)) is compact [17, page 70]. We have from (93) and HFB(v*) = 0 that
()
It follows from (54) that, for v(k) sufficiently close to v*,
()
From (35), the inequality by Step 3 of Algorithm 13, and the local Lipschitz continuity of HFB, we get
()
Since, by Proposition 11, HFB is semismooth, we have from (15) and (17) that
()
Therefore, from (95)–(97) and r > 1, we obtain (91). Moreover, if ∇F is locally Lipschitzian, then, by Proposition 11, HFB is strongly semismooth, and hence, we have
()
Therefore, from (95)–(97) and r ≥ 2, we obtain (92).

Using Lemma 27, we obtain the local Q-superlinear (Q-quadratic) convergence result.

Theorem 28. Suppose that Assumptions (A1)–(A3) hold. Let {v(k)} be a sequence generated by Algorithm 13. If every accumulation point of is nonsingular, then the following statements hold.

  • (a)

    For all k sufficiently large, is satisfied at j = 0 in Step 2.2 of Algorithm 13. Moreover, for all k sufficiently large, v(k+1) = v(k) + d(k) holds, where .

  • (b)

    The whole sequence {v(k)} converges Q-superlinearly to a solution v* of SOCCP (1). Moreover, if ∇F is locally Lipschitz continuous and r ≥ 2, then the sequence {v(k)} converges Q-quadratically.

Proof. Since (b) is directly obtained from (a) and Lemma 27, it suffices to prove (a). Namely, we prove that for all k sufficiently large. We have from (93) that

()
and hence, it follows from the boundedness of {v(k)} and the continuity of that {d(k)} is bounded. Let v* be any accumulation point of v(k), and let v(k) be sufficiently close to v*. By Theorem 23, v* is a solution of SOCCP (1), and thus, HFB(v*) = 0. From the local Lipschitz continuity of HFB, we may assume that ∥HFB(v(k))∥ is small enough. By Lemma 27 and (100), there exists a constant c2 ∈ (0,1) such that
()
Therefore, we have from (35) with that
()
This together with Lemma 27 and the local Lipschitz continuity of HFB yields that
()
Therefore, it follows again from (35) with that
()
From (35), the choices of tk and βk in Step 3 of Algorithm 13, and , we get
()
where , and hence, (104) yields . This implies that . Taking into account and , we have and v(k+1) = v(k) + d(k). Since, by Lemma 27, v(k+1) remains in the neighborhood of v*, the desired results are obtained.

5. Numerical Experiments

In this section, we show some numerical results for Algorithm 13. The program was coded in MATLAB 7, and computations were carried out on a machine with Intel Core i7-3770 K CPU (3.50 GHz×2) and 8.0 GB RAM. We set the parameters , ρ = 0.66, σ = 0.1, r = 2, κ = 1, , and β0 = 2. We also set the function as follows (see [22] for details):
()
for ni = 1, and
()
for ni ≥ 2. For all problems, we randomly chose the initial point (x(0), y(0), p(0)) ∈ R2n+ whose components were distributed on the interval [0,1], by using rand command of MATLAB. The stopping criterion in Step 1 is relaxed to
()
We first solve the following second-order cone programming (SOCP) problem:
()
which is reformulated as SOCCP (1) with
()
equivalently. We generate one hundred test problems randomly such that there exist primal and dual strictly feasible solutions. Specifically, we first choose matrix AR×n whose components are distributed on [−100,100], vectors and randomly, and then set and . Here, each component of is distributed on [−100,100] and is set to , where α ∈ (0,100] is chosen randomly, and is also generated similarly, while each component of is distributed on [0,1]. In order to compare our method with another method, we solve SOCP (109) by SDPT3 [26, 27], which is the software of interior point methods for solving semidefinite, second-order cone, and linear programming problems. We use SDPT3 with the default parameter and option settings. The obtained results are shown in Table 1, in which “Iter” and “CPU” denote the average values of the number of iterations and the CPU time in seconds, respectively. In particular, the value of “Iter” in “our method” denotes the number of times that the Newton equations (49) have been solved. In the column of 𝒦, (𝒦5) 3 denotes 𝒦5 × 𝒦5 × 𝒦5, for example. Since SDPT3 failed to solve some test problems when 𝒦 = (𝒦5) 3 × (𝒦2) 2 × R+, the average values in this case were taken over the successful trials only. We see from Table 1 that our method is superior to or at least comparable with SDPT3 from the viewpoint of the number of iterations. On the other hand, from the viewpoint of CPU time, our method is also superior to SDPT3 for small-scale problems. However, SDPT3 outperforms our method for middle- or large-scale problems. We believe that this is because SDPT3 is coded to reduce the computational costs by means of some fundamental techniques on matrix computation and so forth. For further development of our method, we will need more appropriate tuning of our code. However, it is not the purpose of this paper.
Table 1. Numerical comparison of our method with SDPT3.
𝒦 n Our method SDPT3
Iter CPU Iter CPU
20 5 8.99 0.063 8.84 0.070
50 10 8.28 0.058 9.82 0.078
400 100 7.02 1.521 10.19 0.291
1000 200 7.01 8.253 14.85 2.512

In order to confirm the local behaviors of the sequence generated by Algorithm 13, we list the value of at each outer iteration k in Table 2. In addition, to investigate how the parameter r affects the rate of convergence, we performed the algorithm with r = 1, 1.5, 2. We also investigate the relation between the choices of r and the behavior of {tk}; we list the behaviors of {tk}. We chose one of the above test problems in the case 𝒦 = 𝒦500 × 𝒦200 × (𝒦100) 3. We note that 2.66e − 09 means 2.66 × 10−9, for example. We see from Table 2 that the sequence generated by Algorithm 13 seems to converge Q-quadratically and the parameter r does not affect the convergence of the sequence. On the other hand, we find that the choices of r affect the behavior of {tk}.

Table 2. Numerical behaviors of .
r = 1 r = 1.5 r = 2
k j tk k j tk k j tk
1 0 1.00e + 00 1.62e + 06 1 0 1.00e + 00 1.62e + 06 1 0 1.00e + 00 1.62e + 06
1 1 1.00e + 00 4.35e + 03 1 1 1.00e + 00 3.80e + 03 1 1 1.00e + 00 3.77e + 03
1 2 1.00e + 00 8.03e + 02 1 2 1.00e + 00 7.38e + 02 1 2 1.00e + 00 7.32e + 02
1 3 1.00e + 00 1.37e + 02 1 3 1.00e + 00 1.27e + 02 1 3 1.00e + 00 1.27e + 02
1 4 1.00e + 00 1.38e + 01 1 4 1.00e + 00 1.30e + 01 1 4 1.00e + 00 1.30e + 01
1 5 1.00e + 00 1.38e + 01 1 5 1.00e + 00 1.30e + 01 1 5 1.00e + 00 1.30e + 01
2 1 1.00e − 01 2.85e − 01 2 1 1.00e − 01 2.54e − 01 2 1 6.57e − 02 2.56e − 01
3 1 1.17e − 04 1.17e − 04 3 1 8.80e − 07 9.19e − 05 3 1 9.71e − 09 9.86e − 05
4 1 2.66e − 09 2.66e − 09 4 1 1.20e − 13 2.43e − 09 4 1 5.39e − 18 2.32e − 09
The next experiment is an application of Algorithm 13 to the robust Nash equilibrium problem in the game theory. The robust Nash equilibrium [2, 3, 28, 29] is a new solution concept for noncooperative games with uncertain information. In this model, it is assumed that each player’s cost (pay-off) function and/or the opponents’ strategies are uncertain, but they belong to some uncertainty sets and each player chooses his strategy by taking the worst possible case into consideration. In other words, each player makes decision according to the robust optimization policy. In this experiment, we focus on the following 2-person robust Nash game with quadratic cost functions:
()
()
where for (i, j)∈{1,2}×{1,2} are given matrices, is the vector of ones of appropriate dimension, and and denote the mixed strategies for Players 1 and 2, respectively. Moreover, δx1 and δx2 mean the estimation error or noise, and each player knows that they belong to the uncertainty sets D1 and D2, respectively. Under this situation, the tuple (x1, x2) is called a robust Nash equilibrium when x1 and x2 solve (111) and (112) simultaneously. In this experiment, we set
()
and change the values of ρ1 and ρ2 variously. Since D1 and D2 are defined by means of Euclidean norm, the robust Nash equilibrium problem can be reformulated as an SOCCP equivalently (the reformulated SOCCP is explicitly written in Section  5.1.1 of [3]. We thus omit the details here). Here, we emphasize that the reformulated SOCCP cannot be expressed as any SOCP, and hence existing software such as SDPT3 cannot be applied. Moreover, if the reformulated SOCCP is rewritten of the form (1), then it satisfies neither (4) nor (6). The obtained results are summarized in Table 3, in which x1 and x2 denote the obtained robust Nash equilibria for various choices of uncertainty radiuses ρ1 and ρ2. For all problems, we could calculate the robust Nash equilibria correctly. Moreover, as is discussed in the existing papers, we can observe that the robust Nash equilibria move smoothly as the values of ρ1 and ρ2 change gradually.
Table 3. Robust Nash equilibria with various choices of uncertainty radiuses.
Iter CPU ρ1 ρ2 x1 x2
8 0.078 0.2 0.2 (0.3916, 0.6083, 0) (0.3247, 0.3625, 0.3128)
8 0.094 0.4 0.4 (0.4168, 0.5832, 0) (0.3354, 0.3152, 0.3495)
7 0.078 0.6 0.6 (0.4292, 0.5708, 0) (0.3477, 0.2821, 0.3702)
7 0.078 0.8 0.8 (0.4017, 0.5065, 0.0918) (0.3686, 0.2503, 0.3812)
9 0.078 1.0 1.0 (0.3627, 0.4589, 0.1784) (0.3891, 0.2258, 0.3851)

6. Conclusion

In this paper, we have proposed a smoothing Newton method with appropriate parameter control based on the Fischer-Burmeister function for solving the SOCCP. We have shown its global and Q-quadratic convergence properties under some assumptions. In addition, we have considered some sufficient conditions for the assumptions. In numerical experiments, we have confirmed the effectiveness of the proposed methods.

Acknowledgments

The authors would like to thank the academic editor Prof. Jein-Shan Chen and the anonymous reviewers who helped them improve the original paper. The first author is supported in part by the Grant-in-Aid for Scientific Research (C) 25330030 and Young Scientists (B) 25870239 of Japan Society for the Promotion of Science. The third author is supported in part by the Grant-in-Aid for Young Scientists (B) 22760064 of Japan Society for the Promotion of Science.

      The full text of this article hosted at iucr.org is unavailable due to technical difficulties.