Volume 3, Issue 6 e1192
SPECIAL ISSUE PAPER
Full Access

An efficient family of Steffensen-type methods with memory for solving systems of nonlinear equations

Mona Narang

Mona Narang

D.A.V. College, Chandigarh, India

Search for more papers by this author
Saurabh Bhatia

Saurabh Bhatia

University Institute of Engineering and Technology, Panjab University, Chandigarh, India

Search for more papers by this author
Vinay Kanwar

Corresponding Author

Vinay Kanwar

University Institute of Engineering and Technology, Panjab University, Chandigarh, India

Correspondence Vinay Kanwar, University Institute of Engineering and Technology, Panjab University, Chandigarh 160014, India.

Email: [email protected]

Search for more papers by this author
First published: 22 September 2021
Citations: 2

Abstract

The article discusses derivative-free algorithms with and without memory for solving numerically nonlinear systems. We proposed a family of fifth- and sixth-order schemes and extended them to algorithms with memory. We further discuss the convergence and computational efficiency of these algorithms. Numerical examples of mixed Hammerstein integral equation, discrete nonlinear ordinary differential equation, and Fisher's partial differential equation with Neumann's boundary conditions are discussed to demonstrate the convergence and efficiency of these schemes. Finally, some numerical results are included to examine the performance of the developed methods.

1 INTRODUCTION

Applications in science and engineering which lead to nonlinear systems justify the importance of finding solutions of F ( x ) = 0 in the domain of computational mathematics, where F : n n is sufficiently many times differentiable. The Newton's method1-6 is possibly one of the oldest schemes available in literature for determination of roots of nonlinear systems. The method is given by
x ( j + 1 ) = x ( j ) F ( x ( j ) ) 1 F ( x ( j ) ) . ()
Here, F ( x ) is first-order Fr é chet derivative of the function F ( x ) and [ F ( x ) ] 1 represents its inverse. This algorithm converges quadratically under the assumptions that function F should be continuously differentiable and the initial approximation considered to find this root is close enough to the solution. However, to handle the situations where F ( x ) either does not exist or cannot be evaluated easily, Traub6 proposed quadratically convergent derivative-free scheme given by
x ( j + 1 ) = x ( j ) [ t ( j ) , x ( j ) ; F ] 1 F ( x ( j ) ) . ()
In this case, t ( j ) , x ( j ) ; F 1 is the inverse of first-order divided difference t ( j ) , x ( j ) ; F of the function F and t ( j ) = x ( j ) + η F ( x ( j ) ) , where η 0 represents an arbitrary constant. Here, we recall that the first-order divided difference operator of a multivariable function F is a map [ . , . ; F ] : S × S n × n L ( n ) which is defined3, 7, 8 as given below:
[ θ 1 , θ 2 ; F ] ( θ 1 θ 2 ) = F ( θ 1 ) F ( θ 2 ) ( θ 1 , θ 2 ) n × n , θ 1 θ 2 . ()
Furthermore, in order to obtain the Taylor series expansion of the divided difference operator, we use the Genocchi–Hermite formula7, 9-11 given by
[ θ + h , θ ; F ] = 0 1 F ( θ + g h ) d g , ( θ , h ) n × n , ()
where h = θ 2 θ 1 . The divided difference operator [ . , . ; F ] is a square matrix of order n and the definitions (3) and (4) are equivalent.7 In this article, for computational purposes we use the following definition proposed by Grau et al.12 of first-order divided difference operator [ x , y ; F ] , namely,
[ x , y ; F ] i k = f i ( x 1 , , x k 1 , x k , y k + 1 , , y n ) f i ( x 1 , . . . , x k 1 , y k , y k + 1 , . . . , y n ) ( x k y k ) , 1 i , k n . ()

This is a linear operator and satisfies (3). For η = 1 , algorithm in (2) is the famous Steffensen method.13

Different approaches have been used by researchers to construct computationally efficient root-finding algorithms for both scalar and systems of nonlinear equations. Several iterative methods have been constructed using compositional approach, quadrature approach, adomian decomposition approach, weight function approach to name a few.

Several higher order derivative-free schemes have been proposed in literature. Liu et al.14 and Zheng et al.15 suggested fourth-order derivative-free methods. Grau et al.16 suggested fourth- and sixth-order methods and Sharma and Gupta17 proposed a family of sixth-order derivative-free methods. A derivative-free family of seventh-order methods was proposed by Wang and Zhang.18 Recently, Qiu et al.19 proposed a hybrid genetic algorithm by combining the advantages of genetic algorithms and derivative-free iterative methods to solve a class of complex nonlinear functions with discontinuity, nondifferentiability and multiple roots. This suggests derivative-free algorithms can be extended to the area of soft computing.

Recently, a technique used to achieve higher efficiency, is to develop memory based methods. This concept is given by Traub,6 who introduced algorithms with and without memory for the first time. Though this area has been explored more for scalar nonlinear equations, but very little progress has been made for nonlinear systems. Some of the existing memory based derivative-free methods can be found in works of Ren and Argyros,20 Chicharro et al.,21 and Ahmad et al.22 Here, we propose new multipoint schemes which are not only derivative-free but also with and without memory for solving numerically nonlinear systems.

2 DEVELOPMENT AND CONVERGENCE INVESTIGATION

In this section, we propose new derivative-free schemes using divided differences as follows:
y ( j ) = x ( j ) L j 1 F ( x ( j ) ) , z ( j ) = y ( j ) L j 1 F ( y ( j ) ) , x ( j + 1 ) = z ( j ) ( p I + L j 1 M j q I + r L j 1 M j ) L j 1 F ( z ( j ) ) . ()

Here, p , q , and r are arbitrary constants, I denotes an identity matrix of order n, L j = [ u ( j ) , v ( j ) ; F ] , where u ( j ) = x ( j ) a F ( x ( j ) ) , v ( j ) = x ( j ) + b F ( x ( j ) ) , and M j = [ w ( j ) , s ( j ) ; F ] , where w ( j ) = z ( j ) c F ( z ( j ) ) and s ( j ) = z ( j ) + d F ( z ( j ) ) . Here, a , b , c and d are arbitrary nonzero parameters such that either a 0 or b 0 and either c 0 or d 0 and L j 1 represents the inverse of L j .

To discuss the convergence of proposed schemes, we recall the following Taylor series expansion of vector functions.3

Lemma 1.Let F: S n n be t times Fr é chet differentiable function in a set S n , which is an open convex set. Then, for any x , h n , the following expression holds:

F ( x + h ) = F ( x ) + F ( x ) h + 1 2 ! F ( x ) h 2 + 1 3 ! F ( x ) h 3 + + 1 ( t 1 ) ! F ( t 1 ) ( x ) h t 1 + R t , where R t 1 t ! sup 0 δ 1 F ( t ) ( x + δ h ) h t and h t = ( h , h , . . t . . , h ) .

For a sufficiently continuously differentiable function F on S, the Taylor series expansion of F ( x + t h ) about the point x is substituted in (4) and integration yields
[ x + h , x ; F ] = 0 1 F ( x + t h ) d t = F ( x ) + 1 2 F ( x ) h + 1 6 F ( x ) h 2 + O ( h 3 ) , ()
where h k = ( h , h , . . k . . , h ) , h n .
Let the error in the jth approximation of a solution “ α ” of a nonlinear system F ( x ) = 0 be e ( j ) = x ( j ) α . Under the assumption that Γ = [ F ( α ) ] 1 exists, we develop F ( x ( j ) ) in a neighborhood of “ α ” to obtain
F ( x ( j ) ) = F ( α ) [ e ( j ) + A 2 ( e ( j ) ) 2 + A 3 ( e ( j ) ) 3 + A 4 ( e ( j ) ) 4 + A 5 ( e ( j ) ) 5 + O ( ( e ( j ) ) 6 ) ] , ()
where A k = 1 k ! Γ F ( k ) ( α ) L k ( n , n ) and L k ( n , n ) represents the set of bounded k-linear functions, k = 2 , 3 , . . . , and ( e ( j ) ) k = ( e ( j ) , e ( j ) , . . k . , e ( j ) ) , e ( j ) n and Γ = [ F ( α ) ] 1 L ( n , n ) . With help of (8), derivatives of function F in a neighborhood of “ α ” can be written as follows:23
F ( x ( j ) ) = F ( α ) [ I + 2 A 2 e ( j ) + 3 A 3 ( e ( j ) ) 2 + 4 A 4 ( e ( j ) ) 3 + 5 A 5 ( e ( j ) ) 4 + O ( ( e ( j ) ) 5 ) ] . ()
F ( x ( j ) ) = F ( α ) [ 2 A 2 + 6 A 3 e ( j ) + 12 A 4 ( e ( j ) ) 2 + 20 A 5 ( e ( j ) ) 3 + O ( ( e ( j ) ) 4 ) ] . ()
F ( x ( j ) ) = F ( α ) [ 6 A 3 + 24 A 4 e ( j ) + 60 A 5 ( e ( j ) ) 2 + O ( ( e ( j ) ) 3 ) ] . ()

We now discuss the convergence of the algorithm (6) in the following theorem.

Theorem 1.Let F : S n n be sufficiently Fr é chet differentiable in set S, which is an open convex set and the solution “ α ” of F lies in S. Assume that F ( x ) is continuous and nonsingular at x= α and x ( 0 ) is an initial approximation sufficiently close to “ α .” Then, for all a , b , c , d , such that either a 0 or b 0 and either c 0 or d 0 , the sequence of approximations { x ( j ) } j 1 generated by scheme (6) has local order of convergence five for q = 3 2 p , r = p 2 , p 3 and six for p = 3 , q = 3 , r = 1 .

Proof.Let e u ( j ) = u ( j ) α = e ( j ) a F ( x ( j ) ) and e v ( j ) = v ( j ) α = e ( j ) + b F ( x ( j ) ) .

Using F ( x ( j ) ) from (8), one obtains

e u ( j ) = I a F ( α ) e ( j ) a F ( α ) ( A 2 ( e ( j ) ) 2 + A 3 ( e ( j ) ) 3 + A 4 ( e ( j ) ) 4 + O ( ( e ( j ) ) 5 ) ) , ()
and
e v ( j ) = I + b F ( α ) e ( j ) + b F ( α ) ( A 2 ( e ( j ) ) 2 + A 3 ( e ( j ) ) 3 + A 4 ( e ( j ) ) 4 + O ( ( e ( j ) ) 5 ) ) . ()

Let x + h = u ( j ) , x = v ( j ) and h = e u ( j ) e v ( j ) in (7) and with help of (9)–(11), using the results in Reference 12, one gets

L j = [ u ( j ) , v ( j ) ; F ] = F ( α ) I + A 2 e u ( j ) + e v ( j ) + A 3 ( e u ( j ) ) 2 + e u ( j ) e v ( j ) + ( e v ( j ) ) 2 + A 4 ( e u ( j ) ) 3 + ( e u ( j ) ) 2 e v ( j ) + e u ( j ) ( e v ( j ) ) 2 + ( e v ( j ) ) 3 + O ( ( e ( j ) ) 5 ) , ()
and
L j 1 = [ u ( j ) , v ( j ) ; F ] 1 = [ I A 2 ( e u ( j ) + e v ( j ) ) + ( A 2 2 A 3 ) ( ( e u ( j ) ) 2 + ( e v ( j ) ) 2 ) + ( 2 A 2 2 A 3 ) e u ( j ) e v ( j ) + O ( ( e ( j ) ) 5 ) ] Γ . ()

On premultiplying (8) by (15) and simplifying with help of (12) and (13), one obtains

L j 1 F ( x ( j ) ) = e ( j ) A 2 [ I + ( b a ) F ( α ) ] ( e ( j ) ) 2 + O ( ( e ( j ) ) 3 ) . ()

Using (16) in the first step of algorithm (6), one gets on simplification

e y ( j ) = y ( j ) α = A 2 ( I + ( b a ) F ( α ) ) ( e ( j ) ) 2 + O ( ( e ( j ) ) 3 ) . ()

Taylor series expansion of F ( y ( j ) ) about “ α ” gives

F ( y ( j ) ) = F ( α ) [ e y ( j ) + A 2 ( e y ( j ) ) 2 + A 3 ( e y ( j ) ) 3 + O ( ( e y ( j ) ) 4 ) ] . ()

Premultiplying (18) by (15) and substituting in second step of scheme (6), one obtains

e z ( j ) = z ( j ) α = A 2 ( e u ( j ) + e v ( j ) ) e y ( j ) A 2 ( e y ( j ) ) 2 ( A 2 2 A 3 ) ( ( e u ( j ) ) 2 + ( e v ( j ) ) 2 ) e y ( j ) ( 2 A 2 2 A 3 ) e u ( j ) e v ( j ) e y ( j ) + O ( ( e ( j ) ) 5 ) . ()

Now, the Taylor series of F ( z ( j ) ) about “ α ” gives

F ( z ( j ) ) = F ( α ) ( e z ( j ) + A 2 ( e z ( j ) ) 2 + A 3 ( e z ( j ) ) 3 + O ( ( e z ( j ) ) 4 ) ) . ()

Let e w ( j ) = w ( j ) α = e z ( j ) c F ( z ( j ) ) , and e s ( j ) = s ( j ) α = e z ( j ) + d F ( z ( j ) ) . Using (20), one gets

e w ( j ) = ( I c F ( α ) ) e z ( j ) c F ( α ) A 2 ( e z ( j ) ) 2 + A 3 ( e z ( j ) ) 3 + O ( ( e z ( j ) ) 4 ) , ()
and
e s ( j ) = ( I + d F ( α ) ) e z ( j ) + d F ( α ) ( A 2 ( e z ( j ) ) 2 + A 3 ( e z ( j ) ) 3 + O ( ( e z ( j ) ) 4 ) ) . ()

Using x + h = w ( j ) , x = s ( j ) and h = e w ( j ) e s ( j ) in (7), and (9)–(11), one gets

M j = [ w ( j ) , s ( j ) ; F ] = F ( α ) I + A 2 ( e w ( j ) + e s ( j ) ) + A 3 ( e w ( j ) ) 2 + e w ( j ) e s ( j ) + ( e s ( j ) ) 2 + O ( ( e ( j ) ) 7 ) . ()

Using (15), (20), and (23) in the third step of scheme (6), one can write

e ( j + 1 ) = ( 1 p q r ) e z ( j ) + ( p + 2 q + 3 r ) A 2 ( e u ( j ) + e v ( j ) ) e z ( j ) ( ( p + 3 q + 6 r ) A 2 2 ( p + 2 q + 3 r ) A 3 ) ( ( e u ( j ) ) 2 + ( e v ( j ) ) 2 ) e z ( j ) ( 2 ( p + 3 q + 6 r ) A 2 2 ( p + 2 q + 3 r ) A 3 ) e u ( j ) e v ( j ) e z ( j ) + ( ( p + 4 q + 10 r ) A 2 3 ( p + 3 q + 6 r ) ( A 2 A 3 + A 3 A 2 ) + ( p + 2 q + 3 r ) A 4 ) ( ( e u ( j ) ) 3 + ( e v ( j ) ) 3 ) e z ( j ) + ( 3 ( p + 4 q + 10 r ) A 2 3 2 ( p + 3 q + 6 r ) ( A 2 A 3 + A 3 A 2 ) + ( p + 2 q + 3 r ) A 4 ) e u ( j ) e v ( j ) ( e u ( j ) + e v ( j ) ) e z ( j ) ( p + q + r ) A 2 ( e z ( j ) ) 2 ( q + 2 r ) A 2 ( e w ( j ) + e s ( j ) ) e z ( j ) + O ( ( e ( j ) ) 8 ) . ()

In order to attain higher order of convergence of the proposed scheme (6), the choice of parameters p , q , and r should be such that the coefficients 1 p q r , p + 2 q + 3 r are zero. Solving these equations simultaneously, results in q = 3 2 p , r = p 2 . The values of q and r are substituted in the error equation (24), give

e ( j + 1 ) = ( ( p 3 ) A 2 2 ) ( ( e u ( j ) ) 2 + ( e v ( j ) ) 2 ) e z ( j ) 2 ( p 3 ) A 2 2 e u ( j ) e v ( j ) e z ( j ) + A 2 ( e w ( j ) + e s ( j ) ) e z ( j ) + [ ( p 3 ) ( A 2 A 3 + A 3 A 2 ) + ( 3 p 8 ) A 2 3 ) ] ( ( e u ( j ) ) 3 + ( e v ( j ) ) 3 ) e z ( j ) + [ 2 ( p 3 ) ( A 2 A 3 + A 3 A 2 ) + 2 ( 3 p 8 ) A 2 3 ) ] ( ( e u ( j ) ) 2 e v ( j ) + e u ( j ) ( e v ( j ) ) 2 ) e z ( j ) + O ( ( e ( j ) ) 7 ) . ()

It can be seen easily that the convergence order of the proposed algorithm is five for p 3 and six for p = 3 . For p 3 , the error equation (25) can be simplified with the help of (12), (13), (17), and (19) and one gets

e ( j + 1 ) = ( p 3 ) A 2 2 ( e u ( j ) + e v ( j ) ) 2 e z ( j ) + O ( ( e ( j ) ) 6 ) , = ( p 3 ) A 2 3 ( 2 I + ( b a ) F ( α ) ) 3 ( I + ( b a ) F ( α ) ) ( e ( j ) ) 5 + O ( ( e ( j ) ) 6 ) . ()

Using p = 3 , the error equation using (12), (13), (17), (19), (21), and (22) will become

e ( j + 1 ) = A 2 ( e w ( j ) + e s ( j ) ) e z ( j ) + A 2 3 ( ( e u ( j ) ) 3 + ( e v ( j ) ) 3 ) e z ( j ) + 2 A 2 3 ( ( e u ( j ) ) 2 e v ( j ) + e u ( j ) ( e v ( j ) ) 2 ) e z ( j ) A 2 ( e z ( j ) ) 2 + O ( ( e ( j ) ) 7 ) . = A 2 ( e z ( j ) ) 2 + A 2 ( e w ( j ) + e s ( j ) ) e z ( j ) + A 2 3 ( e u ( j ) + e v ( j ) ) 3 e z ( j ) + O ( ( e ( j ) ) 7 ) . = A 2 5 ( ( I + ( d c ) F ( α ) ) ( I + ( b a ) F ( α ) ) + ( 2 I + ( b a ) F ( α ) ) 2 ) ( 2 I + ( b a ) F ( α ) ) 2 ( I + ( b a ) F ( α ) ) ( e ( j ) ) 6 + O ( ( e ( j ) ) 7 ) . ()

Remark 1.The proposed algorithms require evaluation of two divided differences, at most seven F and inversion of one matrix only per iteration.

3 METHODS WITH MEMORY

In literature, several techniques are used to increment the order of convergence. One such methodology is to convert a method without memory into a method with memory. As the name suggests, this is done with the assistance of current and preceding data stored in memory.

Now, if we are allowed to choose b a = [ F ( α ) ] 1 or b a = 2 [ F ( α ) ] 1 in the error equations (26) and (27), one can observe that the order of convergence of scheme becomes six and seven, respectively. This acceleration requires the knowledge of “ α ” in advance, but in absence of this information, it is not viable to attain this acceleration. But, this acceleration can be achieved partially by considering an appropriate approximation of [ F ( α ) ] 1 based on the current and previous acquired information.

Hence, we consider the following value of varying parametric matrix P ( j ) :
b a = P ( j ) = [ u ( j 1 ) , v ( j 1 ) ; F ] 1 [ F ( α ) ] 1 , j = 1 , 2 , 3 , ()
and define the iterative scheme as follows:
y ( j ) = x ( j ) L j 1 F ( x ( j ) ) , z ( j ) = y ( j ) L j 1 F ( y ( j ) ) , x ( j + 1 ) = z ( j ) ( p I + L j 1 M j ( 3 2 p ) I + ( p 2 ) L j 1 M j ) L j 1 F ( z ( j ) ) . ()

Here, L j = [ u ( j ) , v ( j ) ; F ] , v ( j ) = x ( j ) + δ P ( j ) F ( x ( j ) ) , and u ( j ) = x ( j ) γ P ( j ) F ( x ( j ) ) , where δ and γ are arbitrary constants which satisfy the condition δ γ = 1 , and P ( j ) is given by (28). Note that the value of P ( j ) requires values that are based on preceding steps and hence (29) is a scheme with memory. In the undermentioned theorem, we establish the R-order of convergence of the scheme (29).

Theorem 2.Let F: S n n be sufficiently Fr é chet differentiable function in a set S n which is convex and contains the zero “ α ” of F ( x ) . Assume F ( x ) is continuous and nonsingular at “ α ” and the parameter matrix P ( j ) of the algorithm is recursively calculated using (28). Then, the R-order of convergence the scheme (29), for p 3 and b a = P ( j ) is 5.19, and for p 3 and b a = 2 P ( j ) is 5.54 where c , d , such that either c 0 or d 0 . Further, for p = 3 and b a = P ( j ) algorithm converges with R-order of convergence 6.16, and for b a = 2 P ( j ) , the R-order of convergence is 6.31 where c , d , such that either c 0 or d 0 . Furthermore, for p = 3 , b a = 2 P ( j ) and d c = C ( j ) , the R-order of convergence is 6.46, and for b a = 2 P ( j ) and d c = C ( j + 1 ) , the R-order of convergence is 6.60.

Proof.Assume that the scheme (29) generates a sequence of approximations { x ( j ) } which approaches the zero “ α ” of F and its R-order is σ , thus one can write

e ( j + 1 ) Λ ( j , σ ) ( e ( j ) ) σ . ()

Here, the sequence { Λ ( j , σ ) } approaches asymptotic error constant Λ σ of scheme when j and ã b ˜ symbolize that ã and b ˜ have magnitudes of same order. Thus, one can write

e ( j + 1 ) Λ ( j , σ ) ( Λ ( j 1 , σ ) ( e ( j 1 ) ) σ ) σ = Λ ( j , σ ) ( Λ ( j 1 , σ ) ) σ ( e ( j 1 ) ) σ 2 . ()

Replacing j by j 1 in (15), one obtains

b a = P ( j ) = [ u ( j 1 ) , v ( j 1 ) ; F ] 1 = ( I A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) ) Γ . ()

Similarly, on replacing j by j 1 in (12) and (13), one gets

I + P ( j ) F ( α ) = A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) e ( j 1 ) . ()

Using the value of b a = P ( j ) , and rewriting the error equation (26), one can write

e ( j + 1 ) = ( p 3 ) A 2 3 ( 2 I + P ( j ) F ( α ) ) 3 ( I + P ( j ) F ( α ) ) ( e ( j ) ) 5 + O ( ( e ( j ) ) 6 ) . ()

Using (33) in (34) and letting j , with the help of (31), one can write

e ( j + 1 ) e ( j 1 ) ( e ( j ) ) 5 e ( j 1 ) ( ( e ( j 1 ) ) σ ) 5 ( e ( j 1 ) ) 5 σ + 1 . ()

From (31) and (35), the exponents of e ( j 1 ) are compared and one gets the quadratic equation

σ 2 = 5 σ + 1 . ()

Solving (36), one gets the positive root as 5.19, which gives an estimate of the lower bound of R-order of convergence of the algorithm (29) for all p , c , d such that either c 0 or d 0 and p 3 .

Next, if the matrix P ( j ) = 2 [ u ( j 1 ) , v ( j 1 ) ; F ] 1 then, replacing j by j-1 in (15), one obtains

b a = P ( j ) = 2 [ I A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) ] Γ . ()

Substituting j 1 in (12) and (13), one obtains

2 I + P ( j ) F ( α ) = 2 A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) e ( j 1 ) . ()

Using (34) in (38), as explained earlier, one can write

e ( j + 1 ) ( e ( j 1 ) ) 3 ( e ( j ) ) 5 ( e ( j 1 ) ) 3 ( ( e ( j 1 ) ) σ ) 5 ( e ( j 1 ) ) 5 σ + 3 . ()

On comparing exponents of e ( j 1 ) using (31) and (39), one gets

σ 2 = 5 σ + 3 . ()

The positive root of this quadratic equation (40) is 5.54. This gives the lower bound of the R-order of convergence of scheme (29) for all p , c , d such that either c 0 or d 0 and p 3 .

Next, to explore the convergence behavior of the algorithm (29) for p = 3 , we use (33) in the error equation (27) to obtain

e ( j + 1 ) = A 2 5 ( ( I + ( d c ) F ( α ) ) ( I + P ( j ) F ( α ) ) + ( 2 I + P ( j ) F ( α ) ) 2 ) ( 2 I + P ( j ) F ( α ) ) 2 ( I + P ( j ) F ( α ) ) ( e ( j ) ) 6 + O ( ( e ( j ) ) 7 ) . ()

From (33) and (41), one can write

e ( j + 1 ) e ( j 1 ) ( e ( j ) ) 6 e ( j 1 ) ( ( e ( j 1 ) ) σ ) 6 ( e ( j 1 ) ) 6 σ + 1 . ()

Comparison of exponents of e ( j 1 ) in (31) and (42), one can write the quadratic equation

σ 2 = 6 σ + 1 . ()

A positive solution of (43) is 6.16, which gives the lower bound of the R-order of convergence of the iterative scheme (29) for all c , d such that either c 0 or d 0 .

Next, on substituting (38) in (41), we can write

e ( j + 1 ) ( e ( j 1 ) ) 2 ( e ( j ) ) 6 ( e ( j 1 ) ) 2 ( ( e ( j 1 ) ) σ ) 6 ( e ( j 1 ) ) 6 σ + 2 . ()

From (31) and (44), we compare the exponents of e ( j 1 ) . This yields the quadratic equation

σ 2 = 6 σ + 2 . ()

The positive root of the equation is 6.31, which is the lower bound of the R-order of convergence of the scheme (29) for all c , d such that either c 0 or d 0 .

Now, for d c = C ( j ) = [ u ( j 1 ) , v ( j 1 ) ; F ] 1 and using (15) for j 1 , one can write

d c = C ( j ) = [ I A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) ] Γ . ()

Using (12) and (13) for j 1 , one gets

I + C ( j ) F ( α ) = A 2 ( e u ( j 1 ) + e v ( j 1 ) ) + O ( ( e ( j 1 ) ) 2 ) e ( j 1 ) . ()

Using (46), the error equation (41) can be written as

e ( j + 1 ) = A 2 5 ( ( I + C ( j ) F ( α ) ) ( I + P ( j ) F ( α ) ) + ( 2 I + P ( j ) F ( α ) ) 2 ) ( 2 I + P ( j ) F ( α ) ) 2 ( I + P ( j ) F ( α ) ) ( e ( j ) ) 6 + O ( ( e ( j ) ) 7 ) . ()

Using (38) and (47) in (48), one can write

e ( j + 1 ) ( e ( j 1 ) ( I + P ( j ) F ( α ) ) + ( e ( j 1 ) ) 2 ) ( e ( j 1 ) ) 2 ( e ( j ) ) 6 , ( e ( j 1 ) ) 3 ( ( e ( j 1 ) ) σ ) 6 ( e ( j 1 ) ) 6 σ + 3 . ()

On comparing exponents of e ( j 1 ) in (31) and (49), one gets the quadratic equation

σ 2 = 6 σ + 3 . ()

Solving Equation (50), one gets the positive root as 6.46. This provides a lower bound of the R-order of convergence of the scheme (29).

Further, we consider d c = C ( j + 1 ) = [ u ( j ) , v ( j ) ; F ] 1 and using (15), one can write

d c = C ( j + 1 ) = [ I A 2 ( e u ( j ) + e v ( j ) ) + O ( ( e ( j ) ) 2 ) ] Γ . ()

Simplifying (51) and using (12) and (13), one gets

I + C ( j + 1 ) F ( α ) = A 2 ( e u ( j ) + e v ( j ) ) + O ( ( e ( j ) ) 2 ) e ( j ) . ()

The error equation (41) with the help of (51) can be simplified as

e ( j + 1 ) = A 2 5 ( ( I + C ( j + 1 ) F ( α ) ) ( I + P ( j ) F ( α ) ) + ( 2 I + P ( j ) F ( α ) ) 2 ) ( 2 I + P ( j ) F ( α ) ) 2 ( I + P ( j ) F ( α ) ) ( e ( j ) ) 6 + O ( ( e ( j ) ) 7 ) . ()

Using (38) and (52) in (53), one can write

e ( j + 1 ) ( e ( j ) ( I + P ( j ) F ( α ) ) + ( e ( j 1 ) ) 2 ) ( e ( j 1 ) ) 2 ( e ( j ) ) 6 , ( ( e ( j 1 ) ) σ ( I + P ( j ) F ( α ) ) + ( e ( j 1 ) ) 2 ) ( e ( j 1 ) ) 2 ( e ( j ) ) 6 , ( e ( j 1 ) ) 4 ( ( e ( j 1 ) ) σ ) 6 ( assuming σ 2 ) ( e ( j 1 ) ) 6 σ + 4 . ()

Comparing the exponents of e ( j 1 ) from (31) and (54), yields the quadratic equation

σ 2 = 6 σ + 4 . ()

The positive root of (55) is 6.60, which estimates lower bound of R-order of convergence of the scheme (29).

3.1 Particular cases of family (29)

Here, let the ith scheme of order ρ be represented by Ω ρ , i .

Case 1. For p = 2 in (29), we get the scheme as given below:
y ( j ) = x ( j ) L j 1 F ( x ( j ) ) , z ( j ) = y ( j ) L j 1 F ( y ( j ) ) , x ( j + 1 ) = z ( j ) ( 2 I L j 1 M j ) L j 1 F ( z ( j ) ) , ()
where L j = [ u ( j ) , v ( j ) ; F ] and M j = [ w ( j ) , s ( j ) ; F ] .

For v ( j ) = x ( j ) + 2 δ 1 P ( j ) F ( x ( j ) ) and u ( j ) = x ( j ) 2 γ 1 P ( j ) F ( x ( j ) ) , where γ 1 and δ 1 are arbitrary constants such that δ 1 γ 1 = 1 , and P ( j ) is given by (28), w ( j ) = z ( j ) c F ( z ( j ) ) , s ( j ) = z ( j ) + d F ( z ( j ) ) , where c and d are arbitrary constants such that either c 0 or d 0 , the method is referred as Ω 5 . 54 , i . For ( γ 1 = 0 , δ 1 = 1 , c = 0 , d 0 ) , we denote the method by Ω 5 . 54 , 1 .

Case 2. For p = 3 in (29), we obtain the undermentioned algorithm:
y ( j ) = x ( j ) L j 1 F ( x ( j ) ) , z ( j ) = y ( j ) L j 1 F ( y ( j ) ) , x ( j + 1 ) = z ( j ) ( 3 I + L j 1 M j ( 3 I + L j 1 M j ) ) L j 1 F ( z ( j ) ) , ()
where L j = [ u ( j ) , v ( j ) ; F ] and M j = [ w ( j ) , s ( j ) ; F ] .

Subcase (i) For u ( j ) = x ( j ) γ 1 P ( j ) F ( x ( j ) ) and v ( j ) = x ( j ) + δ 1 P ( j ) F ( x ( j ) ) , where γ 1 and δ 1 are arbitrary constants that satisfy the condition δ 1 γ 1 = 1 , and P ( j ) is given by (28), w ( j ) = z ( j ) c F ( z ( j ) ) , s ( j ) = z ( j ) + d F ( z ( j ) ) , where c and d are arbitrary constants such that either c 0 or d 0 , the method is referred as Ω 6 . 16 , i . For ( γ 1 = 0 , δ 1 = 1 , c = 0 , d 0 ) , the method is addressed by Ω 6 . 16 , 1 .

Subcase (ii) For u ( j ) = x ( j ) 2 γ 1 P ( j ) F ( x ( j ) ) , v ( j ) = x ( j ) + 2 δ 1 P ( j ) F ( x ( j ) ) , w ( j ) = z ( j ) γ 2 C ( j ) F ( z ( j ) ) and s ( j ) = z ( j ) + δ 2 C ( j ) F ( z ( j ) ) , where γ 1 , δ 1 , γ 2 and δ 2 are arbitrary constants satisfying the condition δ 1 γ 1 = 1 and δ 2 γ 2 = 1 , and P ( j ) and C ( j ) are given by (28), the method is denoted by Ω 6 . 46 , i . For ( γ 1 = 0 , δ 1 = 1 , γ 2 = 0 , δ 2 = 1 ) , the method is referred as Ω 6 . 46 , 1 .

Subcase (iii) For u ( j ) = x ( j ) 2 γ 1 P ( j ) F ( x ( j ) ) , v ( j ) = x ( j ) + 2 δ 1 P ( j ) F ( x ( j ) ) , w ( j ) = z ( j ) γ 2 C ( j + 1 ) F ( z ( j ) ) and s ( j ) = z ( j ) + δ 2 C ( j + 1 ) F ( z ( j ) ) where γ 1 , δ 1 , γ 2 , and δ 2 are arbitrary constants which satisfy the condition δ 1 γ 1 = 1 and δ 2 γ 2 = 1 , and P ( j ) and C ( j + 1 ) are given by (28), the method is referred as Ω 6 . 60 , i . For ( γ 1 = 0 , δ 1 = 1 , γ 2 = 0 , δ 2 = 1 ) , we address the method by Ω 6 . 60 , 1 .

4 COMPUTATIONAL EFFICIENCY

Among various parameters that decide the quality of an algorithm, one parameter is its computational efficiency. It is discussed using efficiency index.4 The computational efficiency index is defined by E = ρ 1 C , where ρ is the convergence order and C denotes the computational cost per step. In case of a nonlinear system of n equations in n unknowns, the computational cost12 per step has been modified and is calculated using

C ( μ , n , l ) = A ( n ) μ + P ( n , l ) .

Here, A ( n ) is count of scalar functions while evaluating F and divided difference [ x , y ; F ] , and P ( n , l ) counts number of products and quotients requisite in each step. To evaluate C ( μ , n , l ) , we require the ratios μ > 0 and l > 0 . Here, μ represents the ratio of number of products used for evaluation of the function F and number of scalar function in F, whereas l 1 refers to the ratio of the time taken by one quotient and the time taken by one product.10

We recall that we calculate n scalar functions for F and n ( n 1 ) for divided difference [ x , y ; F ] of F (refer (5)), and F ( x ) and F ( y ) are computed additionally. Also, we add n 2 quotients for each new divided difference, n 2 products are used in case of multiplication of a matrix and vector, whereas n products for multiplication of a vector and scalar.

Another aspect that contributes toward the computational cost is the calculation of number of products and quotients utilized for the evaluation of an inverse operator. Rather than calculating an inverse operator, we solve a linear system. The linear system uses ( n ) ( n 1 ) ( 2 n 1 ) 6 products and n ( n 1 ) 2 quotients for LU decomposition, and further the resolution of two triangular linear systems requires n ( n 1 ) products and n quotients.

The computational efficiency of the proposed methods is discussed by considering some existing derivative-free schemes with and without memory. The derivative-free algorithms without and with memory considered are SA 6 , 2 , SA 6 . 16 , 2 , SA 6 . 46 , 2 , and SA 6 . 60 , 2 proposed by Sharma and Arora,24 which require five functions, two divided differences and one matrix inversion per step to be evaluated. Another derivative-free method GGN 6 , 3 without memory is by Grau et al.12 and this needs five functions, two divided differences and two matrix inverse per step to be evaluated. These schemes are given as follows:

Sharma and Arora method (SA 6 , 2 ):

y ( j ) = x ( j ) [ u ( j ) , x ( j ) ; F ] 1 F ( x ( j ) ) ,

z ( j ) = y ( j ) ( 2 I G ( j ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( y ( j ) ) ,

x ( j + 1 ) = z ( j ) ( 2 I G ( j ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( z ( j ) ) ,

where, G ( j ) = [ u ( j ) , x ( j ) ; F ] 1 [ v ( j ) , y ( j ) ; F ] , u ( j ) = x ( j ) + a F ( x ( j ) ) , v ( j ) = y ( j ) + c F ( y ( j ) ) , and a , c 0 are arbitrary constants.

Grau et al. method (GGN 6 , 3 ):

y ( j ) = x ( j ) [ u 1 ( j ) , v 1 ( j ) ; F ] 1 F ( x ( j ) ) ,

z ( j ) = y ( j ) ( 2 [ y ( j ) , x ( j ) ; F ] [ u 1 ( j ) , v 1 ( j ) ; F ] ) 1 F ( y ( j ) ) ,

x ( j + 1 ) = z ( j ) ( 2 [ y ( j ) , x ( j ) ; F ] [ u 1 ( j ) , v 1 ( j ) ; F ] ) 1 F ( z ( j ) ) ,

where u 1 ( j ) = x ( j ) + F ( x ( j ) ) , v 1 ( j ) = x ( j ) F ( x ( j ) ) .

Sharma and Arora method (SA 6 . 16 , 2 ):

y ( j ) = x ( j ) [ u ( j ) , x ( j ) ; F ] 1 F ( x ( j ) ) ,

z ( j ) = y ( j ) ( 2 I G ( j ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( y ( j ) ) ,

x ( j + 1 ) = z ( j ) ( 2 I G ( j ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( z ( j ) ) .

Sharma and Arora method (SA 6 . 46 , 2 ):

y ( j ) = x ( j ) [ u ( j ) , x ( j ) ; F ] 1 F ( x ( j ) ) ,

z ( j ) = y ( j ) ( 3 I G ( j ) ( 3 I G ( j ) ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( y ( j ) ) ,

x ( j + 1 ) = z ( j ) ( 3 I G ( j ) ( 3 I G ( j ) ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( z ( j ) ) .

Sharma and Arora method (SA 6 . 60 , 2 ):

y ( j ) = x ( j ) [ u ( j ) , x ( j ) ; F ] 1 F ( x ( j ) ) ,

z ( j ) = y ( j ) ( 3 I G ( j ) ( 3 I G ( j ) ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( y ( j ) ) ,

x ( j + 1 ) = z ( j ) ( 3 I G ( j ) ( 3 I G ( j ) ) ) [ u ( j ) , x ( j ) ; F ] 1 F ( z ( j ) ) .

In the methods, SA 6 . 16 , 2 , SA 6 . 46 , 2 , and SA 6 . 60 , 2 mentioned above,

G ( j ) = [ u ( j ) , x ( j ) ; F ] 1 [ v ( j ) , y ( j ) ; F ] , u ( j ) = x ( j ) + B ( j ) F ( x ( j ) ) , v ( j ) = y ( j ) + c F ( y ( j ) ) , B ( j ) = [ u ( j 1 ) , x ( j 1 ) ; F ] 1 and c 0 is an arbitrary constant for SA 6 . 16 , 2 and SA 6 . 46 , 2 , and v ( j ) = y ( j ) + C ( j ) F ( y ( j ) ) , B ( j ) = C ( j ) = [ u ( j 1 ) , x ( j 1 ) ; F ] 1 for SA 6 . 60 , 2 .

For ( p = 2 , q = 1 , r = 0 , a = 0 , b 0 , c = 0 , d 0 ) and ( p = 3 , q = 3 , r = 1 , a = 0 , b 0 , c = 0 , d 0 ) , the methods in (6) are referred as Ω 5 , 1 , Ω 6 , 1 respectively.

From the definition, and taking into consideration all aspects of computational cost discussed above, the values of C ρ , i , E ρ , i (the computational cost and efficiency index of scheme M ρ , i , respectively) can be seen in Table 1.

TABLE 1. Values of C ρ , i , E ρ , i for various schemes
Methods C ρ , i E ρ , i
Ω 5 , 1 1 6 n 17 + 2 n 2 + 3 l ( 7 + 5 n ) + 18 μ + 3 n ( 9 + 4 μ ) ( 5 ) 1 / C 5 , 1
Ω 5 . 54 , 1 1 6 n 17 + 2 n 2 + 3 l ( 9 + 5 n ) + 18 μ + 3 n ( 11 + 4 μ ) ( 5 . 54 ) 1 / C 5 . 54 , 1
Ω 6 , 1 1 6 n 23 + 2 n 2 + 3 l ( 9 + 5 n ) + 18 μ + 3 n ( 13 + 4 μ ) ( 6 ) 1 / C 6 , 1
SA 6 , 2 1 6 n 17 + 2 n 2 + 3 l ( 9 + 5 n ) + 18 μ + 3 n ( 13 + 4 μ ) ( 6 ) 1 / C 6 , 2
GGN 6 , 3 1 3 n 8 + 2 n 2 + l ( 6 + 9 n ) + 9 μ + n ( 9 + 6 μ ) ( 6 ) 1 / C 6 , 3
Ω 6 . 16 , 1 1 6 n 29 + 2 n 2 + 3 l ( 11 + 5 n ) + 18 μ + 3 n ( 15 + 4 μ ) ( 6 . 16 ) 1 / C 6 . 16 , 1
SA 6 . 16 , 2 1 6 n 23 + 2 n 2 + 3 l ( 11 + 5 n ) + 18 μ + 3 n ( 15 + 4 μ ) ( 6 . 16 ) 1 / C 6 . 16 , 2
Ω 6 . 46 , 1 1 6 n 29 + 2 n 2 + 3 l ( 13 + 5 n ) + 18 μ + 3 n ( 17 + 4 μ ) ( 6 . 46 ) 1 / C 6 . 46 , 1
SA 6 . 46 , 2 1 6 n 35 + 2 n 2 + 15 l ( 3 + n ) + 18 μ + 3 n ( 23 + 4 μ ) ( 6 . 46 ) 1 / C 6 . 46 , 2
Ω 6 . 60 , 1 1 6 n 23 + 2 n 2 + 3 l ( 11 + 5 n ) + 18 μ + 3 n ( 15 + 4 μ ) ( 6 . 60 ) 1 / C 6 . 60 , 1
SA 6 . 60 , 2 1 6 n 41 + 2 n 2 + 3 l ( 17 + 5 n ) + 18 μ + 3 n ( 25 + 4 μ ) ( 6 . 60 ) 1 / C 6 . 60 , 2

4.1 Comparison between efficiencies

To carry out the comparison between computational efficiencies of any two schemes M A and M B , we consider the ratio
R M A M B = log E M A log E M B = C M B log ( ρ A ) C M A log ( ρ B ) , ()
where ρ A and ρ B are the convergence order of schemes M A and M B , respectively. Clearly, if R M A M B > 1 , the iterative method M A is more efficient than M B . The boundary between the computational efficiencies is given by the equation R M A M B = 1 , which expresses μ as a function of n and l; ( μ > 0 , n is a positive integer 2 and l 1 ).

The comparison of efficiencies of various new schemes with existing schemes has been tabulated in Table 2, whereas the comparison of new methods among themselves has been tabulated in Table 3. These results can be easily verified using the ratio given in (58).

TABLE 2. Comparison of efficiency index of new and existing schemes
S A 6 , 2 G G N 6 , 3 SA 6 . 16 , 2 S A 6 . 46 , 2 S A 6 . 60 , 2
Ω 5 , 1 E 5 , 1 < E 6 , 2 n 33 E 5 . , 1 > E 6 , 3 n 7 E 5 , 1 < E 6 . 16 , 2 n 49 E 5 , 1 < E 6 . 46 , 2 n 112 E 5 , 1 < E 6 . 60 , 2 n 119
Ω 5 . 54 , 1 E 5 . 54 , 1 < E 6 , 2 n 41 E 5 . 54 , 1 > E 6 , 3 n 9 E 5 . 54 , 1 < E 6 . 16 , 2 n 73 E 5 . 54 , 1 < E 6 . 46 , 2 n 177 E 5 . 54 , 1 < E 6 . 60 , 2 n 182
Ω 6 , 1 E 6 , 1 > E 6 , 2 n 2 E 6 , 1 > E 6 , 3 n 2 E 6 , 1 < E 6 . 16 , 2 n 179 E 6 , 1 < E 6 . 46 , 2 n 338 E 6 , 1 < E 6 . 60 , 2 n 312
Ω 6 . 16 , 1 E 6 . 16 , 1 > E 6 , 2 n 177 E 6 . 16 , 1 > E 6 , 3 n 12 E 6 . 16 , 1 > E 6 . 16 , 2 n 2 E 6 . 16 , 1 < E 6 . 46 , 2 n 430 E 6 . 16 , 1 < E 6 . 60 , 2 n 366
Ω 6 . 46 , 1 E 6 . 46 , 1 > E 6 , 2 n 119 E 6 . 46 , 1 > E 6 , 3 n 14 E 6 . 46 , 1 > E 6 . 16 , 2 n 85 E 6 . 46 , 1 > E 6 . 46 , 2 n 2 E 6 . 46 , 1 < E 6 . 60 , 2 n 1012
Ω 6 . 60 , 1 E 6 . 60 , 1 > E 6 , 2 n 30 E 6 . 60 , 1 > E 6 , 3 n 11 E 6 . 60 , 1 > E 6 . 16 , 2 n 2 E 6 . 60 , 1 > E 6 . 46 , 2 n 2 E 6 . 60 , 1 > E 6 . 60 , 2 n 2
TABLE 3. Comparison of efficiency index of proposed methods among themselves
Ω 5 , 1 Ω 5 . 54 , 1 Ω 6 , 1 Ω 6 . 16 , 1 Ω 6 . 46 , 1 Ω 6 . 60 , 1
Ω 5 , 1 * * * * *
Ω 5 . 54 , 1 n 28 * * * *
Ω 6 , 1 n 32 n 39 * * *
Ω 6 . 16 , 1 n 49 n 73 n 178 * *
Ω 6 . 46 , 1 n 55 n 77 n 120 n 87 *
Ω 6 . 60 , 1 n 32 n 35 n 32 n 3 n 2

Note 1. In Table 3 “*” indicates that the comparison of method M A has been done with M B , so the comparison of M B with M A is not shown.

Note 2. The entries in Table 3 indicate the values of “n” for which one method is more efficient than the other method, for example, the method Ω 5 . 54 , 1 is more efficient than the method Ω 5 , 1 for all n 28 .

Note 3. For the values of “n” other than the ones mentioned in table, the comparison depends on μ and “n.”

5 NUMERICAL RESULTS

In this section, we discuss some numericals to examine convergence behavior of the proposed schemes. The new methods Ω 5 , 1 , Ω 5 . 54 , 1 , Ω 6 , 1 , Ω 6 . 16 , 1 , Ω 6 . 46 , 1 , and Ω 6 . 60 , 1 for particular values of the parameters P ( 0 ) have been compared with the existing schemes S A 6 , 2 , G G N 6 , 3 , S A 6 . 16 , 2 , S A 6 . 46 , 2 , and S A 6 . 60 , 2 . The values of parameters used for existing schemes S A 6 . 16 , 2 , S A 6 . 46 , 2 , and S A 6 . 60 , 2 have been kept same as used by the authors in Reference 24.

The computational work is compiled in the programming software M a t h e m a t i c a 7 ,25 using multiple-precision arithmetic with 4096 digits. For each problem and scheme, we tabulate the number of iterations (Itr) that the initial point uses to converge to the solution, by considering the stopping criterion x ( j + 1 ) x ( j ) + F ( x ( j + 1 ) ) < 1 0 300 , Error ( x ( j + 1 ) x ( j ) + F ( x ( j + 1 ) ) ) at the end of fourth step, computational cost C A and the efficiency index E A using the values given in Table 1 and the values “n” and “ μ ” are given in each problem. To affirm the theoretical order of convergence, we measured the computational order of convergence (COC),26 using the formula
COC = log F ( x ( j + 1 ) ) / F ( x ( j ) ) log F ( x ( j ) ) / F ( x ( j 1 ) ) ,
where the last three approximations are used to calculate COC. The e-Time reported in the last column is the mean CPU time. This is calculated by performing the program 25 times and taking its mean, by considering the above stopping criterion used in single execution of the program.

To compute the values μ and l for numerical examples, we express the evaluation cost of elementary functions in terms of products. This depends on the arithmetic used, computer, and the software27 (www.mpfr.org/mpfr-2.1.0/timings.html). These programs have been run in M a t h e m a t i c a 7 on a processor with specifications: Intel(R) Core (TM) i5-3210M CPU @2.50 GHz (64-bit machine) Microsoft Windows 7 (2009) and are given in Table 4. This shows the estimates of cost of the elementary functions in product units, where the running time of one product is measured in milliseconds (ms). From Table 4, one can observe that the time taken by one quotient is twice the time taken by product and hence l = 2 . The following numerical examples are considered to verify the results.

TABLE 4. CPU-time and estimate of computational cost of elementary functions for x = 3 1 and y = 5
Function x y x / y x exp x ln x sin x cos x arccos x arcsin x arctan x x y
CPU-time 0.04212 0.08424 0.03822 2.7924 2.4383 3.0093 3.0061 5.1387 4.9951 4.9702 5.6972
Cost 1 2 0.91 66.30 57.89 71.45 71.44 122 118.59 118.00 135.26

Problem 1.Here, we study the mixed Hammerstein's integral equation:3

x ( s ) = 1 + 1 5 0 1 G ( s , t ) x ( t ) 3 d t , s , t [ 0 , 1 ] where x C [ 0 , 1 ]
and the kernel G is G ( s , t ) = ( 1 s ) t , t s , s ( 1 t ) , s t .

The above integral equation is discretized into a finite dimensional problem with the help of Gauss–Legendre quadrature formula for abscissas t m and weights ω m determined for n = 12 . Let x m denote the approximation x ( t m ) , ( m = 1 , 2 , . . . , 12 ) . This results in a nonlinear system of equations given below:

5 x i 5 m = 1 12 a i m x m 3 = 0 , i = 1 , 2 , , 12 ,
where
a i m = ω m t m ( 1 t i ) i f m i , ω m t i ( 1 t m ) i f i < m .

The solution of the problem is

α = { 1 . 0009727166180117 , 1 . 0048748186599682 , 1 . 0109092367279116 , 1 . 0176086786577538 , 1 . 0233126345057937 , 1 . 0265822324745664 , 1 . 0265822324745664 , 1 . 0233126345057937 , 1 . 0176086786577538 , 1 . 0109092367279116 , 1 . 0048748186599682 , 1 . 0009727166180117 } t , which is correct up to 16 decimal places. The methods are tested using the parameters ( n , μ ) = ( 12 , 15 ) and the initial approximation of the solution as x ( 0 ) = { 0 . 9 , 0 . 9 , , 0 . 9 } t and the results are displayed in Table 5.

TABLE 5. Numerical results of different schemes for problem 1
Methods ρ Error Itr C A E A COC e-Time
Ω 5 , 1 ( b = d = 0 . 01 ) 5 3.116 ( 181) 5 6854 1.000234844 5.000 7.190
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 5.54 6.859 ( 226) 5 7022 1.00024383 5.645 7.206
Ω 6 , 1 ( b = 0 . 01 , d = 0 . 01 ) 6 2.482 ( 283) 5 7154 1.00025049 6.000 6.806
SA 6 , 2 ( b = c = 0 . 01 ) 6 1.779 ( 245) 5 7166 1.00025007 6.000 7.610
GGN 6 , 3 6 1.428 ( 7) 7 7324 1.00024467 6.000 10.364
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 6.16 5.490 ( 310) 4 7310 1.00024874 6.163 5.884
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.16 3.190 ( 272) 5 7322 1.00024833 6.162 6.625
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.46 3.881 ( 349) 4 7478 1.00024951 6.521 6.104
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.46 1.800 ( 299) 5 7922 1.00023553 6.464 6.407
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 1.254 ( 363) 4 7322 1.00025776 6.701 5.203
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 2.054 ( 318) 4 8078 1.00023363 6.615 6.121

Problem 2.Now, we discuss a system of fifty equations given by:

arctan x i + 1 2 m = 1 , m i 50 x m 2 = 0 , 1 i 50 .
The approximate solution is α = { 0 . 0960567972718922 , 0 . 0960567972718922 , , 0 . 0960567972718922 } t , which is correct up to 16 digits. With initial guess as x ( 0 ) = { 1 3 , 1 3 , 1 3 } t and parameters as ( n , μ ) = ( 50 , 120 ) , the methods are tested. Results for this problem are shown in Table 6.

TABLE 6. Numerical results of different schemes for problem 2
Methods ρ Error Itr C A E A COC e-Time
Ω 5 , 1 ( b = d = 0 . 01 ) 5 2.828 ( 28) 6 683,625 1.0000023543 5.000 481.21
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 5.54 5.996 ( 55) 5 686,225 1.0000024948 5.646 509.44
Ω 6 , 1 ( b = d = 0 . 01 ) 6 1.975 ( 45) 6 688,675 1.0000026018 6.000 540.02
SA 6 , 2 ( b = c = 0 . 01 ) 6 1.887 ( 43) 6 688,725 1.0000026016 6.000 596.58
GGN 6 , 3 6 2.866 ( 48) 6 723,900 1.0000024752 6.000 640.85
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 6.16 4.650 ( 49) 5 691,225 1.0000026302 6.162 346.76
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.16 3.511 ( 47) 6 691,275 1.0000026300 6.162 481.42
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.46 6.862 ( 85) 5 693,825 1.0000026889 6.541 417.29
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.46 2.310 ( 65) 5 701,375 1.0000026600 6.463 452.35
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 1.855 ( 86) 5 691,275 1.0000027298 6.707 443.78
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 1.271 ( 67) 5 703,925 1.0000026808 6.619 489.79

Problem 3.Now, we examine the following boundary value problem:

y = 1 2 y 3 + 3 y 3 2 x + 1 2 , y ( 0 ) = 0 , y ( 1 ) = 1 .

Assuming that the interval [ 0 , 1 ] is partitioned as, x 0 = 0 < x 1 < x 2 < < x m = 1 , x j + 1 = x j + h , h = 1 m , we define y 0 = y ( x 0 ) = 0 , y 1 = y ( x 1 ) , , y m 1 = y ( x m 1 ) , y m = y ( x m ) = 1 . The problem is discretized using the numerical formulas for first and second derivative given by:

y k = y k + 1 y k 1 2 h , y k = y k 1 2 y k + y k + 1 h 2 , k = 1 , 2 , 3 , . . . , m 1 .

This yields a system of m 1 nonlinear equations in m 1 variables:

y k + 1 2 y k + y k 1 h 2 2 y k 3 3 2 h ( y k + 1 y k 1 ) + 3 2 x k h 2 1 2 h 2 = 0 , k = 1 , 2 , 3 , , m 1 .
In this problem, we take m = 200 . Hence, a system of n = 199 nonlinear equations is solved. The initial guess is taken as y ( 0 ) = { 43 100 , 43 100 , . . . , 43 100 } t and we apply different methods to acquire approximate solution. The solution correct up to 16 decimal places is α = { 0 . 0025062505477845 , 0 . 0050250953257465 , , 0 . 98019785632803 , 0 . 99004966827654 } t . The values of the parameters are ( n , μ ) = ( 199 , 5 ) . Numerical results for the problem can be seen in Table 7.

TABLE 7. Numerical results of different schemes for problem 3
Methods ρ Error Itr C A E A COC e-Time
Ω 5 , 1 ( b = d = 0 . 01 ) 5 2.826 ( 250) 5 3,402,900 1.000000472961 5.000 76.48
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 5.54 7.837 ( 308) 4 3,442,899 1.000000497253 5.648 72.48
Ω 6 , 1 ( b = d = 0 . 01 ) 6 1.872 ( 412) 4 3,482,301 1.000000514533 6.001 76.15
SA 6 , 2 ( b = c = 0 . 01 ) 6 9.243 ( 385) 4 3,482,500 1.000000514504 6.000 84.03
GGN 6 , 3 6 4.117 ( 325) 4 6,009,402 1.000000298159 6.003 86.31
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 6.16 1.411 ( 448) 4 3,522,101 1.000000516191 6.164 72.61
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.16 5.067 ( 423) 4 3,522,300 1.000000516161 6.163 78.15
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.46 4.356 ( 504) 4 3,562,100 1.000000523744 6.539 73.74
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.46 2.743 ( 488) 4 3,681,102 1.000000506812 6.466 85.24
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 1.453 ( 528) 4 3,522,300 1.000000535749 6.699 72.48
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 4.591 ( 514) 4 3,720,902 1.000000507153 6.610 89.17

Problem 4.Now, we focus on the Van der Pol equation28, 29 which governs the flow of current in a vacuum tube is defined by:

y μ ˜ ( y 2 1 ) y + y = 0 , μ ˜ > 0 .
The boundary conditions are y ( 0 ) = 0 , y ( 2 ) = 1 . The interval [ 0 , 2 ] is partitioned as follows with nodes as x 0 = 0 < x 1 < x 2 < < x m = 2 , x j + 1 = x 0 + j h , h = 2 m , and we define y 0 = y ( x 0 ) = 0 , y 1 = y ( x 1 ) , , y m 1 = y ( x m 1 ) , y m = y ( x m ) = 1 . The problem is discretized using the numerical formulas for first and second derivative as given above:

This yields a system of nonlinear equations of size ( m 1 ) × ( m 1 ) :

( 2 h 2 4 ) y 1 + ( 2 + μ ˜ h ) y 2 μ ˜ h y 1 2 y 2 = 0 , ( 2 μ ˜ h ) y k 1 + ( 2 h 2 4 ) y k + ( 2 + μ ˜ h ) y k + 1 μ ˜ h y k 2 ( y k + 1 y k 1 ) = 0 , k = 2 , 3 , , m 2 , ( 2 μ ˜ h ) y m 2 + 2 ( 2 h 2 4 ) y m 1 + ( 2 + μ ˜ h ) + μ ˜ h y n 1 2 ( 1 y n 2 ) = 0 .
In this problem, we take μ ˜ = 1 / 2 , m = 51 . Hence, a system of n = 50 nonlinear equations is solved. The initial guess is taken as y ( 0 ) = Table 1 i , { i , 1 , n } and we apply different methods to acquire approximate solution.

The values of the parameters are ( n , μ ) = ( 50 , 9 ) . Numerical results for the problem can be seen in Table 8.

TABLE 8. Numerical results of different schemes for Problem 4
Methods ρ Error Itr C A E A COC e-Time
Ω 5 , 1 ( b = d = 0 . 01 ) 5 4.269 ( 104) 5 111,975 1.000014373 4.998 5.807
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 5.54 2.898 ( 114) 5 114,575 1.00001494224 5.012 6.306
Ω 6 , 1 ( b = d = 0 . 01 ) 6 4.419 ( 164) 5 117,025 1.0000153110 6.011 6.686
SA 6 , 2 ( b = c = 0 . 01 ) 6 2.712 ( 149) 5 117,075 1.0000153045 6.009 7.451
GGN 6 , 3 6 1.692 ( 61) 6 152,250 1.00001176860 4.014 8.123
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 6.16 6.569 ( 185) 6 119,575 1.00001520460 6.165 6.258
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.16 4.881 ( 174) 5 119,625 1.0000151983 6.164 7.125
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.46 7.672 ( 187) 5 122,175 1.00001527026 6.003 6.231
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 6.46 1.495 ( 201) 5 129,725 1.0000143815 6.470 7.896
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 3.752 ( 194) 5 119,625 1.00001577500 5.999 6.208
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) 6.60 2.235 ( 213) 5 132,275 1.00001426636 6.614 7.458

Problem 5.A nonlinear partial differential equation can be solved with help of finite difference discretization. This yields a system of nonlinear algebraic equations. Here, we consider nonlinear PDE known as Fisher's equation30 with homogeneous Neumann's boundary conditions and the diffusion coefficient D give below:

u t = D u x x + u ( 1 u ) , u ( x , 0 ) = 1 . 5 + 0 . 5 C o s ( π x ) , 0 x 1 , u x ( 0 , t ) = 0 t 0 , u x ( 1 , t ) = 0 t 0 . ()

Assume M and N to be number of partitions in x and t directions, and their respective step size as h and k. Let u = u ( x , t ) denote the exact solution of the given system and its approximate solution be w i , j u ( x i , t j ) at the grid points of the mesh created. Applying backward difference to u t and central difference to the other terms, that is, u t ( x i , t j ) ( w i , j w i , j 1 ) k , u x ( x i , t j ) ( w i + 1 , j w i 1 , j ) 2 h and u x x ( x i , t j ) ( w i + 1 , j 2 w i , j + w i 1 , j ) h 2 to solve the PDE with values M = 21 and N = 21 , and D = 1 gives rise to nonlinear system equations of size m n where m = M 1 , n = N 1 . Here, we take two different initial guesses x ( 0 ) = Table 5 4 + i 2 m n , { i , 1 , m n } and x ( 0 ) = Table 12 5 + i 2 m n , { i , 1 , m n } . This has resulted in a system of size 400, which is solved using schemes mentioned earlier.

In case of PDE's, since the precision of solution required is low, so the stopping criterion used is F ( x ) 2 < 1 0 8 . We count iterations and mean CPU time (e-Time) used by schemes to arrive at the stopping criterion. We take average of 25 performances of the program for e-Time. The results are displayed in Table 9, and the approximate solution of the system can be seem in Figure 1.

Details are in the caption following the image
Approximate solution of Fisher's equation
TABLE 9. Numerical results of different schemes for Problem 5
Table 5 4 + i 2 m n , { i , 1 , m n } Table 12 5 + i 2 m n , { i , 1 , m n }
Methods Itr e-Time Itr e-Time
Ω 5 , 1 ( b = d = 0 . 01 ) 3 9.81 7 89.25
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 1 I , d = 0 . 1 ) 3 8.52 4 80.24
Ω 6 , 1 ( b = d = 0 . 01 ) 3 8.06 Diverges
SA 6 , 2 ( b = c = 0 . 01 ) 6 11.65 Diverges
GGN 6 , 3 3 10.64 4 100.25
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 1 I , d = 0 . 01 ) 3 7.1 Diverges
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) Diverges 4 82.17
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 3 7.27 4 72.14
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) Diverges 4 86.12
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 3 7.17 4 70.16
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) Diverges 3 85.75

6 BASINS OF ATTRACTION

The dynamical behavior of the proposed algorithms is examined by generating their basins of attraction and comparing them with the existing methods mentioned earlier. We examine the following test problems which are systems of polynomials.

Example 1.

( x 1 3 ) 2 16 x 2 2 1 = 0 , x 1 2 x 2 2 1 = 0 ,
with approximate solutions { { 1 . 48062 , 1 . 0919 } , { 1 . 48062 , 1 . 0919 } , { 1 . 08062 , 0 . 409573 } , { 1 . 08062 , 0 . 409573 } } .

Example 2.

x 1 x 2 + x 1 x 2 1 = 0 , x 1 x 2 x 1 + x 2 1 = 0 ,
with solutions { { 1 , 1 } , { 1 , 1 } .

Example 3.

x 1 2 1 = 0 , x 2 2 1 = 0 ,
with its solutions as { { 1 , 1 } , { { 1 , 1 } , { { 1 , 1 } , { 1 , 1 } .

For generating basins of attraction, we take a square 2 , 2 × 2 , 2 of 401 × 401 points containing all solutions of system of equations. The iterative algorithm is then applied to all points in the square as initial points. We allocate color to each point depending on the root to which the corresponding orbit of the scheme, starting from the point, converges. In case the corresponding orbit does not approach any root of the polynomial system or tends to infinity, with tolerance 1 0 3 and in a maximal of 50 iterations, black color is assigned to the point. The number of iterations required to attain the solution with the given stopping criterion is shown in brighter or darker colors. The brighter color means number of iterations required is lower.

The attraction basin of examples generated using respective schemes have been displayed in Figures 2, 3, and 4.

Details are in the caption following the image
Basins of attraction for Problem 1
Details are in the caption following the image
Basins of attraction for Problem 2
Details are in the caption following the image
Basins of attraction for Problem 3

To compare the basins quantitatively, we calculate total number of initial points convergent in all test problem and then find the average number of iterations required per point. The average performance results of the schemes in comparison are displayed in Table 10. We can derive following conclusions on considering the average of three examples in Table 10:

TABLE 10. Numerical results of schemes Ω 5 , 1 , Ω 5 . 54 , 1 , Ω 6 , 1 , SA 6 , 2 , GGN 6 , 3 , Ω 6 . 16 , 1 , SA 6 . 16 , 2 , Ω 6 . 46 , 1 , SA 6 . 46 , 2 , Ω 6 . 60 , 1 , and SA 6 . 60 , 2
Average convergent points Average iterations
Methods eg1 eg2 eg3 Average % eg1 eg2 eg3 Average
Ω 5 , 1 ( b = d = 0 . 01 ) 120,404 129,295 123,128 12,4275.67 77.29 15.46 12.80 14.64 14.3
Ω 5 . 54 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 123,222 132,716 127,998 127,978.67 79.59 14.74 11.85 13.38 13.32
Ω 6 , 1 ( P ( 0 ) = 0 . 01 , d = 0 . 01 ) 115,285 122,779 118,233 118,765.67 73.86 16.68 14.48 15.79 15.65
SA 6 , 2 ( b = c = 0 . 01 ) 132,490 141,946 138,360 137,598.67 85.57 12.84 9.78 10.97 11.20
GGN 6 , 3 159,998 160,668 160,000 160,222 99.63 3.38 3.48 3.16 3.34
Ω 6 . 16 , 1 ( P ( 0 ) = 0 . 01 I , d = 0 . 01 ) 112,973 124,303 116,878 118,051.33 73.41 17.23 9.25 16.09 14.19
SA 6 . 16 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 105,756 122,873 114,825 114,484.67 71.20 19.49 14.40 17.06 16.98
Ω 6 . 46 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 149,241 151,659 151,223 150,707.67 93.72 9.49 7.99 8.83 8.77
SA 6 . 46 , 2 ( B ( 0 ) = 0 . 01 I , c = 0 . 01 ) 144,509 146,957 145,159 145,541.67 90.51 14.22 12.56 13.84 13.54
Ω 6 . 60 , 1 ( P ( 0 ) = C ( 0 ) = 0 . 01 I ) 148,918 151,597 151,255 150,704 93.72 9.54 7.99 8.84 8.79
SA 6 . 60 , 2 ( B ( 0 ) = C ( 0 ) = 0 . 01 I ) 144,509 146,957 145,160 145,542 90.51 14.22 12.56 13.84 13.54

(i) Both average number of convergent points and average number of iterations of schemes Ω 5 , 1 and Ω 5 . 54 , 1 indicate that the new algorithm with memory Ω 5 . 54 , 1 is better than the method without memory Ω 5 , 1 .

(ii) By considering convergent points and average number of iterations in case schemes of order 6 and above, methods with memory Ω 6 . 60 , 1 and Ω 6 . 46 , 1 are best and then followed by SA 6 . 60 , 2 , SA 6 . 46 , 2 . This is followed by Ω 6 . 16 , 1 and further followed by SA 6 . 16 , 2 .

Therefore, from Table 10 one can say that the new schemes Ω 6 . 60 , 1 and Ω 6 . 46 , 1 are best among all methods.

7 CONCLUSIONS

In this article, we proposed two families of fifth-order and sixth-order derivative-free schemes to solve nonlinear systems of equations, numerically. Further, from these families we derived several derivative-free schemes with memory. The proposed algorithms involve evaluation of only two divided differences and inverse of only one matrix per step. Consequently, the schemes are low in computational cost and result in high efficiency index. Some special members of the collection have been compared with some of the existing schemes using numerical examples and comparison of efficiencies has been shown. Various parameters like the error in approximation of the roots at end of fourth iterations, the number of iterations required, and the e-Time (CPU time) utilized by the scheme to accomplish the stopping criterion, verify that the new schemes give faster and closer approximation of roots than the already existing schemes. These results confirmed that the proposed schemes with memory Ω 6 . 60 , 1 and Ω 6 . 46 , 1 are best and more efficient than the existent schemes. Finally, these results have been revalidated with help of their basins of attraction. Thus, it can be ascertained that these novel schemes exhibit highly efficient nature particularly for large-scale systems.

ACKNOWLEDGMENT

The authors would like to thank all the five anonymous reviewers for their useful comments and suggestions, which have improved the final version of this manuscript.

    CONFLICT OF INTEREST

    The authors declare no potential conflict of interests.

    Biographies

    • biography image

      Mona Narang received her B.Sc.(Hons. School), M.Sc.(Hons. School), and Ph.D. in numerical analysis from Department of Mathematics, Panjab University, Chandigarh. Currently, she is working as an Associate Professor in the Department of Mathematics, D.A.V. College, Chandigarh. Her research interests include iterative schemes for numerical solution of system of nonlinear equations and its applications in solving real-life problems, dynamical study of iterative methods.

    • biography image

      Saurabh Bhatia received his B.Sc.(Hons. School), M.Sc.(Hons. School), and Ph.D. in valuation theory from Department of Mathematics, Panjab University, Chandigarh. Currently, he is working in the Department of Mathematics, University Institute of Engineering and Technology, Panjab University, Chandigarh. His research interests include iterative schemes for numerical solution of nonlinear equations and system of nonlinear equations and its applications in solving real-life problems.

    • biography image

      Vinay Kanwar did his Ph.D. in fluid dynamics in the year 1996 from Department of Mathematics, Himachal Pradesh University, Shimla, India. Currently, he is working as a Professor in Mathematics at University Institute of Engineering and Technology, Panjab University, Chandigarh. Presently, he is working in numerical analysis, fixed point theory, specifically in iterative techniques for solving nonlinear equations and nonlinear systems, numerical solution of differential equations, and their applications in solving real-life problems.

      The full text of this article hosted at iucr.org is unavailable due to technical difficulties.