Optimal efficiency indexes for iterative methods of interpolatory type

Abstract

The paper is concerned with the order of convergence and the efficiency index of iterative methods of interpolatory type for solving scalar equations. Some classes of such methods are presented and then, using well defined criteria, the methods having the optimal efficiency index (i.e. those which practically are most efficient) are determined. For these methods the efficiency indexes are effectively determined.

Authors

Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)

Keywords

PDF

Cite this paper as:

I. Păvăloiu, Optimal efficiency indexes for iterative methods of interpolatory type, Computer Science Journal of Moldova, 5 (1997) no. 1(13), pp. 20-43.

About this paper

Print ISSN

1561-4042

Online ISSN

Not available yet.

References

?

Paper (preprint) in HTML form

OPTIMAL EFFICIENCY INDEXES FOR ITERATIVE METHODS OF INTERPOLATORY TYPE AMS Subject Classification (1991): Primary 65Y20, 68Q25, 65H05.This work has been supported by the Romanian Academy of Sciences.

OPTIMAL EFFICIENCY INDEXES FOR ITERATIVE METHODS OF INTERPOLATORY TYPE thanks: AMS Subject Classification (1991): Primary 65Y20, 68Q25, 65H05.thanks: This work has been supported by the Romanian Academy of Sciences.

Abstract

The paper is concerned with the order of convergence and the efficiency index of iterative methods of interpolatory type for solving scalar equations. Some classes of such methods are presented and then, using well defined criteria, the methods having the optimal efficiency index (i.e. those who practically are the most efficient) are determined. For these methods the efficiency indexes are effectively determined.

1 Introduction

In this paper we propose a unitary approach concerning the computational complexity for the numerical solving of scalar equations by iterative methods of interpolatory type. We shall consider some classes of such methods from which, using well defined criteria, we shall choose the optimal ones.

As a measure for the complexity of a method we shall adopt the efficiency  index (see [4]).

For this purpose we shall start by presenting some general considerations concerning the convergence order and the efficiency index of an iterative method. Then we shall specify the interpolatory methods to be studied. Finally, we shall select from the interpolatory classes those having the highest efficiency index, and for the classes for which the selection method can’t be applied, we shall give delimitations for the efficiency index.

2 The convergence order and the efficiency index

Denote by I=[a,b],a,b𝐑,a<b and consider the equation

(1) f(x)=0

where f:I𝐑. In the following we shall suppose, for simplicity, that the equation (2.1) has a unique solution x¯I. Let g:II be a function having a unique fixed point in the interval I, which coincides with x¯.

For the approximation of the root x¯ of equation (2.1), under certain conditions, we may consider the elements of the sequence (xp)p0 generated by the following iterative process

(2) xs+1=g(xs)x0I,s=0,1,

More generally, if G:IkI is a function of k variables whose restriction to the diagonal of the set Ik coincides with g, i.e.

G(x,x,,x)=g(x),for all xI,

then we may consider the following iterative process:

(3) xs+k=G(xs,xs+1,,xs+k1),s=0,1,,x0,x1,,xk1I,

The convergence of the sequence (xp)p0 generated by (2.2) or (2.3) depends on certain properties of the functions f and g, respectively G. The amount of time needed by a computer to obtain a suitable approximation of x¯, depends both on the convergence order of the sequence (xp)p0 and on the number of elementary operations that must be performed at each iteration step in (2.2) or (2.3). Concerning the convergence order, it can be exactly computed in almost all the cases. Hence it remains to solve the difficult problem of determining the number of elementary operations that must be performed at each iteration step. A general approach to this problem of course can’t be successful. That’s why A.M. Ostrowski proposed in [4] a simplification of this problem, by considering the number of function evaluations at each iteration step. At first sight this approach seems to be strange, taking into account that some functions may be more complicated and others may be simpler from the computational standpoint. But our purpose is to compare different methods applied to the same equation, and such an approach can give results.

We consider an arbitrary sequence (xp)p0, satisfying together with f and g the following properties:

a)

xsI and g(xs)I for s=0,1,;

b)

the sequences (xp)p0 and (g(xp))p0 are convergent and limxp=limg(xp)==x¯, wherex¯ is the solution of (2.1);

c)

For all x,yI,0<|[x,y;f]|m,m𝐑,m>0, where we have denoted by [x,y;f] the first order divided difference of f on the nodes x and y;

d)

f is derivable at x¯.

Definition 2.0.1

The sequence (xp)p0 has the convergence order ω𝐑, ω1, in respect to the function g, if there exists the limit:

(4) α=limpln|g(xp)x¯|ln|xpx¯|

and α=ω.

Remark 2.0.1

If the sequence (xp)p0 is generated by the iterative method (2.2), then the Definition 2.1. reduces to the known one [4].

For a unitary treatment of the determination of the convergence order of the studied methods, we shall use the following lemmas.

Lemma 2.0.0

If the sequence (xp)p0 and the functions f and g satisfy the properties a) - d) then the necessary and sufficient condition for the sequence (xp)p0 to have the convergence order ω𝐑,ω1, with respect to the function g is that the following limit exists:

(5) β=limpln|f(g(xp))|ln|f(xp)|

and β=ω.

Proof. Supposing that one of the equalities (2.4) or (2.5) is true and taking into account the properties a) - d), we obtain:

limln|g(xp)x¯|ln|xpx¯|=limln|f(g(xp))|ln|[g(xp),x¯;f]|ln|f(xp)|ln|[xp,x¯;f]|=
=limln|f(g(xp))|ln|f(xp)|1ln|[g(xp),x¯;f]|ln|f(g(xp))|1ln|[xp,x¯;f]|ln|f(xp)|=limln|f(g(xp))|ln|f(xp)|.

Lemma is proved.

Lemma 2.0.0

If (up)p0 is a sequence of real positive numbers satisfying:

i. The sequence (up)p0 is convergent and limup=0;

ii. There exist the nonnegative real numbers α1,,αn+1 and a bounded sequence (cp)p0 with cs>0 for all s=0,1,, and  0<inf{cp} such that the elements of (up)p0 satisfy

(6) us+n+1=csusα1us+1α2us+nαn+1,s=0,1,;

iii. The sequence lnup+1lnup is convergent and ω=limlnup+1lnup>0.

Then ω is the positive solution of the equation:

tn+1αn+1tnαntn1α2tα1=0.

Proof. From (2.6) we obtain

limslnun+s+1lnun+s=limslncslnun+s+i=0nαi+1limslnus+ilnus+n.

But it can be easily seen that

limslncslnun+s=0 and limslnus+ilnus+n=1ωni,i=0,n¯

whence it follows thatω=i=0nαi+11ωni, i.e. ωn+1i=0nαi+1ωi=0. Lemma is proved.

We shall denote in the following by mp the number of function evaluations that must be performed at each iteration step p in (2.2), respectively (2.3), for p=0,1,.

In the hypotheses of Lemma 2.1 and taking into account the definition given in [4], we have:

Definition 2.0.2

The real number E is called the efficiency index of the iterative method (2.2) or (2.3) if there exists

L=lim(ln|f(xp+1)|ln|f(xp)|)1mp

and L=E.

Remark 2.0.2

If for the methods (2.2) and (2.3) there exists a natural number s0 such that ms=r for all s>s0 and ω is the convergence order of these methods, then the efficiency index E is given by the following expression:

(7) E=ω1r.

3 Iterative methods of interpolatory type

In the following we shall briefly present the Lagrange-Hermite-type inverse interpolatory polynomial. It is well known that this leads us to general classes of iterative methods from which, by suitable particularizations we obtain usual methods as Newton’s method, chord’s method, Chebyshev’s method, etc.

For the sake of simplicity we prefer to treat separately the Hermite polynomial and the Lagrange polynomial, though the last is a particular case of the first.

As we shall see, a suitable choice of the nodes enables us to improve the convergence orders of Lagrange-Hermite-type methods. We shall call such methods Steffensen-type methods.

3.1 Lagrange-type inverse interpolation

Denote by F=f(I) the range of f for xI. Suppose f is n+1 times differentiable and f(x)0 for all xI. It follows that f is invertible and there exists f1:FI. Consider n+1 interpolation nodes in I:

(8) x1,x2,,xn+1,xixj,for i,j=1,n+1¯,ij.

In the above hypotheses it follows that the solution x¯ of equation (2.1) is given by

x¯=f1(0).

Using the Lagrange interpolatory polynomial for the function f1 at the nodes f(x1),,f(xn+1) we shall determine an approximation for f1(0), i.e. for x¯.

Denote yi=f(xi),i=1,n+1¯ and let L(y1,y2,,yn+1;f1y)be the mentioned polynomial, which is known to have the form

L(y1,y2,,yn+1;f1y)=i=1nxiω1(y)(yyi)ω1(yi),

where ω1(y)=i=1n+1(yyi).

The following equality holds

(9) f1(y)=L(y1,y2,,yn+1;f1y)+R(f1,y)

where

R(f1,y)=[f1(θ1)](n+1)(n+1)!ω1(y)

and min{y,f(x1),,f(xn+1)}<θ1<max{y,f(x1),,f(xn+1)}.

It is also known that in the mentioned hypotheses concerning the derivability of f on I, the function f1 admits derivatives of any order k, 1kn+1 for all yF and the following equality holds [4], [8]:

(10) [f1(y)](k)=(2ki12)!(1)k+i11i2!i3!ik![f(x)]2k1(f(x)1!)i1(f′′(x)2!)i2(f(k)(x)k!)ik,k=1,n+1¯

where y=f(x) and the above sum extends over all nonnegative integer solutions of the system

{i2+2i3++(k1)ik=k1i1+i2++ik=k1.

From (3.1.2), neglecting R(f1,0) we obtain the following approximation for x¯

x¯L(y1,y2,,yn+1;f10).

Denoting

xn+2=L(y1,y2,,yn+1;f10),

we obtain

|xn+2x¯|=|[f1(θ1)](n+1)|(n+1)!|ω1(0)|,

where min{0,f(x1),,f(xn+1)}<θ1<max{0,f(x1),,f(xn+1)}.

It is clear that if xs,xs+1,,xs+n are n+1 distinct approximations of the solution x¯ of equation (2.1) then a new approximation xs+n+1 can be obtained as above, i.e.

(11) xs+n+1=L(ys,ys+1,,ys+n;f10)s=1,2,

with the error estimate given by

(12) |xs+n+1x¯|=|[f1(θs)](n+1)|(n+1)!i=0n|f(xs+i)|,s=1,2,

where θs belongs to the smallest open interval containing 0,f(xs),,f(xs+n).

If we replace in (3.1.5) |xs+n+1x¯|=|f(xs+n+1)||f(αs)|, we obtain for the sequence (f(xp))p0 the relations:

(13) |f(xs+n+1)|=|f(αs)||[f1(θs)](n+1)|(n+1)!i=0n|f(xs+i)|,

where αs belongs to the open interval determined by x¯ and xs+n+1.

Suppose that cs=|f(αs)||[f1(θs)](n+1)|(n+1)!,s𝐍, satisfies the hypotheses of Lemma 2.1 and that the sequence (f(xp))p0, converges to zero, where (xp)p0 is generated by (3.1.4). Then the convergence order of this sequence is equal to the positive solution of the equation:

tn+1tntn1t1=0

Considering the set of all equations of the above form for n1,n𝐍, and denoting by ωn+1 its corresponding positive solution it is known that the following relations hold [4]:

a)

2(n+1)n+2<ωn+1<2n=1,2,;

b)

ωn<ωn+1n=1,2,;

c)

limωn=2.

3.2 Hermite-type inverse interpolation

Consider in the following, besides the interpolation nodes (3.1.1), n+1 natural numbers a1,a2,,an+1, where ai1,i=1,n+1¯ and

a1+a2++an+1=m+1.

We shall suppose here too, for simplicity, that f is m+1 times differentiable on I. From this and from f(x)0 for all xI, it follows, by (3.1.3), that f1 is also m+1 times differentiable on F. Denoting yi=f(xi), i=1,n+1¯, the Hermite polynomial for the nodes yi,i=1,n+1¯, with multiplicity orders ai,i=1,n+1¯, has the following form:

H(y1,a1;y2,a2;;yn+1,an+1;f1y)=
=i=1n+1j=0ai1k=0aij1[f1(yi)](j)1k!j![(yyi)aiω1(y)]y=yi(k)ω1(y)(yyi)aijk

where

(14) ω1(y)=i=1n+1(yyi)ai.

If xs,xs+1,,xs+n are n+1 distinct approximations of the solution x¯ of the equation (2.1), then the next approximation xs+n+1 can be obtained as before in the following way:

(15) xs+n+1=H(ys,a1;;ys+n,an+1;f1y),s=1,2,

where, as in (3.2.1),

ωs(y)=i=ss+n(yyi)ai.

It can be easily seen that the following equality holds:

(16) |f(xs+n+1)|=|f(βs)||[f1(θs′′)](m+1)|(m+1)!i=0n|f(xs+i)|ai+1,s=1,2,,

where θs′′ belongs to the smallest open interval containing 0,ys,ys+1,,ys+n and βs belongs to the open interval determined by x¯ and xs+n+1.

If we suppose that cs=|f(βs)||[f1(θs′′)](m+1)|(m+1)!,s𝐍, verifies the hypotheses of Lemma 2.1 and, moreover, limsf(xs)=0, then it is clear that the convergence order of the method (3.2.2) is given by the positive solution of the equation

(17) tn+1an+1tnantn1a2ta1=0.

In the following we shall consider the following particular cases of (3.2.2):

For a1=a2==an+1=q, from (3.2.2) we obtain

(18) xs+n+1=H(ys,q;ys+1,q;;ys+n,q;f10),

method having the convergence order given by the positive solution of the equation

(19) tn+1qtnqtn1qtq=0

Let γn+1(q) denote the positive solution of equation (3.2.6). It is easy to prove that the following properties hold (see [7]):

a′′)

γn(q)<γn+1(q)n=1,2,;

b′′)

max{q,n+1n+2(q+1)}<γn+1(q)<q+1n=1,2,;

c′′)

limnγn(q)=q+1.

Taking n=0 in (3.2.2) we obtain again Chebyshev’s method, i.e.

(20) xs+1=xs[f1(ys)]1!f(xs)+[f1(ys)]′′2!f2(xs)++(1)m[f1(ys)](m)m!fm(xs),s=1,2,,

where ys=f(xs), the convergence order being m+1.

Concerning the positive solution of equation (3.2.4) we state the following lemma.

Lemma 3.2.0

The positive solution δn+1 of equation (3.2.4) verifies the relations:

(21) (m+1)m+1(n+1)(m+1)i=1n+1(i1)aiδn+11+max1in+1{ai},n=1,2,

Proof. Let

(22) α=(m+1)m+1(n+1)(m+1)i=1n+1(i1)ai.

It is sufficient to prove that Pn+1(α)0, where Pn+1(t)=tn+1an+1tna2ta1. We shall use for this the inequality between the arithmetic mean and the geometric mean, i.e.

i=1n+1αipii=1n+1pi(i=1n+1αipi)1i=1npi,αi>0,pi0,i=1,n+1¯,i=1n+1pi>0.

Using this inequality we obtain

Pn+1(α)=αn+1i=1n+1aiαi1=αn+1i=1n+1aiαi1i=1n+1aii=1n+1aiαn+1(i=1n+1ai)(i=1n+1α(i1)ai)1i=1n+1ai=αn+1(m+1)(i=1n+1α(i1)ai)1m+1==αn+1(m+1)(αi=1n+1(i1)ai)1m+1=αi=1n+1(i1)aim+1[αn+1i=1n+1(i1)aim+1(m+1)]=0,

i.e. Pn+1(α)0.

Remark 3.2.3

It can be easily seen that the number α given by (3.2.9) can be exprimed using Pn+1(1):

α=(m+1)m+1m(n+1)+Pn+1(1).

The second part of relations (3.2.8) follows easily from the inequality Pn+1(a)>0, where a=1+max1in+1{ai}.

3.3 Steffensen-type iterative methods

The convergence orders of methods (3.1.4), (3.2.2), respectively (3.2.5) can be improved if the interpolation nodes in the corresponding formulae are chosen in a special way. For this purpose we consider a continuous function φ:II, whose unique fixed point in the interval I is x¯. We also suppose that f and φ verify the equality

(23) f(φ(x))=g(x)f(x),for all xI,

where g:I𝐑, g(x)0 for all xI.

Let xsI be an approximation of the solution x¯. Denote us=xs, us+1=φ(us),us+2=φ(us+1),,us+n=φ(us+n1) and y¯s=f(us),y¯s+1=f(us+1),, y¯s+n=f(us+n).

Considering now as interpolation nodes the numbers y¯s,y¯s+1,,y¯s+n, by (3.1.4) we obtain

(24) xs+1=L(y¯s,y¯s+1,,y¯s+n;f10)s=0,1,,x1I,

and from (3.2.2) we have

(25) xs+1=H(y¯s,a1;y¯s+1,a2;;y¯s+n,an+1;f0),s=0,1,,x1I.

The iterative methods (3.3.2) and (3.3.3) are generalizations of the Steffensen’s method, which can be obtained from (3.3.2) for n=1 (see [4], [5]).

From (3.3.1) one obtains the following representations for y¯s+i,i=1,n¯:

y¯s+i=f(us+i)=ps,i1f(xs),i=1,2,,n,

where

ps,i1=j=ss+i1g(uj).

Considering (3.1.6) we obtain:

(26) |f(xs+1)|=|f(αs)||(f1(μs))(n+1)|(n+1)!i=1s+1|ps,i1||f(xs)|n+1,s=0,1,,

and, analogously, from (3.2.3) we get

(27) |f(xs+1)|=|f(βs)||[f1(μs)](m+1)|(m+1)!i=1n+1|ps,i1|ai|f(xs)|m+1.

In the relations (3.3.4) and (3.3.5), αs and βs are contained in the open interval determined by x¯ and xs+1 from (3.3.2) and (3.3.3) respectively and μs and μs belong to the smallest open interval containing 0,y¯s,y¯s+1,,y¯s+n from (3.3.2), respectively (3.3.3).

If we suppose that the sequences (us)s0 and (vs)s0 given by

us=|f(αs)||[f1(μs)](n+1)|(n+1)!i=1n+1|ps,i1|,vs=|f(βs)||[f1(μs)](m+1)|(m+1)!i=1n+1|ps,i1|ai

are bounded and inf{us}0, respectively inf{vs}0, then we clearly have that the convergence orders of methods (3.3.2), respectively (3.3.3) are equal to n+1, respectively m+1.

Remark 3.3.4

For the way of choosing the function φ with the mentioned properties see for example [5].

4 Optimal efficiency

We shall analyze in the following the efficiency index of each of the methods described and in the hypotheses adopted below we shall determine the optimal methods, i.e. those having the highest efficiency index.

As we have seen, the formulae for computing the derivatives of f1 have a complicated form and they depend on the successive derivatives of f. Though, in the case where the orders of the derivatives of f1 are low, the values of these derivatives are obtained by only a few elementary operations. Taking into account the generality of the problem we shall consider each computation of the values of any derivative of f1 by (3.1.3) as a single function evaluation. For similar reasons we shall also consider each computation of the inverse interpolatory polynomials as a single function evaluation.

As it will follow from our reasonings, the methods having the optimal efficiency index are generally the simple ones, using one or two interpolation nodes and the derivatives of f1 up to the second order.

Remark that in our case we can use for the efficiency index the relation (2.7).

4.1 Optimal Chebyshev-type methods

Observe that for passing from the s-th iteration step to the s+1, in method (3.2.7) must be performed the following evaluations:

f(xs),f(xs),,f(m)(xs),

i.e. m+1 values.

Then, by (3.1.3), we perform the following m function evaluations:

[f1(ys)],[f1(ys)]′′,,[f1(ys)](m),

where ys=f(xs). Finally, for the right hand expression of relation (3.2.7) we perform another function evaluation, so that 2(m+1) function evaluations must be performed.

By (2.7) the efficiency index of method (3.2.7) has the form

E(m)=(m+1)12(m+1),E:𝐍𝐑.

Considering the function h:(0,+)𝐑,h(t)=t12t, we observe that it attains its maximum at t=e, so that the maximum value of E is attained for m=2. We have proved the following result:

Theorem 4.1.4

Among the Chebyshev-type iterative methods having the form (3.2.7) the method with the highest efficiency index is the third order method, i.e.

(28) xs+1=xsf(xs)f(xs)12f′′(xs)f2(xs)[f(xs)]3,s=0,1,,x0I.

In the following table some approximate values of E are listed :

m     1   2   3   4   5
E(m) 1.1892 1.2009 1.1892 1.1746 1.1610

Table 4.1.1

We note that E(2)1.2009.

4.2 The efficiency of Lagrange-type methods

We shall study the methods of the form (3.1.4), for which the convergence order verifies a)c) from 3.1. Taking into account Remark 2.2, it can be easily seen that we can use relation (2.7) for the efficiency index of these methods. For each s+n+1 step,s2, in (3.1.4) in order to pass to the next step, only f(xs+n+1) must be evaluated, the other values from (3.1.4) being already computed. We have also another function evaluation in computing the right-hand side of relation (3.1.4). So there are needed two function evaluations. Taking into account that the convergence order ωn+1 of each method satisfies a)c), and denoting by En+1 the corresponding efficiency index, we have

En+1=ωn+112,n=1,2,;
En<En+1,n=2,3,

and

limEn=2.

We have proved:

Theorem 4.2.5

For the class of iterative methods of the form (3.1.4) the efficiency index is increasing with respect to the number of interpolation nodes, and we have the equality

limEn=2.

4.3 Optimal Hermite-type particular methods

We shall study the class of iterative methods of the form (3.2.5) for q>1. Taking into account the remarks from 4.2. it is clear that we can use again relation (2.7) for the efficiency index.

If xn+j is an approximation for the solution x¯ obtained by (3.2.5) then for passing to the following iteration step we need

f(xn+j),f(xn+j),,f(q1)(xn+j),

i.e. q function evaluations. Then, by (3.1.3) we must compute the derivatives of the inverse function [f(yn+j)1](i), i=1,q1¯, where yn+j=f(xn+j). Another function evaluation is needed for computing the right hand side of relation (3.2.5). We totally have 2q function evaluations, the other values in (3.2.5) being already computed.

By a′′)b′′) from 3.2. and denoting by E(γn+1(q),q) the efficiency of this method, we get:

(29) E(γn+1(q),q)>E(γn(q),q)n1,q>1;
(30) (max{q,n+1n+2(q+1)})12q<E(γn+1(q),q)<(q+1)12q,n1,q>1.

For a fixed q, by (4.3.1) it follows that the efficiency index is an increasing function with respect to n and

limE(γn(q),q)=(q+1)12q.

In the following we shall study E(γn(q),q) as a function of q>1 and n2, q,n𝐍.

By (4.3.2) we have

q12q<E(γn+1(q),q)<(q+1)12q, for qn+1

and

(31) [n+1n+2(q+1)]12q<E(γn+1(q),q)<(q+1)12q, for q<n+1.

For qn+1 consider the functions h:(0,+)𝐑, h(t)=t12t and l:(0,+)𝐑,l(t)=(t+1)12t.

Some elementary considerations show that h and l satisfy limt0h(t)=0,limth(t)=1, h is increasing on (0,e) and decreasing on (e,+) and limt0l(t)=e12,limtl(t)=1, l is decreasing on (0,). The maximum value of h is h(e)=e12e.

Let t¯ be the solution of the equation

(32) (t+1)12te12e=0.

It can be easily seen that t¯ exists and it is the unique solution for equation (4.3.4). For t>t¯, l(t)>e12e, so it is clear that the maximum value of E(γn+1(q),q) can be obtained for qt¯,q𝐍. It is easy to prove that t¯(4,5) and t¯4.76. Taking into account the properties of h and l it is clear that in order to determine the greatest value of E(γn+1(q),q) it will be sufficient to consider only those q𝐍 verifying 1<q4, and nq1.

Table 4.3.1. contains the approximate values of the efficiency indexes corresponding to these values of q and n.

q\n 1 2 3
2 1.2856
3 1.2487 1.2573
4 1.2175 1.2218 1.2226

Table 4.3.1

The highest value for the efficiency index is hence obtained for q=2 and n=1. We shall precise explicitly the method (3.2.5) for these values. For this purpose it is convenient to use the divided differences on multiple nodes. The following table contains the divided differences for the inverse function f1 on the nodes ys=f(xs), ys+1=f(xs+1) having the multiplicity orders 2.

f(x) x [u,v;f1] [u,v,w;f1] [u,v,w,z;f1]
ys xs
ys xs [ys,ys;f1]
ys+1 xs+1 [ys,ys+1;f1] [ys,ys,ys+1;f1]
ys+1 xs+1 [ys+1,ys+1;f1] [ys,ys+1,ys+1;f1] [ys,ys,ys+1ys+1;f1]

Table 4.3.2

Here [ys,ys;f1]=1f(xs),[ys+1,ys+1;f1]=1f(xs+1) and [ys,ys+1;f1]=1[xs,xs+1;f], and the other divided differences are computed using the well-known recurrence formula.

In this case the method has the following form:

(33) xs+2=xs[ys,ys;f1]ys+[ys,ys,ys+1;f1]ys2[ys,ys,ys+1,ys+1;f1]ys2ys+1,s=1,2,,x1,x2I.

The following theorem holds:

Theorem 4.3.6

Among the methods given by relation (3.2.5) for n1 and qn+1, the method with the highest efficiency index is given by (4.3.5), and corresponds to the case n=1 and q=2.

We shall analyze the case q<n+1. In this case the efficiency index verifies (4.3.3). We also consider, besides the function l already defined, the functions pn:(0,)𝐑,pn(t)=[n+1n+2(t+1)]12t, which satisfy the following properties limt0pn(t)=0,limtpn(t)=1 and

pn(t)=12[n+1n+2(t+1)]12ttt+1lnn+1n+2(t+1)t2.

It can be easily shown that the equation pn(t)=0 has a unique positive solution, denoted by τn. We also have   pn(t)>0 for t<τn and    pn(t)< 0 for t>τn, i.e. pn attains its maximum value at t=τn.

We also have that pn+1(τn)<0, showing that τn+1<τn for all n2. But since 1<q<n+1 it follows that we must examine only the cases when n2. Taking into account that τn is the solution of the equation pn(t)=0 we get that the maximum value of the function pn is equal to e12(τn+1).

Let vn:(0,+)𝐑, vn(t)=(t+1)12te12(τn+1). An elementary reasoning leads us to the following conclusions: vn is decreasing on (0,+); the equation vn(x)=0 has a unique solution μn on the interval (0,+) and μn+1<μn.

Since for t>μn, we have pn(τn)>pn(t), it follows that the values of n and q for which E attains its maximum must be searched in the set

(34) {q𝐍2q<min{n+1,μn}}.

Table 4.3.3. below contains the approximate values of the solutions τn and μn,the error being smaller than 102.

n τn μn
2 1.3816 3.6711
3 1.1201 2.8679
4 0.9566 2.3871
5 0.8436 2.0649
6 0.7601 1.8327

Table 4.3.3

Since q𝐍, we shall be interested only in the integer parts of the solutions μn.

From the above Table and by (4.3.6) we can see that E(γn+1(q),q) attains its maximum at q=2. Taking into account that E(γn(2),2)<E(γn+1(2),2) for n2 then we observe that E is increasing with respect to n.

Hence the following theorem holds:

Theorem 4.3.7

Taking q<n+1 in (3.2.5), the greatest values of the efficiency indexes E(γn+1(q),q), n2, are obtained for q=2. In this case the efficiency index is increasing with respect to n, and we have

limE(γn(2),2)=34.

4.4 Bounds for the efficiency index of the general Hermite-type methods

As it was shown in [6], the method (3.2.2) have the highest convergence order when the natural numbers a1,a2,,an+1verify the inequalities a1a2an+1. More exactly consider the equations:

(35) tn+1an+1tnantn1a2ta1=0;
(36) tn+1a1tna2tn1antan+1=0;
(37) tn+1ai1tnai2tn1aintain+1=0,

where ai0,i=1,n+1¯,i=1n+1ai>1 and (i1,i2,,in+1) is an arbitrary permutation of the numbers 1,2,,n+1.

If a,b,c are the corresponding positive solutions for equations (4.4.1)(4.4.3) then the following Lemma holds:

Lemma 4.4.0

If a1a2an+1 then 1<bca, i.e., among all equations of the form (4.4.3),equation (4.4.1) has the greatest positive root.

In the following we shall assume that the multiplicity orders of the interpolation nodes of the Hermite polynomial which leads to the method (3.2.2) satisfy

a1a2an+1.

From the above assumptions, at each iteration step there must be performed 2an+1 function evaluations. Denoting by E(δn+1) the efficiency index of (3.2.2) and taking into account Lemma 3.2.1 we get:

Theorem 4.4.9

If a1a2an+1 and δn+1 is the positive solution of (3.2.4) then the efficiency index of the method (3.2.2) satisfies

(38) (m+1)m+12[m(n+1)+Pn+1(1)]an+1E(δn+1)(1+an+1)12an+1.

Taking into account the properties of the function l given in 4.3. and that an+1>1, it follows that the expression (1+an+1)12an+1 attains its maximum value for an+1=2. Taking into account the inequalities from (4.4.4) the fact that (1+an+1)12an+1 attains its maximum value at an+1=2 do not imply the maximality of E(δn+1).

4.5 Optimal Steffensen-type methods

In the following we shall determine the optimal efficiency index for the class of iterative methods given by (3.3.3). First, we observe that at each iteration step s in (3.3.3), we must compute n values of the function φ, us+i=φ(us+i1), i=1,n¯, us=xs being an already computed approximation of the solution x¯.

We then compute y¯s+i=f(us+i),i=0,n¯, i.e. n+1 function evaluations. In order to compute the successive values of f and f1 at the nodes us+i,i=0,n¯ we need 2(mn) function evaluations. Finally, there is another function evaluation in computing the right hand side of (3.3.3). Totally there are 2(m+1) function evaluations.

If we denote by E(m) the efficiency index of (3.3.3) then

E(m)=(m+1)12(m+1),

which, taking into account the results from 4.1., attains its maximum at m=2.

Remark 4.5.5

If we take ai1 in (3.3.3), then method (3.3.2) is a particular case of (3.3.3), since for a1=a2==an+1=1 in (3.3.3) we get (3.3.2).

By the above remark, if m=2 then from a1+a2++an+1=3, it follows n2. Hence we have to analyze the following cases:

i)

a1+a2+a3=3, i.e. a1=a2=a3=1;

ii)

a1+a2=3,i.e. a1=1,a2=2or a1=2,a2=1;

iii)

a1=3.

i) For a1=a2=a3=1, by (3.3.2) we get the following method:

(39) xk+1=xkf(xk)[xk,φ(xk);f][xk,φ(xk),φ(φ(xk));f]f(xk)f(φ(xk))[xk,φ(xk);f][xk,φ(φ(xk));f][φ(xk),φ(φ(xk));f],k=0,1,,x0I.

ii) For a1=2,a2=1 we get the method

(40) xk+1=xkf(xk)f(xk)[xk,xk,φ(xk);f]f2(xk)f(xk)[xk,φ(xk);f]2,k=0,1,,x0I

and for a1=1,a2=2 we get

(41) xk+1=xkf(xk)[xk,φ(xk);f][xk,φ(xk),φ(xk);f]f(xk)f(φ(xk))[xk,φ(xk);f]2f(φ(xk)),k=0,1,,x0I.

iii) For a1=3 we get the method (4.1.1), i.e. the Chebyshev’s method of third order.

We have proved the following theorem:

Theorem 4.5.10

Among Steffensen-type iterative methods given by (3.3.3), the methods (4.5.1)(4.5.3) have the optimal efficiency index.

Remark 4.5.6

In the particular case when a1=a2==an+1=q the condition imposed to obtain an optimal method leads us to two possibilities, namely: q=3 and n=0, i.e. method (4.1.1) or q=1 and n=2, i.e. method (4.5.1).

References

  • [1] R. Brent, S. Winograd, F. Wolfe, Optimal Iterative Processes for Root-Finding, Numer. Math. 20 (1973), 327-341.
  • [2] Gh. Coman, Some Practical Approximation Methods for Nonlinear Equations, Mathematica - Revue d’Analyse Numérique et de Théorie de l’Approximation, Tome 11, No 1-2, (1982), 41-48.
  • [3] H.T. Kung, J.F. Traub, Optimal Order and Efficiency for Iterations with Two Evaluations, SIAM J. Numer. Anal., Vol. 13, No 1, (1976), 84-99.
  • [4] A.M. Ostrowski, Solution of Equations and Systems of Equations, Academic Press, New York and London, (1966).
  • [5] I. Pǎvǎloiu, Bilateral Approximations for the Solutions of Scalar Equations, Revue d’Analyse Numérique et de Théorie de l’Approximation, (23), 1, (1994), 95-100.
  • [6] I. Pǎvǎloiu, Optimal Problems Concerning Interpolation Methods of Solution of Equations, Publ. Inst. Math., 52 (66) (1992), 113-126.
  • [7] Traub J.F., Iterative methods for solution of equations, Prentice-Hall Inc., Englewood Cliffs, New Jersey, 1964.
  • [8] Turowicz B.A., Sur les derivées d’ordre superieur d’une fonction inverse, Ann. Polon. Math. 8 (1960), 265-269.
1997

Related Posts