On computational complexity in solving equations by interpolation methods

Abstract

 

Authors

Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)

Keywords

PDF

Scanned paper: on the journal website.

Latex version of the paper.

Cite this paper as:

I. Păvăloiu, On computational complexity in solving equations by interpolation methods, Rev. Anal. Numér. Théor. Approx., 24 (1995) no. 1, pp. 201-214.

About this paper

Print ISSN

1222-9024

Online ISSN

2457-8126

References

[1] Brent, R., Winograd, S. and Wolfe, Ph., Optimal Iterative Processes for Root-Finding. Numer. Math. 20(5) (1973), pp. 327-341.

[2] Casulli, V. and Trigiante, D., Sui procedimenti Iterativi Compositi. Calcolo, XIII. IV (1976), pp. 403-420.

[3] Coman, Gh., Some practical Approximation Methods for Nonlinear

Equations. Mathematica Revue d’Analyse Num'{e}rique et de Th'{e}orie de l’Approximation, 11, 1-2 (1982), pp. 41-48.

[4] Kacewitcz, B., An Integral-Interpolation Iterative Method for the Soltuion of Scalar Equations, Numer. Math. 26 (4), (1976), pp. 355-365.

[5] Kung, H.T. and Traub, J.F., Optimal Order and Efficiency for Iterations With Two Evaluations. SIAM, Numer Anal. 13, 1, (1976), pp. 84-99.

[6] Ostrowski, M.A., Soltuion of Equations and Systems of Equations. Academic Press. New York and London (1960).

[7] Păvăloiu I., The Soltuion of the Equations by Interpolation (Romanian) Ed. Dacia Cluj-Napoca (1981).

[8] Păvăloiu I., Optimal Problems Concerning Interpolation Methods of Solution of Equations. Publications de L’Institut Mathematique Beograd. 52, (66), (1992), pp. 113-126.

[9] Traub, J.F., Iterative Methods for the Solution of Equations. Prentice-Hall, New York (1964).

[10] Traub, J.F., Theory of Optimal algorithms, Preprint. Department of Computer Science Carnegie-Mellon University (1973), pp. 1-22.

[11] Traub, J.F. and Wozniakowski, H., Optimal Radius of Convergence of Interpolatory Iterations for Operator Equations, Aequationes Mathematicae, 21 (1980), pp. 159-172.

[12] Turowicz, A.B., Sur les derivees d’ordre suprieur d’une fonction inverse, Ann. Polon Math.8 (1960), pp. 265-269.

Paper (preprint) in HTML form

On computational complexity in Solving Equations by Interpolation Methods

On computational complexity in Solving Equations
by Interpolation Methods

Ion Păvăloiu
(Cluj-Napoca)

1. Introduction

Let I be an interval of the real axis and f:I a function. Consider the equation:

(1) f(x)=0,

supposed to have a solution x¯I. Also get g:II be a function whose fixed points from I coincide with the root of (1).

For solving (1) one can usually use an iterative method of the form:

(2) xs+1=g(xs),x0I,s=0,1,

More generally, if G:IkI is a function depending on k variables, whose restriction to the diagonal of the set Ik coincides with g, i.e.

(3) g(x)=G(x,x,,x),for all xI

then one can consider the following iterative method for solving equation (1):

(4) xk+s=G(xs,xs+1,,xs+k1),x0,x1,,xk1I,s=0,1,,

The convergence of the sequences (xn)n0 generated by (2) or (4) to a solution of equation (1) depends obviously on the properties of the functions f,g respectively G, and the amount of time necessary to obtain a suitable approximation for the solution x¯ is influenced both by the convergence order of the methods (2), resp. (4) and by the amount of elementary operations that must be performed at each iteration step. This last aspect belongs to a chapter of the calculus theory and practice, chapter concerning the computational complexity.

Many authors ([1], [2], [3], [5], [6], [9], [10], [11]), who studied the computational complexity of the iteration processes, have defined different notions, as: the efficiency of a method, the efficiency index of a method or the cost of a method, which they have quantitatively expressed by different scalar magnitudes.

Throughout this paper we shall adopt the following definition for the convergence order of an iteration method:

Definition 1.1.

The real number p1 is called the convergence order of the sequence (xn)n0 generated by an iterative method if the following limit exists and is not zero:

(5) limn|xn+1x¯||xnx¯|p=a0.

where x¯ is the solution of equation (1).

Concerning the calculus complexity, using the convergence order we can define now the following notion:

Definition 1.2.

[6]. The number I is called the efficiency index of the method (2) or (4), if the following limit exists and is finite:

(6) limn(ln|xn+1x¯|ln|xnx¯|p)1/mn=I,

where mn represents the number of function evaluations that must be performed when passing from the step n to the step n+1.

If we suppose that mn is the same for all n, and take into account that (6) has an asymptotical character, then there results for I the following expression:

(7) I=I(p,m)=p1m.

In the following we shall study certain classes of iteration methods,namely the methods obtained by interpolation, among which we shall select those for which the efficiency index given by (7) is optimal, i.e. the greatest. For this purpose in the next section we shall briefly recall the classes of methods that we want to study.

2. Interpolation Iterative Methods

2.1. Lagrange’s inverse interpolation polynomial

Let I be an interval and f:I a function. Denote by F=f(I) the set of all values of f for xI. Suppose that f is one-to-one, i.e. there exists the inverse function f1;FI. Consider in I,n+1 interpolation nodes:

(8) x1,x2,,xn+1,with xixj;for ij;i,j=1,n+1.¯

Suppose that equation (1) has the unique solution x¯I.

Obviously,

(9) x¯=f1(0),

and so the problem of approximation the solution x¯ reduces to the approximation of f1(0).

A simple and efficient approximation method for the functions is given by the interpolating approximation.

Denote by:

(10) y1,y2,,yn+1,yi=f(xi),i=1,n+1,¯

the values of the f on the nodes xi from (8).

The Lagrange interpolation polynomial corresponding to the function f1 on the nodes from (10) (taking into account that yiyj,ij;i,j=1,n+1¯) has the form:

(11) L(y1,y2,,yn+1;f1|y)=i=1n+1xiω1(y)(yyi)ω1(yi),

where ω1(y)=i=1n+1(yyi).

If we suppose that the function f has derivatives up to the order k,k and f(x)0 for all xI, then we have the following formula for the computation of the k-th derivative of f1 at the point y=f(x),xI ([7], [12]):

(12) [f1(y)](k)=(2ki12)(1)k+i11i2!i3!ik!f(x)2k1(f(x)1!)i1(f(x)(k)k!)ik,

where the above sum extends to all nonnegative integer numbers, solutions of the system:

(13) i2+2i3++(k1)ik =k1;
i1+i2++ik =k1.

If we suppose that f admits derivatives up to the order n+1 on the interval I and the n+1-th derivative is bounded on I, then by (12) we obtain,

(14) x¯=f1(0)=L(y1,y2,,yn+1;f1|0)+[f1(c)](n+1)(n+1)!ω1(0),

where c is a point belonging to the smallest interval containing 0,y1,y2,,yn+1 and ω1(0)=(1)n+1f(x1)f(x2)f(xn+1).

Denote by xn+2 the number:

(15) xn+2=L(y1,y2,,yn+1;f1|0),

and by (14) we get:

(16) |x¯xn+2|Mn+1(n+1)!|f(x1)||f(x2)||f(xn+1)|,

from which we see that if x1,x2,,xn+1 are chosen in a neighbourhood of x¯ such that |f(xi)|<1,i=1,n+1¯, then xn+1 can be considered as a new approximation for x¯. We have denoted in inequality (16), Mn+1=supyF|[f1(y)](n+1)|.

Let now xk,xk+1,,xk+nI be n+1 approximations for x¯. Then the Lagrange polynomial corresponding to the function f1 on the nodes yi=f(xi),i=k,n+k¯ has the form:

(17) L(yk,yk+1,,yk+n;f1|y)=i=kk+nxiωk(y)(yyi)ωk(yi)

where ωk(y)=i=kn+k(yyi). From this relation, for y=0, we obtain a new approximation for x¯, namely

(18) xn+k+1=L(yk,yk+1,,yk+n;f1|0),k=1,2,

which satisfies the delimitation

(19) |x¯xn+k+1|Mn+1(n+1)!|f(xk)||f(xk+1)||f(xk+n)|.

It is well known that the iterative method given by (18) has the convergence order θn+1, which is the unique positive root of the equation, [6]:

(20) tn+1tntn1t1=0.

It is also known (see [6]) that θn+1 verifies:

(21) 2(n+1)n+2<θn+1<2

and

(22) θn<θn+1;limnθn=2,for all n1.
Remark 2.1.

In the successive computation of the elements of the sequence (xn)n0 generated by (18) it is necessary to compute at each step k the values ωk(0) and ωk(yi),i=k,k+1,,k+n.

We observe that practically there exists a connection both between ωk(0) and ωk+1(0) and between ωk(yi) and ωk+1(yi).

Indeed:

ωk(y)=i=kn+k(yyi)

and

ωk+1(y)=i=k+1n+k+1(yyi)

hence we get

(23) ωk+1(y)=ωk(y)(yyn+k+1)yyk,

which for y=0 yields

(24) ωk+1(0)=ωk(0)yn+k+1yk

From (23) we obtain:

(25) ωk+1(y)=[ωk(y)(yyn+k+1)+ωk(y)](yyk)ωk(y)(yyn+k+1)(yyk)2,

which gives us the following recurrence formula:

(26) ωk+1(yi)={ωk(yi)(yiyn+k+1)yiyk,i=k+1,k+2,,k+nωk(yi)yiyk,i=k+n+1

Recurrence formulae (24) and (26) hold for all k=1,2,,

2.2. Hermite Inverse Interpolating Polynomial

We consider, besides the interpolatory nodes (8), the natural numbers a1,a2,,an+1 with ai1i=1,n+1¯ and

(27) a1+a2++an+1=m+1,m.

Suppose that f admits derivatives up to the order m+1 on the interval I and f0. Then, by (12) it follows that the function f1 also admits derivatives up to the order m+1. The Hermite polynomial of degree m associated to the function f1 on the nodes yi=f(xi),i=1,n+1¯, assuming that f(x)0 for all xI, is:

(28) H(y1;a1,y2;a2,,yn+1;an+1;f1|y)=
=i=1n+1j=0ai1k=0aij1[f1(yi)](j)1k!j![(yyi)aiω1(y)]y=yi(k)ω1(y)(yyi)aijk.

where

(29) ω1(y)=i=1n+1(yyi)ai

This polynomial satisfies:

(30) H(j)(y1;a1,y2;a2,,yn+1;an+1;f1|yi)=[f1(yi)](j)

for all j=0,1,,ai1;i=1,2,,n+1.

As in 2.1 we obtain from (28) the following iterative method for solving equation (8):

(31) xn+k+1=H(yk;a1,yK+1;a2,,yk+n;an+1;f1|0),k=1,2,,

where in the polynomial H,ωk=i=kn+k(yyi)ai.

Using the differentiability assumptions for f, we obtain:

(32) |xn+k+nx¯|Mm+1(m+1)!|f(yk)|a1|f(yk+1)|a2|f(yk+n)|an+1

where Mm+1=supyF|[f1(y)](m+1)|.

It is well known that the convergence order of (31) is given by the positive root ωn+1 of the equation:

(33) tn+1an+1tnantn1a2ta1=0.

In the particular cases when a1=a2==an+1 we have iterative method:

(34) xn+k+1=H(yk;q,yk+1;q,,yk+n;q;f1|0),

and when n=0 we get the Chebyshev iterative method of order m+1:

(35) xk+1=xk[f1(y)]1!f(xk)++(1)m[f1(yk)]m!(m)fm(xk),k=1,2,

which has the convergence order p=m+1.

For the method (34), by (33) it follows that the convergence order is given by the positive root ωn+1 of the equation:

(36) tn+1qtnqtn1qtq=0.

which satisfies:

(37) ωn<ωn+1,n=1,2,,
(38) max{q,n+1n+2(q+1)}<ωn+1<q+1;n=1,2,,
(39) limnωn+1=q+1.

3. The Efficiency Index of the Chebyshev Method of Order m+1

In the following we shall make the assumptions:

  • a)

    Consider as a function evaluation, the evaluation of the derivatives [f1(y)](k), assuming f(k)(x),k=1,m¯ as having been computed.

  • b)

    Consider as a function evaluation, the evaluation of the right hand side of expression (35) assuming f(x) and [f(1)(y)](k),k=1,m¯ as having been computed.

  • c)

    Consider as a function evaluation the evaluation of the function f or of any of its derivatives.

In this hypothesis, it is necessary for an iteration step with method (35) to compute firstly the values of the functions: f,f,,f(m) at the point xk, altogether m+1 function for the calculus of the values of the successive derivatives of f1, by (12). If we take into account that for evaluating the right hand side expression from relation (35) is computed another function value, we have altogether 2(m+1) function evaluations at each iteration step.

Using definition (7) for the efficiency index, method (35) has the following index:

(40) I(m+1,2(m+1))=(m+1)12(m+1)

(We have taken into account that the convergence order is m+1).

We are searching for the maximum value of the index I from (40), for m.

For this purpose we consider the auxiliary function φ:(0,)+φ(t)=t12t and we note that: limt0φ(t)=0, limtφ(t)=1, φ is increasing for t(0,e) and decreasing for t(e,);t=e is a maximum point for the function φ.

For t, the function φ attains its maximum for t=3, so I(m+1,2(m+1)) attains its maximum form m=2.

So, the following holds:

Theorem 3.1.

In the above assumptions a)–c) among all the Chebyshev iterative methods of the form (35) the method with the greatest efficiency index is the one of 3rd order, namely

(41) xk+1=xkf(xk)f(xk)12f′′(xk)f2(xk)[f(xk)]3,k=0,1,,x0I.

In the following table are shown the approximations of the efficiency index of Chebyshev methods for some values of m.

m 2 3 4 5 6
I(m+1,2(m+1)) 1.1892 1.2009 1.1892 1.1746 1.1610

We see that I(3,6)1.2009.

4. The Efficiency Index for the Lagrange-Hermite Methods

In the following we shall study the case when n1, i.e. when number of nodes is greater than one. In the beginning we shall take the method given by (18) for which, if we take into account the assumptions a) - c) and neglect the first step, we get that at each iteration step we have first one function evaluation, namely the value of the function f at xn+k, and then we have another function evaluation, for the right hand side of relation (18), hence altogether two function evaluations. Recalling that the convergence order for (18) is θn+1, which satisfies (21) and (22) we get for the efficiency index the relation

(42) I(θn+1,2)=θn+1.

By (22) we have that I(θn,2)<I(θn+1,2) for all n1.

So we conclude:

Theorem 4.1.

If the assumptions a)–c) hold, for the Lagrange methods given by (17) the efficiency index is increasing with respect to the number of interpolation nodes and

limnI(θn,2)=2.

Now we study the efficiency index for the methods given by (34), for which the convergence order ωn+1 verifies (36)–(39). Obviously, we suppose that q>1,q=1 in (34) giving (18).

There are two aspects that must be considered: the efficiency with respect to the number of interpolation nodes, when their multiplicity order q is kept fixed and, on the second hand, the efficiency with respect to the multiplicity order q for fixed n,n1.

We again suppose that assumptions a)–c) hold. So from the right hand side of (34), at each iteration step, excepting the first one, we have the following function evaluations: we compute f(xn+k),f(xn+k),,f(q1)(xn+k), i.e. q function evaluations, and then by (12) we compute [f1(yn+k)], [f1(yn+k)]′′,,[f1(yn+k)](q1) i.e., q1 function evaluations, and finally we compute the right hand side of (34). so altogether, 2q function evaluations.

Using (37)–(39) we get:

(43) I(ωn+1,2q)>I(ωn,2q),for all n1,q>1,

and

(44) (max{q,n+1n+2(q+1)})12q<I(ωn+1,2q)<(q+1)12q.

for all n1,q>1.

For a fixed q by (43) we get that the efficiency index is increasing as a function of n, and by (44)

limnI(ωn+1,2q)=(1+q)12q.

From (44) we also obtain:

(45) q12q<I(ωn+1,2q)<(q+1)12q,for qn+1

and

(46) [n+1n+2(q+1)]12q<I(ωn+1,2q)<(q+1)12q,,for q<n+1.

A. For the first case, when qn+1 we consider the auxiliary functions φ,ψ;(0,)+ given by

φ(t)=t12t

and

ψ(t)=(1+t)12t.

As we have seen before, φ satisfies:

limt0t>0φ(t)=0,limtφ(t)=1,

is increasing on (0,e) and decreasing on (e,+), so at t=e attains its maximum.

One can establish the following relations for ψ:

limt0ψ(t)=e;limtψ(t)=1

and ψ is decreasing on (0,+).

Recalling that φ attains its maximum value at t=e, let t¯ be the solution of the equation

(47) (t+1)12te12e=0.

then for t>t¯ we have (t+1)12t<e12e, hence, by (45), we obtain that the values of q for which I(ωn+1,2q) attains its maximum lie in the set {q;1<qt¯}. One can easily prove that t¯(46) so we shall study the cases q=2,q=3, and q=4. Since qn we study 1) q=2,n=1;2) q=3,n=1;q=3,n=2 and 3) q=4,n=1;n=2,n=3.

1) The corresponding equation from (36) for q=2,n=1, is t22t2=0, with the positive solution ω2=1+3. So I(ω2,4)=1+341,2856

2) The convergence orders corresponding for this case are the solutions of the equations t23t3=0 for n=1,q=3 respectively t33t23t3=0 for n=2, q=3.

We obtain I(ω2,6)1.2487, respectively I(ω3,6)1.2573

3) The corresponding equations give us:

I(ω2,8)1,2175;I(ω3,8)1,2218 and I(ω4,8)1,2226

So the greatest efficiency index when qn+1 is obtained for n=1,q=2 i.e I(ω2,4)=1+34.

B. Let q<n+1, so (46) holds. We shall again consider two auxiliary functions

φ,ψ: (0,+)+,
φ(t) =[n+1n+2(t+1)]12t,ψ(t)=(t+1)12t

which possess the following properties: limt0φ(t)=0;limtφ(t)=1;

φ(t)=12[n+1n+2(t+1)]12ttt+1lnn+1n+2(t+1)t2,

and one can easily prove that: the equation φ(t)=0 has a unique positive solution, denoted by τn,φ(t)<0 for t>τn and φ(t)>0 for t(0,τn), i.e.φ attains its maximum value at t=τn.

Taking into account the properties of ψ one can see that the equations

(48) (t+1)12te12(τn+1)=0,n=2,3,

have a unique positive solution μn for each n2.

In the table below we give the approximative values μn and τn for n[2,10].

n τn μn
2 1.3816 3.6711
3 1.1201 2.8679
4 0.9566 2.3871
5 0.8436 2.0649
6 0.7601 1.8327
7 0.6955 1.6566
8 0.6438 1.5180
9 0.6013 1.4056
10 0.5656 1.3125

An elementary reasoning proves that τn and μn are decreasing functions of n, n2, as we can see in the above table.

If t>μn then φ(τn)>φ(t) so the optimal values for q must lie in the set {q;2q<max{n+1,μn}}.

It can be shown that for n6,μn<2 and for n[2,5], 2<μn<4. It follows that the only suitable value for q is q=2. In this case we get that I(ωn,4)<I(ωn+1,4), n2, i.e. the efficiency index increases with n, but anyway the best results hold for q=2.

5. Margins for the Efficiency Index in the Case of Lagrange-Hermite Methods

We end this note by indicating, in the above assumptions, left and right margins of the efficiency index of the Lagrange-Hermite methods.

In this respect we shall first establish an inequality that will give left margins for the positive roots of equations (33).

Let

(49) P(t)=tn+1an+1tnantn1a1=0,

where a1+a2++an+1=m+1,aifor i=1,n+1¯.

The following Lemma holds.

Lemma 5.1.

The positive root ωn+1 of (49) satisfies:

(50) ωn+1[m+1]m+1(n+1)(m+1)i=1m+1(i1)ai
Proof.

Let

α=[m+1]m+1(n+1)(m+1)i=1m+1(i1)ai

It will suffice to show P(α)0, using the inequality between the weighted arithmetic and geometric mean, i.e.

(51) i=1n+1αipii=1n+1pi(i=1n+1aipi)1i=1n+1pi

We get

P(α) =αn+1i=1n+1aiαi1=αn+1i=1n+1aiαi1i=1n+1aii=1n+1ai
αn+1(i=1n+1ai)[i=1n+1α(i1)ai]1i=1n+1ai=αn+1(m+1)αi=1n+1(i1)aim+1
=αi=1n+1(i1)aim+1[αn+1i=1n+1(i1)aim+1(m+1)]=0,

i.e. P(α)0.

We also get ωn+1a+1, where a=max1in+1{ai}.

In the hypotheses of 3., using (31) in generating the sequence (xn)n0 then the number of function evaluations at each step is 2(m+1)n.

The efficiency index (31) then satisfies

α12(m+1)nI(ωn+1,2(m+1)n)(a+1)12(m+1)n,

with α and a specified above. ∎

References

  • [1] Brent, R., Winograd, S. and Wolfe, Ph., Optimal iterative processes for root-finding. Numer Math. 20 (5) (1973), 327–341.
  • [2] Casulli, V. and Trigiante, D., Sui procedimenti iterativi compositi. Calcolo, XIII, IV (1976), pp. 403–420.
  • [3] Coman, Gh., margin: available soon,
    refresh and click here
    Some practical approximation methods for nonlinear equations. Anal. Numér. Théor. Approx., 11, 1–2 (1982), pp. 41–48.
  • [4] Kacewitcz, B., An integral-interpolation iterative method for the solution of scalar equations, Numer. Math. 26 (4), (1976), pp. 355–365.
  • [5] Kung, H. T. and Traub, J. F., Optimal order and efficiency for iterations with two evaluations. SIAM. J. Numer Anal. 13, 1, (1976), pp. 84–99.
  • [6] Ostrowski, M. A., Solution of Equations and Systems of Equations. Academic Press. New York and London (1960).
  • [7] Păvăloiu, I., margin: available soon,
    refresh and click here
    The Solution of the Equations by Interpolation (Romanian) Ed. Dacia Cluj-Napoca (1981).
  • [8] margin: clickable Păvăloiu, I., Optimal problems concerning interpolation methods of solution of equa- tions, Publications de L’Institut Mathématique Beograd. 52, (66), (1992), pp. 113–126.
  • [9] Traub, J. F., Iterative Methods for the Solution of Equations. Prentice-Hall, New York (1964).
  • [10] Traub, J. F., Theory of Optimal Algorithms, Preprint Department of Computer Science Carnegie-Mellon University (1973) pp. 1–22.
  • [11] Traub, J, F. and Wozniakowski, H., Optimal radius of convergence of interpolatory iterations for operator equations. Aequationes Mathematicae, 21 (1980), pp. 159–172.
  • [12] Turowicz, A. B., Sur les dérivées d’ordre supérieur d’une fonction inverse, Ann. Polon Math. 8 (1960), pp. 265–269.
  • [13]

Received 1 X 1994

Academia Română

Institutul de Calcul

”Tiberiu Popoviciu”

P.O. Box 68

3400 Cluj-Napoca 1

România

1995

Related Posts