On some Aitken-Steffensen-Halley-type method for approximating the roots of scalar equations

Abstract

We extend the Aitken-Steffensen method to the Halley transformation. Under some rather simple assumptions we obtain error bounds for each iteration step; moreover, the convergence order of the iterates is 3, i.e. higher than for the Aitken-Steffensen case.

Author

Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)

Keywords

nonlinear equations in R; Aitken-Steffensen-Halley method.

PDF

PDF-LaTeX file on the journal website.

Cite this paper as:

I. Păvăloiu, On some Aitken-Steffensen-Halley-type method for approximating the roots of scalar equations, Rev. Anal. Numér. Théor. Approx., 30 (2001) no. 2, pp. 207-212.

About this paper

Print ISSN

1222-9024

Online ISSN

2457-8126

References

[1] Argyros, I.K. On the convergence of some projection methods with perturbations, J. Comp. Appl. Math. 36, (1991), 255–258.

[2] Argyros, I.K. On an application of the Zincenko method to the approximation of implicit functions, Z.A.A. 10, 3, (1991), 391– 396.

[3] Argyros, I.K. and Szidarovszky, F. The Theory and Application of Iteration Methods, C.R.C. Press, Inc., Boca Raton, Florida, 1993.

[4] Catinas, E. On some iterative methods for solving nonlinear equations, Revue d’analyse Numerique et de theorie de l’approximation, 23, 1, (1994), 47–53.

[5] Kantorovich, L.V. The method of successive approximation for functional equations, Acta Math. 71 (1939), 63–97.

[6] Pavaloiu, I. Sur une generalisation de la methode de Steffensen, Revue d’analyse Numerique et de theorie de l’approximation, 21, 1, (1992), 59–65.

[7] Pavaloiu, I. Bilateral approximations for the solutions of scalar equations, Revue d’analyse numerique et de theorie de l’approximation, 23, 1, (1994), 95–100.

Paper (preprint) in HTML form

ON SOME AITKEN-STEFFENSEN-HALLEY-TYPE METHODS FOR APPROXIMATING THE ROOTS OF SCALAR EQUATIONS

ON SOME AITKEN-STEFFENSEN-HALLEY-TYPE METHODS
FOR APPROXIMATING THE ROOTS OF SCALAR EQUATIONS

ION PĂVĂLOIU
Abstract.

In this note we extend the Aitken-Steffensen method to the Halley transformation. Under some rather simple assumptions we obtain error bounds for each iteration step; moreover, the convergence order of the iterates is 3, i.e. higher than for the Aitken-Steffensen case.

1991 Mathematics Subject Classification:
65H05
”T. Popoviciu” Institute of Numerical Analysis, P.O. Box 68-1, 3400 Cluj–Napoca, Romania, e-mail: pavaloiu@ictp-acad.math.ubbcluj.ro.

1. Introduction

Let f:[a,b], a,b, a<b and suppose that fC4[a,b], and f(x)>0, x[a,b]. Consider the function h:[a,b] given by

h(x)=f(x)f(x).

As it was shown in [2], the Halley method for solving:

(1) f(x)=0,

is given by

(2) xn+1=xnh(xn)h(xn),n=0,1,,x0[a,b].

This sequence is in fact generated by the Newton method for solving h(x)=0.

The first and second order derivatives of h are given by

(3) h(x) =2(f(x))2f′′(x)f(x)2(f(x))3/2
(4) h′′(x) =3(f′′(x))22f′′′(x)f(x)4(f(x))5/2f(x)

which yield the following equalities for a solution x¯ of (1):

(5) h(x¯) =(f(x¯))1/2,and
(6) h′′(x¯) =0.

Relation (6) ensures the convergence order 3 for the sequence (xk)k0.

In the papers [2][8] and [12] there are studied the convergence and the convergence order of some sequences generated by some interpolatory methods applied to equation h(x)=0.

We shall consider other two equations equivalent to (1) of the form:

(7) xφ1(x) =0,and
(8) xφ2(x) =0

The Aitken method for solving h(x)=0 is given by the iteration

(9) xn+1=φ1(xn)h(φ1(xn))[φ1(xn),φ2(xn);h],n=0,1,,x0[a,b].

In this note we shall study the convergence of these iterates. We shall show that the functions φ1 and φ2 may be chosen in order to obtain bilateral approximations at each iteration step; this fact allows the control of the errors. On the other hand, the convergence order of (xn)n0 given by (9) is at least equal to 3.

Hypotheses fC4[a,b] and f(x)>0, x[a,b] imply, taking into account (5), that there exist α,β, aα<x¯<βb such that h(x)>0,x[α,β].

2. ERROR EVALUATION AND LOCAL CONVERGENCE

Consider the interval [α,β] given above. We shall make the following assumptions on φ1 and φ2:

  • i.

    the function fC4[a,b];

  • ii.

    equation (1) has the solution x¯[a,b];

  • iii.

    the inequality f(x)>0 holds for x[α,β];

  • iv.

    the function φ1verifies the relation 0<[x,y;φ1]<1 for all x,y[α,β], where [x,y;φ] denotes the first order divided difference of φ on x and y;

  • v.

    the function φ2 verifies the relations 1<[x,y;φ2]<0 for all x,y[α,β].

We can state the following result:

Theorem 2.1.

Assume that i–v hold, and for some x0[α,β] sufficiently close to x¯ we have φ1(x0),φ2(x0)[α,β]. Then the following relations hold:

  • j.

    the sequences (xn)n0,(φ1(xn))n0 and (φ2(xn))n0 converge to x¯;

  • jj.

    for any n=0,1,, one has

    |x¯xn+1|max{|xn+1φ1(xn)|,|xn+1φ2(xn)|};
  • jjj.

    there exists k, k>0,which does not depend on n such that

    |xn+1x¯|k|xnx¯|3,n=0,1,
Proof.

By φ1(x0),φ2(x0)[α,β] it obviously follows that h(φ1(x0))>0 and h(φ2(x0))>0. Denote by I0 the interval having the extremities φ1(x0) and φ2(x0). We notice, taking into account the mean formula, that

[φ1(x0),φ2(x0);h]>0.

When x0<x¯, by iv. and x¯=φ1(x¯) it follows φ1(x0)<x¯. Analogously, φ1(x0)>x¯ for x0>x¯. Taking into account v. and x¯=φ2(x¯) we get φ2(x0)>x¯ for x0<x¯ and φ2(x0)<x¯ for x0>x¯. It is obvious that in both situations x¯I. It can be easily seen that for all n=0,1, we have

φ1(xn)h(φ1(xn))[φ1(xn),φ2(xn);h]=φ2(xn)h(φ2(xn))[φ1(xn),φ2(xn);h],

which, for n=0 imply x1>φ1(x0) and x1<φ2(x0) if x0<x¯, respectively x1<φ1(x0) and x1>φ2(x0) if x0>x¯, i.e., x1intI0. It is clear now that, analogously, x1<φ1(x1)<x¯<φ2(x1) if x1<x¯ or x1>φ1(x1)>x¯>φ2(x1) if x1>x¯. Denoting by I1 the interval determined by φ1(x1) and φ2(x1) then

I1I0,

and the element x2 constructed by (9) satisfies x2,x¯I1.

Repeating the above reason we get xn+1,x¯In, the interval being determined by φ1(xn),φ2(xn), and also that

In+1In.

and x¯In+1. It is clear now that jj. holds. In order to obtain jjj. we shall use the identity

h(x¯)= h(φ1(xn))+[φ1(xn),φ2(xn);h](x¯φ1(xn))
+[x¯,φ1(xn),φ2(xn);h](x¯φ1(xn))(x¯φ2(xn)),

which, together with (9) and h(x¯)=0, imply

x¯xn+1=[x¯,φ1(xn),φ2(xn);h][φ1(xn),φ2(xn);h](x¯φ1(xn))(x¯φ2(xn)).

For the difference x¯φ1(xn), by iv. one gets

x¯φ1(xn)=[x¯,xn;φ1](x¯xn),

i.e.,

x¯φ1(xn)<|x¯xn|.

Analogously, by v. we get

|x¯φ2(xn)|<|x¯xn|.

The mean formula for divided differences implies

[x¯,φ1(xn),φ2(xn);h] =12h′′(ξn),with ξnIn,and
[φ1(xn),φ2(xn);h] =h(ηn),ηnIn.

For h′′(ξn) we have

|h′′(ξn)|=|h′′(ξn)h′′(x¯)|=|h′′′(θn)||x¯ξn|.

Since ξnIn it follows

|x¯ξn|<|x¯xn|.

Denoting m1=infx[α,β]|h(x)|, M3=supx[α,β]|h′′′(x)|, the above relations lead to

|x¯xn+1|M32m1|x¯xn|3,n=0,1,,

i.e., jjj. for k=M32m1.

Since the initial approximation x0 was supposed sufficiently close to the solution x¯, then

M32m1|x¯x0|<1

implies, together with jjj., statement j. ∎

Remark 2.1.

Supposing that f′′(x)0 for all x[a,b], and if instead of iii. we assume that f(x)>0, then obviously f(x)<0 for ax<x¯, and so in Theorem 2.1 we may take α=a. From the above conditions it follows that h(x)>0 for x[a,x¯], and since f(x¯)=0, one gets β>x¯.

3. DETERMINING THE FUNCTIONS φ1 AND φ2

Under reasonable hypotheses on f, we shall show that there exist two classes of functions among we can choose the functions φ1 and φ2 such that hypotheses iv. and v. to be satisfied.

Besides assumptions i.–iii. on f, we shall suppose that f is strictly convex, i.e., f′′(x)>0, x[a,b]. This condition implies that f is increasing on [a,b]. If, moreover, f(x)<2λ, with 0<λfr(a), then we may consider the functions

(10) φ1(x) =xf(x)μand
(11) φ2(x) =xf(x)λ

where μ may be taken as any real number greater than the left derivative of f at b, fl(b).

In the following we shall show that the functions φ1(x) and φ2(x) chosen above obey iv. and v. The derivatives of these functions are given by

φ1(x)=1f(x)μ, resp. φ1(x)=1f(x)λ.

Obviously, 0φ1(x)<1. Also, the monotonicity of f implies φ2(x)<0, while f(x)<2λ implies 1<φ2(x).

Taking into account Remark 2.1, under the above hypotheses it is obvious that if x0<x¯, then condition φ1(x0)[α,β] from Theorem 2.1 is obviously satisfied. Indeed, this fact follows from φ1(x)<1, since φ1(x0)x¯=φ1(x0)φ1(x¯)=φ1(ξ)(x0x¯)<0,x0<θ<x¯, i.e., φ1(x0)<x¯.

On the other hand, |φ1(x0)x¯|<|x0x¯|, and so the relations x0<φ2(x0)<x¯ hold. The hypothesis φ2(x0)[α,β] must be kept.

References

  • [1] A. Ben-Israel, Newton’s method with modified functions, Contemp. Math., 204, pp. 39–50, 1997.
  • [2] G.H. Brown, Jr., On Halley’s variation of Newton’s method, Amer. Math. Monthly, 84, pp. 726–728, 1977.
  • [3] V. Candela and A. Marquina, Recurrence relations for rational cubic methods I: The Halley’s method, Computing, 44, pp. 169–184, 1990.
  • [4] G. Deslauries and S. Dubuc, Le calcul de la racine cubique selon Héron, El. Math., 51, pp. 28–34, 1996.
  • [5] W.F. Ford and J.A. Pennline, Accelerated convergence in Newton method, SIAM Rev., 38, pp. 658–659, 1996.
  • [6] J. Gerlach, Accelerated convergence in Newton’s method, SIAM Rev., 36, pp. 272–276, 1994.
  • [7] margin: clickable D. Luca and I. Păvăloiu, On the Heron’s method for the approximation of the cubic root of a real number, Rev. Anal. Numér. Théor. Approx., 28, pp. 103–108, 1997.
  • [8] A. Melman, Geometry and convergence of Euler’s and Halley’s methods, SIAM Rev., 39, pp. 728–735, 1997.
  • [9] A.M. Ostrowski, The Solution of Equations and Systems of Equations, Academic Press, New York–London, 1960.
  • [10] margin: clickable I. Păvăloiu, On the monotonicity of the sequences of approximations obtained by Stef- fensen method, Mathematica (Cluj), 35 (58), pp. 171–76, 1993.
  • [11] margin: clickable I. Păvăloiu, Approximation of the roots of equations by Aitken–Steffensen-type mono- tonic sequences, Calcolo, 32, pp. 69–82, 1995.
  • [12] margin: clickable I. Păvăloiu, On a Halley-Steffensen method for approximating the solutions of scalar equations, Rev. Anal. Numér. Théor. Approx., 30, 2001 no. 1, pp. 69–74.
  • [13] T. Popoviciu, margin: available soon,
    refresh and click here
    Sur la délimitation de l’erreur dans l’approximation des racines d’une équation par interpolation linéaire ou quadratique, Rev. Roumaine Math. Pures Appl., 13, pp. 75–78, 1968.
  • [14]

Received June 13, 2000.

2001

Related Posts