Abstract
We study the conditions under which the well-known Aitken-Steffensen method for solving equations leads to monotonic sequences whose terms approximate (from the left and from the right) the root of an equation. The convergence order and efficiency index of this method are also studied in the general case and then in various particular cases.
Authors
Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)
Keywords
nonlinear equations in R; Aitken-Steffensen method; monotone sequences; iterative methods; convergence order.
Scanned paper (soon).
Latex version of the paper (soon).
Cite this paper as:
I. Păvăloiu, Approximation of the roots of equations by Aitken-Steffensen-type monotonic sequences, Calcolo, 32 (1995) nos. 1-2, pp. 69-82.
About this paper
Print ISSN
0008-0624
Online ISSN
1126-5434
References
Paper (preprint) in HTML form
Approximation of the Roots of Equations by Aitken-Steffensen-Type Monotonic Sequences
Abstract.
The aim of this paper is to study the conditions under which the well-known Aitken-Steffensen method for solving equations leads to monotonic sequences whose terms approximate (from the left and from the right) the root of an equation. The convergence order and efficiency index of this method are also studied in the general case and then in various particular cases.
1. Introduction
From a practical standpoint, in order to approximate the roots of equations it is advantageous to use methods which lead to monotonic sequences. In this paper we shall use a single iterative process to determine an increasing sequence and a decreasing one , both converging to a root of the equation Such a procedure has the advantage of allowing, at each iteration step, an approximation error checking, i.e.
As it is well known, such sequences can be generated by applying simultaneously two methods, e.g. Newton’s method and the chord method. In the sequel we shall show that, in conditions similar to those imposed to the above mentioned methods, the Aitken-Steffensen method generates two sequences which fulfil the above inequality.
In Section 2 we give sufficient conditions for the general method of Aitken-Steffensen ((3) below) to generate two monotonic sequences, both converging to the solution of equation (1) below. In this way some similar results obtained in [4] are completed, and the proofs given in [5] are made simpler to a certain extent.
In Section 3 we study some particular cases of the method (3), namely the methods (7) and (9) below. In Section 4 there is indicated a way to construct the auxiliary functions or required by (3) or (7), in relatively simple conditions as the monotonicity and convexity of the function In Section 5 the convergence orders and the efficiency indices of the methods (3) and (7) are studied, concluding that the method (7) has a higher efficiency index. This one is the same as that of Newton’s method, but is given conditions, the method (7) as well as (3) and (9), provide in addition bilateral approximations for the root of the equation (1).
So, let be an interval of the real axis, and consider the equation
(1) |
where . Besides (1), consider two more equations
(2) | ||||
where .
The Aitken-Steffensen method consists in constructing the sequences and by means of the following iterative process
(3) |
where stands for the first order divided difference of on the points We shall denote by the second order divided difference of on
As it is well known [6, pp. 268–269], and will also be shown in Section 5, the convergence order of the sequence generated by (3) is at least 2, but it generally depends on the convergence orders of the sequences and generated by [2], [5].
Another advantage of the iteration method (3) is the following: if equation (1) is given, then (as we shall see in Section 4) we have at our disposal many possibilities to construct the functions and from equations (2). On the other hand, hypotheses concerning the differentiability of on the whole interval are not needed.
If is a function, we shall adopt the following definitions concerning the monotonicity and convexity of :
Definition 1.1.
The function is said to be increasing (nondecreasing; decreasing; nonincreasing) on if for every the relations hold respectively.
Definition 1.2.
The function is convex (nonconcave; concave; nonconvex) on if for every the relations hold respectively.
2. Convergence of the Aitken-Steffensen method
We shall adopt the following hypotheses concerning the functions and
-
(a)
are continuous on
-
(b)
is increasing on
-
(c)
is decreasing on
- (d)
-
(e)
for every the relation holds.
Concerning the problem stated in Section 1, the following theorem holds:
Theorem 2.1.
Suppose that and fulfil conditions (a)–(e) and, in addition,
-
(i1)
is increasing and convex on and there exists
-
(ii1)
there exists such that and
Then the sequences , generated by (3), where fulfils condition (ii1), have the following properties:
-
(j1)
the sequences and are increasing and bounded;
-
(jj1)
the sequence is decreasing and bounded;
-
(jjj1)
-
(jv1)
and
Proof.
Since is increasing on and it follows that For by (e) we get which, for leads to if and, analogously, if By b) and we obtain i.e. while from the above inequalities it follows By (c) and g we get that is, By (i1) and we obtain which, together with and (3), shows that
It is easy to see that the following identities
(4) |
(5) |
hold for every
Since it follows that and, using (4) and (3), we get If we put in (5) and take into account (3), we get
from which, considering the convexity of and the inequalities and one obtains that is,
In this way we have obtained
(2.2′) |
In order that the above reasoning may be repeated we still have to show that verifies the last condition of (ii1), namely Indeed, from and (b) it follows which, by (c), leads to It remains to show that Suppose that that is, implying which contradicts (2.2′).
Consider for which and . Repeating the above reasoning, we obtain the inequalities
(6) |
In this way conclusions (j1), (jj1) and (jv1) of Theorem 2.1 are proved. To prove (jjj1), denote We shall show that
In can be shown analogously that the following theorems hold, too:
Theorem 2.2.
Suppose that fulfil conditions (a)–(e), and in addition
-
(i2)
is increasing and concave on and there exists
-
(ii2)
there exists for which and
Then the sequences generated by (3), have the following properties:
-
(j2)
are decreasing and bounded;
-
(jj2)
is increasing and bounded;
-
(jjj2)
and the following relations hold
Theorem 2.3.
Suppose that fulfil conditions (a)–(e) and in addition
-
(i3)
is decreasing and convex on and there exists
-
(ii3)
there exists for which and
3. Particular cases
An interesting particular case of the procedure (3) is obtained taking
for every In this way one obtains Steffensen’s method, namely
(7) |
where stands for
One observes easily that in this case hypotheses (a), (b), (d) and (e) are automatically fulfilled by For the study of the convergence of the sequence generated by (7), we have to adopt the following hypotheses:
-
(a1)
the functions and are continuous on
-
(b1)
the function is decreasing on
-
(c1)
equations (1) and have only one common root
With these specifications, the theorems stated in Section 2 yield:
Corollary 3.1.
Suppose that and fulfil conditions (a1)–(c1) and in addition is increasing and convex on there exists and the initial point in (7) can be chosen such that and Then the sequence is increasing and bounded, while the sequence is decreasing and bounded. Moreover, the relations and
hold.
Corollary 3.2.
Suppose that and fulfil conditions (a1)–(c1) and in addition is increasing and concave on there exists and from (7) can be chosen such that and Then is decreasing and bounded, while is increasing and bounded. Moreover, the following
relations: and
hold.
Corollary 3.3.
Corollary 3.4.
Another interesting particular case is that in which has the form:
(8) |
In this case the iterative method (7) will have the form
(9) |
whose convergence follows easily by Corollaries 3.1–3.4. The following results simplify the conditions imposed to and in [1]:
Corollary 3.5.
4. Determination of auxiliary functions
In what follows we shall show that, under reasonable hypotheses concerning the monotonicity and convexity of we have at our disposal various ways to choose and such that the hypotheses of the theorems and corollaries stated above are verified.
1. Let us admit that is differentiable on and denote by the right-hand derivative in a and by the left-hand derivative in we also admit that is increasing and convex on Suppose that the equation has a root Then we choose and
Since is convex, if follows that is increasing on hence for every In this case we have for every and for every We choose then a subinterval for which If we suppose in addition that there exists for which and then the above constructed functions and verify the hypotheses of Theorem 2.1. It is easy to see that if we choose and with , and then and also verify the hypotheses of Theorem 2.1.
2. If is differentiable, decreasing and convex on then and can be chosen as in case If there exists such that and then the hypotheses of Theorem 2.3 are verified.
3. If is differentiable, decreasing and concave on then we may put and If there exists for which and then and fulfil the conditions of Theorem 2.4.
4. If is differentiable, increasing and concave on then and can be chosen as in case 3. If there exists such that and then and verify the hypotheses of Theorem 2.2.
5. Convergence order and efficiency index
To fix the ideas, we shall consider hereafter, besides (1), another equation, equivalent to (1), of the form
(10) |
where . We shall also consider a sequence which, together with and verifies the properties:
-
(a)
for every ;
-
(b)
and are convergent and where is the root of equation (1);
-
(c)
for every
-
(d)
is differentiable at
We shall adopt the following definition of the convergence order of the sequence
Definition 5.1.
The sequence is said to have the convergence order with respect to if there exists
(11) |
and
The following theorem holds:
Theorem 5.1.
If and and verify the properties (a)–(d), then the necessary and sufficient condition for to have the convergence order , with respect to is to exist
(12) |
and
Proof.
Consider now the functions appearing in equations (2) and let the function be given by
(13) |
Concerning the convergence order of the sequence generated by (3), the following theorem holds:
Theorem 5.2.
Suppose that the functions and and the initial point fulfil the conditions of Theorem 2.1. If there exists and in addition the sequence has the convergence order with respect to while has the convergence order with respect to then the sequence has the convergence order with respect to the function given by (13).
Proof.
At first observe that, under the stated hypotheses, we may use (12) to determine the convergence order. The same hypotheses also lead to
(14) |
(15) |
Using (4), (5) and the procedure (3), we obtain
from which, by (14) and (15), it follows easily the equality
∎
Remark 5.1.
An analogous theorem holds in the case of the procedure (3), too.
Theorem 5.3.
Remark 5.2.
Remark 5.3.
Returning to the functions and determined in Section 4, where or , according to the case under consideration), it is easy to see
that the convergence order of the sequence generated by (3) is equal to .
In the case of the procedure (7), for or (according to the considered situation), the convergence order is also equal to .
In the sequel we shall deal with the efficiency index of the iterative procedures studied along the previous sections.
According to A. M. Ostrowski’s [3] definition for the efficiency index of an iterative procedure, and taking into account Theorems 5.1, 5.2 or 5.3 for the iterative procedures (3), (7), or (9), the efficiency index is expressed as
Here is the convergence order of the sequence generated by one of these procedures, while stands for the number of values of functions to be computed at each iteration step.
Taking into consideration Remark 5.3, it follows that, choosing or as in Section 4, the efficiency index of the procedure (7) is while that of (9) is
From this standpoint, the procedure (7) has the efficiency index equal to that of Newton’s method.
6. Numerical examples
1. Consider the equation
for Since is increasing and convex on we can choose and as in the case 1 (Section 4). We obtain
Taking the functions and verify the hypotheses of Theorem 2.1 on the interval Applying the procedure (3), we get the following results:
0 | 1.50000000000000 | 2.08198430811832 | 2.50854785469606 | -4.610-01 |
---|---|---|---|---|
1 | 2.32357265230323 | 2.33006829103803 | 2.33195667567199 | -5.110-03 |
2 | 2.33112222668589 | 2.33112235050042 | 2.33112238618252 | -9.910-08 |
3 | 2.33112237041442 | 2.33112237041442 | 2.33112238041442 | -3.510-17 |
2. Consider the equation
for Observe that is differentiable on and the derivative of from the left on is The function is increasing and convex on and we may take An elementary calculation shows that and ; and therefore the hypotheses of Corollary 3.1 are verified. The procedure (7) leads to the following results:
0 | -2.000000000000000 | -1.37420481033188 | -7.853 981 633 974 4810-01 |
---|---|---|---|
1 | -1.406051288716128 | -1.40401615840899 | -7.509 542 276 017 4610-01 |
2 | -1.404223647476550 | -1.40422359726392 | -2.442 156 368 565 0410-03 |
3 | -1.404223602391970 | -1.40422360239197 | -6.025 515 505 460 5810-08 |
4 | -1.404223602391970 | -1.40422360239197 | -3.718 813 451 625 2810-17 |
References
-
[1]
M. Balázs,
††margin:
available soon,
refresh and click here A bilateral approximating method for finding the real roots of real equations, Rev. Anal. Numér. Théor. Approx., (21), 2 (1992), 111–117. - [2] V. Casulli, D. Trigiante, The convergence order for iterative multipoint procedures, Calcolo, (13), 1 (1977), 25–44.
- [3] A. M. Ostrowski, Solution of Equations and Systems of Equations, (1960), Academic Press, New York and London.
- [4] ††margin: clickable I. Păvăloiu, On the monotonicity of the sequences of approximations obtained by Steffensen’s method, Mathematica (Cluj), (35), (58), 1 (1993), 71–76.
- [5] ††margin: clickable I. Păvăloiu, Bilateral approximations for the solutions of scalar equations, Rev. Anal. Numér. Théor. Approx., (23), 1 (1994), 95–100.
- [6] F. J. Traub, Iterative Methods for the Solution of Equations, (1964), Prentice-Hall, Inc. Englewood Cliffs, N.J.