Abstract
The paper is concerned with the order of convergence and the efficiency index of iterative methods of interpolatory type for solving scalar equations. Some classes of such methods are presented and then, using well defined criteria, the methods having the optimal efficiency index (i.e. those which practically are most efficient) are determined. For these methods the efficiency indexes are effectively determined.
Authors
Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)
Keywords
Cite this paper as:
I. Păvăloiu, Optimal efficiency indexes for iterative methods of interpolatory type, Computer Science Journal of Moldova, 5 (1997) no. 1(13), pp. 20-43.
About this paper
Publisher Name
Article on the journal website
Print ISSN
1561-4042
Online ISSN
Not available yet.
References
?
Paper (preprint) in HTML form
OPTIMAL EFFICIENCY INDEXES FOR ITERATIVE METHODS OF INTERPOLATORY TYPE ††thanks: AMS Subject Classification (1991): Primary 65Y20, 68Q25, 65H05.††thanks: This work has been supported by the Romanian Academy of Sciences.
Abstract
The paper is concerned with the order of convergence and the efficiency index of iterative methods of interpolatory type for solving scalar equations. Some classes of such methods are presented and then, using well defined criteria, the methods having the optimal efficiency index (i.e. those who practically are the most efficient) are determined. For these methods the efficiency indexes are effectively determined.
1 Introduction
In this paper we propose a unitary approach concerning the computational complexity for the numerical solving of scalar equations by iterative methods of interpolatory type. We shall consider some classes of such methods from which, using well defined criteria, we shall choose the optimal ones.
As a measure for the complexity of a method we shall adopt the efficiency index (see [4]).
For this purpose we shall start by presenting some general considerations concerning the convergence order and the efficiency index of an iterative method. Then we shall specify the interpolatory methods to be studied. Finally, we shall select from the interpolatory classes those having the highest efficiency index, and for the classes for which the selection method can’t be applied, we shall give delimitations for the efficiency index.
2 The convergence order and the efficiency index
Denote by and consider the equation
(1) |
where . In the following we shall suppose, for simplicity, that the equation has a unique solution . Let be a function having a unique fixed point in the interval , which coincides with
For the approximation of the root of equation , under certain conditions, we may consider the elements of the sequence generated by the following iterative process
(2) |
More generally, if is a function of variables whose restriction to the diagonal of the set coincides with i.e.
then we may consider the following iterative process:
(3) |
The convergence of the sequence generated by or depends on certain properties of the functions and , respectively . The amount of time needed by a computer to obtain a suitable approximation of , depends both on the convergence order of the sequence and on the number of elementary operations that must be performed at each iteration step in or . Concerning the convergence order, it can be exactly computed in almost all the cases. Hence it remains to solve the difficult problem of determining the number of elementary operations that must be performed at each iteration step. A general approach to this problem of course can’t be successful. That’s why A.M. Ostrowski proposed in [4] a simplification of this problem, by considering the number of function evaluations at each iteration step. At first sight this approach seems to be strange, taking into account that some functions may be more complicated and others may be simpler from the computational standpoint. But our purpose is to compare different methods applied to the same equation, and such an approach can give results.
We consider an arbitrary sequence , satisfying together with and the following properties:
- a)
-
and for
- b)
-
the sequences and are convergent and where is the solution of
- c)
-
For all , where we have denoted by the first order divided difference of on the nodes and ;
- d)
-
is derivable at
Definition 2.0.1
The sequence has the convergence order , , in respect to the function , if there exists the limit:
(4) |
and .
Remark 2.0.1
If the sequence is generated by the iterative method , then the Definition 2.1. reduces to the known one [4].
For a unitary treatment of the determination of the convergence order of the studied methods, we shall use the following lemmas.
Lemma 2.0.0
If the sequence and the functions and satisfy the properties a) - d) then the necessary and sufficient condition for the sequence to have the convergence order with respect to the function is that the following limit exists:
(5) |
and
Proof. Supposing that one of the equalities or is true and taking into account the properties a) - d), we obtain:
Lemma is proved.
Lemma 2.0.0
If is a sequence of real positive numbers satisfying:
i. The sequence is convergent and
ii. There exist the nonnegative real numbers and a bounded sequence with for all and such that the elements of satisfy
(6) |
iii. The sequence is convergent and
Then is the positive solution of the equation:
Proof. From we obtain
But it can be easily seen that
whence it follows that i.e. Lemma is proved.
We shall denote in the following by the number of function evaluations that must be performed at each iteration step in , respectively for .
In the hypotheses of Lemma 2.1 and taking into account the definition given in [4], we have:
Definition 2.0.2
The real number is called the efficiency index of the iterative method or if there exists
and
Remark 2.0.2
If for the methods and there exists a natural number such that for all and is the convergence order of these methods, then the efficiency index is given by the following expression:
(7) |
3 Iterative methods of interpolatory type
In the following we shall briefly present the Lagrange-Hermite-type inverse interpolatory polynomial. It is well known that this leads us to general classes of iterative methods from which, by suitable particularizations we obtain usual methods as Newton’s method, chord’s method, Chebyshev’s method, etc.
For the sake of simplicity we prefer to treat separately the Hermite polynomial and the Lagrange polynomial, though the last is a particular case of the first.
As we shall see, a suitable choice of the nodes enables us to improve the convergence orders of Lagrange-Hermite-type methods. We shall call such methods Steffensen-type methods.
3.1 Lagrange-type inverse interpolation
Denote by the range of for Suppose is times differentiable and for all . It follows that is invertible and there exists Consider interpolation nodes in :
(8) |
In the above hypotheses it follows that the solution of equation is given by
Using the Lagrange interpolatory polynomial for the function at the nodes we shall determine an approximation for , i.e. for
Denote and let be the mentioned polynomial, which is known to have the form
where
The following equality holds
(9) |
where
and
It is also known that in the mentioned hypotheses concerning the derivability of on , the function admits derivatives of any order , for all and the following equality holds [4], [8]:
(10) |
where and the above sum extends over all nonnegative integer solutions of the system
From , neglecting we obtain the following approximation for
Denoting
we obtain
where
It is clear that if are distinct approximations of the solution of equation then a new approximation can be obtained as above, i.e.
(11) |
with the error estimate given by
(12) |
where belongs to the smallest open interval containing
If we replace in we obtain for the sequence the relations:
(13) |
where belongs to the open interval determined by and .
Suppose that satisfies the hypotheses of Lemma 2.1 and that the sequence , converges to zero, where is generated by . Then the convergence order of this sequence is equal to the positive solution of the equation:
Considering the set of all equations of the above form for , and denoting by its corresponding positive solution it is known that the following relations hold [4]:
- a′)
-
- b′)
-
- c′)
-
.
3.2 Hermite-type inverse interpolation
Consider in the following, besides the interpolation nodes , natural numbers where and
We shall suppose here too, for simplicity, that is times differentiable on . From this and from for all , it follows, by , that is also times differentiable on . Denoting , the Hermite polynomial for the nodes with multiplicity orders has the following form:
where
(14) |
If are distinct approximations of the solution of the equation , then the next approximation can be obtained as before in the following way:
(15) |
where, as in
It can be easily seen that the following equality holds:
(16) |
where belongs to the smallest open interval containing and belongs to the open interval determined by and .
If we suppose that verifies the hypotheses of Lemma 2.1 and, moreover, , then it is clear that the convergence order of the method is given by the positive solution of the equation
(17) |
In the following we shall consider the following particular cases of :
For from we obtain
(18) |
method having the convergence order given by the positive solution of the equation
(19) |
Let denote the positive solution of equation . It is easy to prove that the following properties hold
- a′′)
-
- b′′)
-
- c′′)
-
Taking in we obtain again Chebyshev’s method, i.e.
(20) |
where the convergence order being .
Concerning the positive solution of equation we state the following lemma.
Lemma 3.2.0
The positive solution of equation verifies the relations:
(21) |
Proof. Let
(22) |
It is sufficient to prove that where We shall use for this the inequality between the arithmetic mean and the geometric mean, i.e.
Using this inequality we obtain
i.e.
Remark 3.2.3
It can be easily seen that the number given by can be exprimed using :
The second part of relations follows easily from the inequality where
3.3 Steffensen-type iterative methods
The convergence orders of methods (3.1.4), (3.2.2), respectively (3.2.5) can be improved if the interpolation nodes in the corresponding formulae are chosen in a special way. For this purpose we consider a continuous function , whose unique fixed point in the interval is . We also suppose that and verify the equality
(23) |
where , for all .
Let be an approximation of the solution . Denote and
Considering now as interpolation nodes the numbers by we obtain
(24) |
and from we have
(25) |
The iterative methods and are generalizations of the Steffensen’s method, which can be obtained from for (see [4], [5]).
From one obtains the following representations for :
where
Considering we obtain:
(26) |
and, analogously, from we get
(27) |
In the relations and and are contained in the open interval determined by and from and respectively and and belong to the smallest open interval containing from respectively .
If we suppose that the sequences and given by
are bounded and respectively then we clearly have that the convergence orders of methods respectively are equal to , respectively .
Remark 3.3.4
For the way of choosing the function with the mentioned properties see for example [5].
4 Optimal efficiency
We shall analyze in the following the efficiency index of each of the methods described and in the hypotheses adopted below we shall determine the optimal methods, i.e. those having the highest efficiency index.
As we have seen, the formulae for computing the derivatives of have a complicated form and they depend on the successive derivatives of . Though, in the case where the orders of the derivatives of are low, the values of these derivatives are obtained by only a few elementary operations. Taking into account the generality of the problem we shall consider each computation of the values of any derivative of by as a single function evaluation. For similar reasons we shall also consider each computation of the inverse interpolatory polynomials as a single function evaluation.
As it will follow from our reasonings, the methods having the optimal efficiency index are generally the simple ones, using one or two interpolation nodes and the derivatives of up to the second order.
Remark that in our case we can use for the efficiency index the relation .
4.1 Optimal Chebyshev-type methods
Observe that for passing from the -th iteration step to the , in method must be performed the following evaluations:
i.e. values.
Then, by we perform the following function evaluations:
where Finally, for the right hand expression of relation we perform another function evaluation, so that function evaluations must be performed.
By the efficiency index of method has the form
Considering the function , we observe that it attains its maximum at , so that the maximum value of is attained for . We have proved the following result:
Theorem 4.1.4
Among the Chebyshev-type iterative methods having the form the method with the highest efficiency index is the third order method, i.e.
(28) |
In the following table some approximate values of are listed :
m | 1 2 3 4 5 |
E(m) | 1.1892 1.2009 1.1892 1.1746 1.1610 |
Table 4.1.1
We note that
4.2 The efficiency of Lagrange-type methods
We shall study the methods of the form , for which the convergence order verifies from Taking into account Remark 2.2, it can be easily seen that we can use relation for the efficiency index of these methods. For each step in in order to pass to the next step, only must be evaluated, the other values from being already computed. We have also another function evaluation in computing the right-hand side of relation . So there are needed two function evaluations. Taking into account that the convergence order of each method satisfies , and denoting by the corresponding efficiency index, we have
and
We have proved:
Theorem 4.2.5
For the class of iterative methods of the form the efficiency index is increasing with respect to the number of interpolation nodes, and we have the equality
4.3 Optimal Hermite-type particular methods
We shall study the class of iterative methods of the form for . Taking into account the remarks from 4.2. it is clear that we can use again relation for the efficiency index.
If is an approximation for the solution obtained by then for passing to the following iteration step we need
i.e. function evaluations. Then, by we must compute the derivatives of the inverse function , where . Another function evaluation is needed for computing the right hand side of relation . We totally have function evaluations, the other values in being already computed.
By from 3.2. and denoting by the efficiency of this method, we get:
(29) |
(30) |
For a fixed , by it follows that the efficiency index is an increasing function with respect to and
In the following we shall study as a function of and , .
By we have
and
(31) |
For consider the functions , and
Some elementary considerations show that and satisfy is increasing on and decreasing on and is decreasing on . The maximum value of is .
Let be the solution of the equation
(32) |
It can be easily seen that exists and it is the unique solution for equation For , so it is clear that the maximum value of can be obtained for . It is easy to prove that and Taking into account the properties of and it is clear that in order to determine the greatest value of it will be sufficient to consider only those verifying, and .
Table 4.3.1. contains the approximate values of the efficiency indexes corresponding to these values of and
qn | 1 | 2 | 3 |
---|---|---|---|
2 | 1.2856 | ||
3 | 1.2487 | 1.2573 | |
4 | 1.2175 | 1.2218 | 1.2226 |
Table 4.3.1
The highest value for the efficiency index is hence obtained for and . We shall precise explicitly the method for these values. For this purpose it is convenient to use the divided differences on multiple nodes. The following table contains the divided differences for the inverse function on the nodes having the multiplicity orders 2.
Table 4.3.2
Here and and the other divided differences are computed using the well-known recurrence formula.
In this case the method has the following form:
(33) |
The following theorem holds:
Theorem 4.3.6
Among the methods given by relation for and the method with the highest efficiency index is given by , and corresponds to the case and .
We shall analyze the case In this case the efficiency index verifies . We also consider, besides the function already defined, the functions , which satisfy the following properties and
It can be easily shown that the equation has a unique positive solution, denoted by We also have for and for i.e. attains its maximum value at
We also have that , showing that for all But since it follows that we must examine only the cases when . Taking into account that is the solution of the equation we get that the maximum value of the function is equal to
Let , An elementary reasoning leads us to the following conclusions: is decreasing on the equation has a unique solution on the interval and
Since for we have , it follows that the values of and for which attains its maximum must be searched in the set
(34) |
Table 4.3.3. below contains the approximate values of the solutions and the error being smaller than
Table 4.3.3
Since we shall be interested only in the integer parts of the solutions .
From the above Table and by we can see that attains its maximum at . Taking into account that for then we observe that is increasing with respect to .
Hence the following theorem holds:
Theorem 4.3.7
Taking in , the greatest values of the efficiency indexes , , are obtained for . In this case the efficiency index is increasing with respect to , and we have
4.4 Bounds for the efficiency index of the general Hermite-type methods
As it was shown in [6], the method have the highest convergence order when the natural numbers verify the inequalities . More exactly consider the equations:
(35) |
(36) |
(37) |
where and is an arbitrary permutation of the numbers .
If are the corresponding positive solutions for equations then the following Lemma holds:
Lemma 4.4.0
If then , i.e., among all equations of the form equation has the greatest positive root.
In the following we shall assume that the multiplicity orders of the interpolation nodes of the Hermite polynomial which leads to the method satisfy
From the above assumptions, at each iteration step there must be performed function evaluations. Denoting by the efficiency index of and taking into account Lemma 3.2.1 we get:
Theorem 4.4.9
If and is the positive solution of then the efficiency index of the method satisfies
(38) |
Taking into account the properties of the function given in 4.3. and that , it follows that the expression attains its maximum value for . Taking into account the inequalities from the fact that attains its maximum value at do not imply the maximality of .
4.5 Optimal Steffensen-type methods
In the following we shall determine the optimal efficiency index for the class of iterative methods given by . First, we observe that at each iteration step in , we must compute values of the function , , , being an already computed approximation of the solution .
We then compute i.e. function evaluations. In order to compute the successive values of and at the nodes we need function evaluations. Finally, there is another function evaluation in computing the right hand side of . Totally there are function evaluations.
If we denote by the efficiency index of then
which, taking into account the results from 4.1., attains its maximum at .
Remark 4.5.5
If we take in , then method is a particular case of , since for in we get .
By the above remark, if then from , it follows . Hence we have to analyze the following cases:
- i)
-
, i.e.
- ii)
-
i.e. or
- iii)
-
i) For , by we get the following method:
(39) |
ii) For we get the method
(40) |
and for we get
(41) |
iii) For we get the method , i.e. the Chebyshev’s method of third order.
We have proved the following theorem:
Theorem 4.5.10
Among Steffensen-type iterative methods given by , the methods have the optimal efficiency index.
Remark 4.5.6
In the particular case when the condition imposed to obtain an optimal method leads us to two possibilities, namely: and , i.e. method or and , i.e. method .
References
- [1] R. Brent, S. Winograd, F. Wolfe, Optimal Iterative Processes for Root-Finding, Numer. Math. 20 (1973), 327-341.
- [2] Gh. Coman, Some Practical Approximation Methods for Nonlinear Equations, Mathematica - Revue d’Analyse Numérique et de Théorie de l’Approximation, Tome 11, N 1-2, (1982), 41-48.
- [3] H.T. Kung, J.F. Traub, Optimal Order and Efficiency for Iterations with Two Evaluations, SIAM J. Numer. Anal., Vol. 13, N 1, (1976), 84-99.
- [4] A.M. Ostrowski, Solution of Equations and Systems of Equations, Academic Press, New York and London, (1966).
- [5] I. Pǎvǎloiu, Bilateral Approximations for the Solutions of Scalar Equations, Revue d’Analyse Numérique et de Théorie de l’Approximation, (23), 1, (1994), 95-100.
- [6] I. Pǎvǎloiu, Optimal Problems Concerning Interpolation Methods of Solution of Equations, Publ. Inst. Math., 52 (66) (1992), 113-126.
- [7] Traub J.F., Iterative methods for solution of equations, Prentice-Hall Inc., Englewood Cliffs, New Jersey, 1964.
- [8] Turowicz B.A., Sur les derivées d’ordre superieur d’une fonction inverse, Ann. Polon. Math. 8 (1960), 265-269.