Abstract
Authors
Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis)
Keywords
Scanned paper: on the journal website.
Latex version of the paper.
Cite this paper as:
I. Păvăloiu, On computational complexity in solving equations by interpolation methods, Rev. Anal. Numér. Théor. Approx., 24 (1995) no. 1, pp. 201-214.
About this paper
Publisher Name
paper on journal website
Print ISSN
1222-9024
Online ISSN
2457-8126
References
[1] Brent, R., Winograd, S. and Wolfe, Ph., Optimal Iterative Processes for Root-Finding. Numer. Math. 20(5) (1973), pp. 327-341.
[2] Casulli, V. and Trigiante, D., Sui procedimenti Iterativi Compositi. Calcolo, XIII. IV (1976), pp. 403-420.
[3] Coman, Gh., Some practical Approximation Methods for Nonlinear
Equations. Mathematica Revue d’Analyse Num'{e}rique et de Th'{e}orie de l’Approximation, 11, 1-2 (1982), pp. 41-48.
[4] Kacewitcz, B., An Integral-Interpolation Iterative Method for the Soltuion of Scalar Equations, Numer. Math. 26 (4), (1976), pp. 355-365.
[5] Kung, H.T. and Traub, J.F., Optimal Order and Efficiency for Iterations With Two Evaluations. SIAM, Numer Anal. 13, 1, (1976), pp. 84-99.
[6] Ostrowski, M.A., Soltuion of Equations and Systems of Equations. Academic Press. New York and London (1960).
[7] Păvăloiu I., The Soltuion of the Equations by Interpolation (Romanian) Ed. Dacia Cluj-Napoca (1981).
[8] Păvăloiu I., Optimal Problems Concerning Interpolation Methods of Solution of Equations. Publications de L’Institut Mathematique Beograd. 52, (66), (1992), pp. 113-126.
[9] Traub, J.F., Iterative Methods for the Solution of Equations. Prentice-Hall, New York (1964).
[10] Traub, J.F., Theory of Optimal algorithms, Preprint. Department of Computer Science Carnegie-Mellon University (1973), pp. 1-22.
[11] Traub, J.F. and Wozniakowski, H., Optimal Radius of Convergence of Interpolatory Iterations for Operator Equations, Aequationes Mathematicae, 21 (1980), pp. 159-172.
[12] Turowicz, A.B., Sur les derivees d’ordre suprieur d’une fonction inverse, Ann. Polon Math.8 (1960), pp. 265-269.
Paper (preprint) in HTML form
On computational complexity in Solving Equations
by Interpolation Methods
1. Introduction
Let be an interval of the real axis and a function. Consider the equation:
(1) |
supposed to have a solution . Also get be a function whose fixed points from coincide with the root of (1).
For solving (1) one can usually use an iterative method of the form:
(2) |
More generally, if is a function depending on variables, whose restriction to the diagonal of the set coincides with , i.e.
(3) |
then one can consider the following iterative method for solving equation (1):
(4) |
The convergence of the sequences generated by (2) or (4) to a solution of equation (1) depends obviously on the properties of the functions respectively and the amount of time necessary to obtain a suitable approximation for the solution is influenced both by the convergence order of the methods (2), resp. (4) and by the amount of elementary operations that must be performed at each iteration step. This last aspect belongs to a chapter of the calculus theory and practice, chapter concerning the computational complexity.
Many authors ([1], [2], [3], [5], [6], [9], [10], [11]), who studied the computational complexity of the iteration processes, have defined different notions, as: the efficiency of a method, the efficiency index of a method or the cost of a method, which they have quantitatively expressed by different scalar magnitudes.
Throughout this paper we shall adopt the following definition for the convergence order of an iteration method:
Definition 1.1.
The real number is called the convergence order of the sequence generated by an iterative method if the following limit exists and is not zero:
(5) |
where is the solution of equation (1).
Concerning the calculus complexity, using the convergence order we can define now the following notion:
Definition 1.2.
If we suppose that is the same for all , and take into account that (6) has an asymptotical character, then there results for the following expression:
(7) |
In the following we shall study certain classes of iteration methods,namely the methods obtained by interpolation, among which we shall select those for which the efficiency index given by (7) is optimal, i.e. the greatest. For this purpose in the next section we shall briefly recall the classes of methods that we want to study.
2. Interpolation Iterative Methods
2.1. Lagrange’s inverse interpolation polynomial
Let be an interval and a function. Denote by the set of all values of for . Suppose that is one-to-one, i.e. there exists the inverse function . Consider in interpolation nodes:
(8) |
Suppose that equation (1) has the unique solution .
Obviously,
(9) |
and so the problem of approximation the solution reduces to the approximation of .
A simple and efficient approximation method for the functions is given by the interpolating approximation.
The Lagrange interpolation polynomial corresponding to the function on the nodes from (10) (taking into account that ) has the form:
(11) |
where .
If we suppose that the function has derivatives up to the order and for all then we have the following formula for the computation of the -th derivative of at the point ([7], [12]):
(12) |
where the above sum extends to all nonnegative integer numbers, solutions of the system:
(13) | ||||
If we suppose that admits derivatives up to the order on the interval and the -th derivative is bounded on , then by (12) we obtain,
(14) |
where is a point belonging to the smallest interval containing and
Denote by the number:
(15) |
and by (14) we get:
(16) |
from which we see that if are chosen in a neighbourhood of such that then can be considered as a new approximation for . We have denoted in inequality (16), .
Let now be approximations for . Then the Lagrange polynomial corresponding to the function on the nodes has the form:
(17) |
where . From this relation, for , we obtain a new approximation for , namely
(18) |
which satisfies the delimitation
(19) |
It is well known that the iterative method given by (18) has the convergence order , which is the unique positive root of the equation, [6]:
(20) |
It is also known (see [6]) that verifies:
(21) |
and
(22) |
Remark 2.1.
In the successive computation of the elements of the sequence generated by (18) it is necessary to compute at each step the values and .
We observe that practically there exists a connection both between and and between and .
2.2. Hermite Inverse Interpolating Polynomial
We consider, besides the interpolatory nodes (8), the natural numbers with and
(27) |
Suppose that admits derivatives up to the order on the interval and . Then, by (12) it follows that the function also admits derivatives up to the order . The Hermite polynomial of degree associated to the function on the nodes , assuming that for all , is:
(28) | |||
where
(29) |
This polynomial satisfies:
(30) |
for all .
As in 2.1 we obtain from (28) the following iterative method for solving equation (8):
(31) |
where in the polynomial .
Using the differentiability assumptions for , we obtain:
(32) |
where .
It is well known that the convergence order of (31) is given by the positive root of the equation:
(33) |
In the particular cases when we have iterative method:
(34) |
and when we get the Chebyshev iterative method of order :
(35) |
which has the convergence order .
3. The Efficiency Index of the Chebyshev Method of Order
In the following we shall make the assumptions:
-
a)
Consider as a function evaluation, the evaluation of the derivatives assuming as having been computed.
-
b)
Consider as a function evaluation, the evaluation of the right hand side of expression (35) assuming and as having been computed.
-
c)
Consider as a function evaluation the evaluation of the function or of any of its derivatives.
In this hypothesis, it is necessary for an iteration step with method (35) to compute firstly the values of the functions: at the point , altogether function for the calculus of the values of the successive derivatives of , by (12). If we take into account that for evaluating the right hand side expression from relation (35) is computed another function value, we have altogether function evaluations at each iteration step.
Using definition (7) for the efficiency index, method (35) has the following index:
(40) |
(We have taken into account that the convergence order is ).
We are searching for the maximum value of the index from (40), for .
For this purpose we consider the auxiliary function and we note that: , , is increasing for and decreasing for is a maximum point for the function .
For , the function attains its maximum for so attains its maximum form .
So, the following holds:
Theorem 3.1.
In the above assumptions a)–c) among all the Chebyshev iterative methods of the form (35) the method with the greatest efficiency index is the one of order, namely
(41) |
In the following table are shown the approximations of the efficiency index of Chebyshev methods for some values of .
We see that
4. The Efficiency Index for the Lagrange-Hermite Methods
In the following we shall study the case when , i.e. when number of nodes is greater than one. In the beginning we shall take the method given by (18) for which, if we take into account the assumptions a) - c) and neglect the first step, we get that at each iteration step we have first one function evaluation, namely the value of the function at , and then we have another function evaluation, for the right hand side of relation (18), hence altogether two function evaluations. Recalling that the convergence order for (18) is , which satisfies (21) and (22) we get for the efficiency index the relation
(42) |
By (22) we have that for all .
So we conclude:
Theorem 4.1.
If the assumptions a)–c) hold, for the Lagrange methods given by (17) the efficiency index is increasing with respect to the number of interpolation nodes and
Now we study the efficiency index for the methods given by (34), for which the convergence order verifies (36)–(39). Obviously, we suppose that in (34) giving (18).
There are two aspects that must be considered: the efficiency with respect to the number of interpolation nodes, when their multiplicity order is kept fixed and, on the second hand, the efficiency with respect to the multiplicity order for fixed .
We again suppose that assumptions a)–c) hold. So from the right hand side of (34), at each iteration step, excepting the first one, we have the following function evaluations: we compute i.e. function evaluations, and then by (12) we compute i.e., function evaluations, and finally we compute the right hand side of (34). so altogether, function evaluations.
For a fixed by (43) we get that the efficiency index is increasing as a function of , and by (44)
From (44) we also obtain:
(45) |
and
(46) |
A. For the first case, when we consider the auxiliary functions given by
and
As we have seen before, satisfies:
is increasing on and decreasing on so at attains its maximum.
One can establish the following relations for :
and is decreasing on .
Recalling that attains its maximum value at let be the solution of the equation
(47) |
then for we have hence, by (45), we obtain that the values of for which attains its maximum lie in the set . One can easily prove that so we shall study the cases and . Since we study 1) 2) and 3) .
1) The corresponding equation from (36) for is , with the positive solution . So
2) The convergence orders corresponding for this case are the solutions of the equations for respectively for , .
We obtain , respectively
3) The corresponding equations give us:
and
So the greatest efficiency index when is obtained for i.e .
B. Let , so (46) holds. We shall again consider two auxiliary functions
which possess the following properties: ;
and one can easily prove that: the equation has a unique positive solution, denoted by for and for i.e. attains its maximum value at .
Taking into account the properties of one can see that the equations
(48) |
have a unique positive solution for each .
In the table below we give the approximative values and for .
An elementary reasoning proves that and are decreasing functions of , , as we can see in the above table.
If then so the optimal values for must lie in the set .
It can be shown that for and for , . It follows that the only suitable value for is . In this case we get that , i.e. the efficiency index increases with , but anyway the best results hold for .
5. Margins for the Efficiency Index in the Case of Lagrange-Hermite Methods
We end this note by indicating, in the above assumptions, left and right margins of the efficiency index of the Lagrange-Hermite methods.
In this respect we shall first establish an inequality that will give left margins for the positive roots of equations (33).
Let
(49) |
where for .
The following Lemma holds.
Lemma 5.1.
The positive root of (49) satisfies:
(50) |
Proof.
Let
It will suffice to show , using the inequality between the weighted arithmetic and geometric mean, i.e.
(51) |
We get
i.e. .
We also get , where .
In the hypotheses of 3., using (31) in generating the sequence then the number of function evaluations at each step is .
References
- [1] Brent, R., Winograd, S. and Wolfe, Ph., Optimal iterative processes for root-finding. Numer Math. 20 (5) (1973), 327–341.
- [2] Casulli, V. and Trigiante, D., Sui procedimenti iterativi compositi. Calcolo, XIII, IV (1976), pp. 403–420.
-
[3]
Coman, Gh.,
††margin:
available soon,
refresh and click here Some practical approximation methods for nonlinear equations. Anal. Numér. Théor. Approx., 11, 1–2 (1982), pp. 41–48. - [4] Kacewitcz, B., An integral-interpolation iterative method for the solution of scalar equations, Numer. Math. 26 (4), (1976), pp. 355–365.
- [5] Kung, H. T. and Traub, J. F., Optimal order and efficiency for iterations with two evaluations. SIAM. J. Numer Anal. 13, 1, (1976), pp. 84–99.
- [6] Ostrowski, M. A., Solution of Equations and Systems of Equations. Academic Press. New York and London (1960).
-
[7]
Păvăloiu, I.,
††margin:
available soon,
refresh and click here The Solution of the Equations by Interpolation (Romanian) Ed. Dacia Cluj-Napoca (1981). - [8] ††margin: clickable Păvăloiu, I., Optimal problems concerning interpolation methods of solution of equa- tions, Publications de L’Institut Mathématique Beograd. 52, (66), (1992), pp. 113–126.
- [9] Traub, J. F., Iterative Methods for the Solution of Equations. Prentice-Hall, New York (1964).
- [10] Traub, J. F., Theory of Optimal Algorithms, Preprint Department of Computer Science Carnegie-Mellon University (1973) pp. 1–22.
- [11] Traub, J, F. and Wozniakowski, H., Optimal radius of convergence of interpolatory iterations for operator equations. Aequationes Mathematicae, 21 (1980), pp. 159–172.
- [12] Turowicz, A. B., Sur les dérivées d’ordre supérieur d’une fonction inverse, Ann. Polon Math. 8 (1960), pp. 265–269.
- [13]
Received 1 X 1994
Academia Română
Institutul de Calcul
”Tiberiu Popoviciu”
P.O. Box 68
3400 Cluj-Napoca 1
România