Local convergence of a two-step Gauss-Newton Werner-type method for solving least squares problems
Abstract.
The aim of this paper is to extend the applicability of a two-step Gauss-Newton-Werner-type method (TGNWTM) for solving nonlinear least squares problems. The radius of convergence, error bounds and the information on the location of the solution are improved under the same information as in earlier studies. Numerical examples further validate the theoretical results.
Key words and phrases:
Gauss-Newton method, Werner’s method, local convergence, least squares problem, average Lipschitz condition.2005 Mathematics Subject Classification:
65G99, 65J15, 65H10, 49M15, 47H17. 65N35, 47H17, 49M15.1. Introduction
Let be natural numbers with Let also be an open and convex subset of We are concerned with the solution of the least squares problem [4, 5, 6, 7, 8, 9]:
(1.1) |
where is a Fréchet-differentiable mapping. Numerous problems can be brought in the form (1.1) using Mathematical Modeling [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]. The closed form solutions can only be found in special cases. That explains why most solution methods for problem (1.1) are iterative. Let and set In the present study, we provide the local convergence analysis of GNWTM defined for each by
(1.2) | |||||
where If TGNWTM reduces to a Gauss-Newton-Werner type method [3, 8, 9]. Notice that in each iteration the inversion of is required only once. Therefore, the computational cost is essentially the same as in the Gauss- Newton method. The decomposition of costs floating-point operations (Flops) leading to the computation of It then follows from the second substep of method (1) that Flops are needed for the computation of
The local convergence analysis of method (1) was given in the elegant paper by Shakhno et.al. in [9] (see also related work in [3, 8]). Their convergence analysis uses average Lipschitz continuity condition as well as Lipschitz conditions.
Using the concept of the average Lipschitz continuity [12] and our new idea of restricted convergence domains, we present a local convergence analysis with the following advantages (A) over works using the similar information [3, 4, 8, 9, 11, 12, 13]:
-
(a)
Larger radius of convergence;
-
(b)
Tighter error bounds on the distances
-
(c)
An at least as precise information on the location of the solution
Achieving (a)-(c) is very important in computational sciences, since: (a)′ We obtain a wider choice of initial guesses; (b)′ Fewer iterates are required to obtain a desired error tolerance; (c)′ Better information about the ball of convergence is obtained.
2. Local convergence analysis
Set to be the open ball in and by to denote its closure. Let Define parameter by The convergence analysis of numerous iterative methods has been given using the following concept due to Wang [12]:
Definition 1.
Mapping satisfies the Lipschitz condition with average on if
where is a positive non-decreasing function.
It turns out that the convergence analysis of iterative methods based on the preceding notion can be improved as follows:
Definition 2.
The mapping satisfies the center-Lipschitz condition with average on if
where is a positive non-decreasing function.
Clearly, we have that
(2.1) |
and can be arbitrary small [2, 3, 4]. Let be a parameter. Suppose that equation
(2.2) |
has positive solutions. Denote by the smallest such solution. Notice for example that exists, if
(2.3) |
Indeed, function is such that and The existence of follows from the intermediate value theorem.
Definition 3.
The mapping satisfies the restricted Lipschitz condition with average on if
where is a positive non-decreasing function.
We have that
(2.4) |
since Throughout this paper, we suppose that
(2.5) |
unless, otherwise stated. Otherwise, i.e., if
(2.6) |
then the results that follow hold with replacing Moreover, we need the definitions:
Definition 4 ([12]).
Let be a twice Fréchet-differentiable mapping. We say that mapping satisfies the Lipschitz condition with average on if
where is a positive nondecreasing function.
Definition 5.
Let be twice Fréchet-differentiable mapping. We say that mapping satisfies the restricted Lipschitz condition with average on if
where is a positive nondecreasing function.
We have that
(2.7) |
It is worth noticing that the definition of functions and (based on and ) was not possible in earlier studies using and That is, whereas and It turns out that can replace the less precise in the computation of the upper bounds on the inverses of the operators involved and can replace respectively in the proofs of such results. Moreover, notice that the iterates lie in which is a more precise location than used in earlier studies [2, 3, 4, 8, 11, 12, 13]. We shall make the paper as self contained as possible by stating some standard auxiliary concepts and results.
Denote by the set of all matrices. The Moore-Penrose pseudo-inverse is defined by for each full rank [6].
Lemma 2.2 ([5]).
Let Assume that and then
Lemma 2.3 ([12]).
Let where is a positive integrable function and monotonically non-decreasing on Then, is monotonically non-decreasing with respect to
Lemma 2.4 ([12]).
Let where is a positive integrable function and monotonically non-decreasing on Then, is monotonically non-decreasing with respect to
As in [9], it is convenient for the local convergence analysis that follows to introduce some functions and parameters:
and
Notice that if and then the preceding definitions reduce to the corresponding ones in [9].
The local convergence analysis is based on the conditions ():
-
()
Mapping is twice Fréchet-differentiable, has full rank and solves problem (1.1).
-
()
satisfies: the center-Lipschitz condition with average on and the restricted Lipschitz condition with average on satisfies the restricted Lipschitz condition with average on where and are positive non-decreasing functions on
-
()
Function has a minimal zero in which also satisfies
Then, we can show the following local convergence result for TGNWTM under the conditions () and the preceding notation.
Theorem 2.5.
Suppose that conditions () hold. Then, sequences generated for by TGNWTM are well defined in remain in for each and converge to Moreover, the following estimates hold
(2.8) |
(2.9) |
and
(2.10) |
Proof. The proof follows the corresponding one in [9] but there are differences where we use (), instead of respectively used in [9]. We shall use mathematical induction to show that iterates are well defined converge to and the error estimates (2.8)–(2.10) are satisfied. Using TGNWTM for we can write
and
where
In view of the estimate
for and we obtain in turn
and
By the central Lipschitz condition, we have that
Moreover, by 2.1 and 2.2 and (), for all we get
and
By the monotonicity of and with 2.3 and 2.4, functions and are non-decreasing in That is, by ()
Thus, by 2.1–2.9 and condition (), we have in turn
In an analogous way, we get in turn
hold, where so and we also have that
so (2.10) is satisfied. Suppose that and (2.10) hold for By TGNWTM for we get in turn that
and
where Furthermore, we obtain
so
and
Concerning the uniqueness of the solution we have:
Proposition 2.6.
Under the conditions () further suppose that
(2.11) |
where Then, limit point is the only solution of problem (1.1) in
The proof follows from the corresponding one in [5] but we only use the center-Lipschitz condition.
3. Special cases and applications
Remark 3.1.
- (a)
-
(b)
If are constants, then we can obtain results of special cases.
- (c)
Therefore, our radius of convergence is larger and our ratio of convergence is smaller. Moreover the information on the location of the solution is more precise, since only is used in (2.11) [9]. Notice that these advantages are obtained under the same computational cost, since in practice the computation of and require the computation of the rest of the functions and as special cases.
Remark 6.
In particular, using the error estimates, it follows that for we have and
Also, for sufficiently large ,
leading to the equation
so the order of iterative method (1) is the positive root of the preceding equation which is
Next, we present an example to show that (3.1)–(3.3) hold as strict inequalities justifying the advantages as claimed at the introduction of this study.
EXAMPLE 3.2.
Let Define function on for by
Then, the Fréchet-derivative is given by
Notice that using the (2.9) conditions, we get Then
which justify the improvements as stated in the introduction of this paper.
References
- [1] S. Amat, M.A. Hernández, N. Romero, Semilocal convergence of a sixth order iterative method for quadratic equations, Appl. Numer. Math., 62 (2012), pp. 833–841.https://doi.org/10.1016/j.apnum.2012.03.001
- [2] I.K. Argyros, S. Hilout, Extending the applicability of the Gauss-Newton method under average Lipschitz-type conditions, Numer. Algorithms, 58 (2011), pp. 23–52.https://doi.org/10.1007/s11075-011-9446-9
- [3] I.K. Argyros, S. Hilout, On the semilocal convergence of Werner’s method for solving equations using recurrent functions, Punj. Univ. J. Math., 43 (2011), pp. 19–28.https://pu.edu.pk/images/journal/maths/PDF/vol43/paper3_Argyros_WMethod.pdf
- [4] I.K. Argyros, A. Magreñan, Iterative methods and their dynamics with applications, CRC Press, New York, 2017.
- [5] J. Chen, W. Li, Convergence of Gauss-Newton method’s and uniqueness of the solution, Appl. Math. Comput., 170 (2005), pp. 686–705. https://doi.org/10.1016/j.amc.2004.12.055
- [6] J.M. Dennis, R.B. Schnabel, Numerical methods for Unconstrained optimization and nonlinear equations, Prentice-Hall, N.Y, 1983.
- [7] W.M. Häubler, A Kantorovich-type convergence analysis for the Gauss-Newton method, Numer. Math., 48 (1986), pp. 119–125. https://doi.org/10.1007/BF01389446
- [8] R. Iakymchuk, S.M. Shakhno, On the convergence analysis of a two-step modification of the Gauss-Newton method, PAMM, 14 (2014), pp. 813–814. https://doi.org/10.1002/pamm.201410387
- [9] R.P. Iakymchuk, S.M. Shakhno, H.P. Yarmola, Convergence analysis of a two-step modification of the Gauss-Newton method and its applications, J. Numer. Appl. Math., 3(126) (2017), pp. 61–74. http://jnam.lnu.edu.ua/pdf/y2017_no3(126)_art05_iakymchuk_shakhno_yarmola.pdf
- [10] A.A. Magreñan, A new tool to study real dynamics: The convergence plane, Appl. Math. Comput., 248 (2014), pp. 29–38. https://doi.org/10.1016/j.amc.2014.09.061
- [11] G.N. Silva, Local convergence of Newton’s method for solving generalized equations with monotone operator, Appl. Anal., 97 (2018), pp. 1094–1105. https://doi.org/10.1080/00036811.2017.1299860
- [12] X. Wang, Convergence of Newton’s method and uniqueness of the solution of equations in Banach space, IMA J. Numer. Anal., 20 (2000), pp. 123–134. https://doi.org/10.1007/s10114-002-0238-y
- [13] W. Werner, Über ein Verfahren der Ordnung zur Nullstellenbestimmung, Numer. Math., 32 (1979), pp. 333–342. https://doi.org/10.1007/BF01397005