GMBACK is a Krylov solver for large linear systems, which is based on backward error minimization properties. The minimum backward error is guaranteed (in exact arithmetic) to decrease when the subspace dimension is increased. In this paper we consider two test problems which lead to nonlinear systems which we solve by the Newton‐GMBACK. We notice that in floating point arithmetic the mentioned property does not longer hold; this leads to nonmonotone behavior of the errors, as reported in a previous paper. We also propose a remedy, which solves this drawback
Authors
Emil Cătinaș Tiberiu Popoviciu Institute of Numerical Analysis, Academy Romanian
Keywords
linear systems; Krylov subspace method; backward error; GMBACK.
References
See the expanding block below.
Paper coordinates
E. Cătinaş, On the nonmonotone behavior of the Newton-GMBACK method, AIP Conf. Proc., 2008, vol. 1046, pp. 87-90 https://doi.org/10.1063/1.2997323
On the Nonmonotone Behavior of the Newton-Gmback Method
Emil CătinaşRomanian Academy, "T. Popoviciu" Institute of Numerical Analysis, P.O. Box 68-1, Cluj-Napoca, Romania, e-mail: ecatinas@ictp.acad.ro
Abstract
GMBACK is a Krylov solver for large linear systems, which is based on backward error minimization properties. The minimum backward error is guaranteed (in exact arithmetic) to decrease when the subspace dimension is increased. In this paper we consider two test problems which lead to nonlinear systems which we solve by the Newton-GMBACK. We notice that in floating point arithmetic the mentioned property does not longer hold; this leads to nonmonotone behavior of the errors, as reported in a previous paper. We also propose a remedy, which solves this drawback.
Keywords: linear/nonlinear systems, Krylov solvers, Newton method.
PACS: 02.60.Cb
NONLINEAR SYSTEMS.
The Newton method for solving a nonlinear system F(x)=0,quad F:D subeR^(N)rarrR^(N)F(x)=0, \quad F: D \subseteq \mathbb{R}^{N} \rightarrow \mathbb{R}^{N} leads to the solving of a linear system at each iteration step:
Under the following conditions (which will be implicitly assumed throughout this paper) the Newton method converge locally at qq-superlinear rate (see [7]): -EEx^(**)in int D-\exists x^{*} \in \operatorname{int} D such that F(x^(**))=0F\left(x^{*}\right)=0
the mapping FF is Fréchet differentiable on int D\operatorname{int} D, with F^(')F^{\prime} continuous at x^(**)x^{*};
the Jacobian F^(')(x^(**))F^{\prime}\left(x^{*}\right) is invertible.
When one considers approximate Jacobians at each step, F^(')(x_(k))+Delta_(k)inR^(N xx N)F^{\prime}\left(x_{k}\right)+\Delta_{k} \in \mathbb{R}^{N \times N}, we are lead to the quasi-Newton (QN) iterates
We have characterized the superlinear convergence of these iterates in the following result:
Theorem 1. [3] Assume that the QN iterates converge to x^(**)x^{*}. Then the convergence is superlinear if and only if
{:(1)||Delta_(k)s_(k)||=o(||F(x_(k))||)","quad" as "k rarr oo.:}\begin{equation*}
\left\|\Delta_{k} s_{k}\right\|=o\left(\left\|F\left(x_{k}\right)\right\|\right), \quad \text { as } k \rightarrow \infty . \tag{1}
\end{equation*}
THE GMBACK METHOD
When the dimension NN is large, the numerical solving of a linear system
Au=b,quad A inR^(N xx N)" nonsingular, "b inR^(N)A u=b, \quad A \in \mathbb{R}^{N \times N} \text { nonsingular, } b \in \mathbb{R}^{N}
becomes a difficult task. The Krylov solvers are popular choices for accomplishing this task, since they may offer good approximations at low computational cost.
We shall consider here the GMBACK solver introduced by Kasenally in [6]. For a given subspace dimension m in{1,dots,N}m \in\{1, \ldots, N\} and an initial approximation u_(0)inR^(N)u_{0} \in \mathbb{R}^{N} having the residual r_(0)=b-Au_(0)r_{0}=b-A u_{0}, it finds u_(m)^(GB)inu_(0)+K_(m)=u_(0)+span{r_(0),Ar_(0),dots,A^(m-1)r_(0)}u_{m}^{G B} \in u_{0}+\mathscr{K}_{m}= u_{0}+\operatorname{span}\left\{r_{0}, A r_{0}, \ldots, A^{m-1} r_{0}\right\} by the following minimization property:
Here ||*||_(F)\|\cdot\|_{F} denotes the Frobenius norm of a matrix, ||Z||_(F)=tr(ZZ^(t))^(1//2)\|Z\|_{F}=\operatorname{tr}\left(Z Z^{t}\right)^{1 / 2} while ||*||_(2)\|\cdot\|_{2} will denote the Euclidean norm from R^(N)\mathbb{R}^{N} and its induced operator norm.
The following steps are performed for determining u_(m)^(GB)u_{m}^{G B} :
Arnoldi
Let r_(0)=b-Ax_(0),beta=||r_(0)||_(2)r_{0}=b-A x_{0}, \beta=\left\|r_{0}\right\|_{2} and v_(1)=(1)/(beta)r_(0)v_{1}=\frac{1}{\beta} r_{0};
Form the Hessenberg matrix bar(H)_(m)inR^((m+1)xx m)\bar{H}_{m} \in \mathbb{R}^{(m+1) \times m} with the (possible) nonzero elements h_(ij)h_{i j} determined above, and the matrix V_(m)inR^(N xx m)V_{m} \in \mathbb{R}^{N \times m} having as columns the vectors v_(j):V_(m)=[v_(1)dotsv_(m)]v_{j}: V_{m}=\left[v_{1} \ldots v_{m}\right].
GMBACK
Determine an eigenvector v_(m+1)v_{m+1} corresponding to the smallest eigenvalue lambda_(m+1)^(GB)\lambda_{m+1}^{G B} of the generalized eigenproblem Pv=lambda Qv;P v=\lambda Q v ;
If the first component v_(m+1)^((1))v_{m+1}^{(1)} is nonzero, compute the vector y_(m)^(GB)inR^(m)y_{m}^{G B} \in \mathbb{R}^{m} by scaling v_(m+1)v_{m+1} such that
Set u_(m)^(GB)=x_(0)+V_(m)y_(m)^(GB)u_{m}^{G B}=x_{0}+V_{m} y_{m}^{G B}.
We shall assume in the following analysis that u_(m)^(GB)u_{m}^{G B} exists (this may not be the case when all the eigenvectors of the smallest eigenvalue have the first component 0 ).
Kasenally proved that for any u_(0)inR^(N)u_{0} \in \mathbb{R}^{N} and m in{1,dots,N}m \in\{1, \ldots, N\}, the backward error Delta_(m)^(GB)\Delta_{m}^{G B} corresponding to the GMBACK solution satisfies
Regarding the induced operator Euclidean norm, the following inequality is known:
{:(3)||Z||_(2) <= ||Z||_(F)","quad" for all "Z inR^(N xx N):}\begin{equation*}
\|Z\|_{2} \leq\|Z\|_{F}, \quad \text { for all } Z \in \mathbb{R}^{N \times N} \tag{3}
\end{equation*}
The eigenvalues at steps mm and m+1m+1 in the Arnoldi algorithm interlace as follows.
Lemma 2. [6] Suppose that m+1m+1 steps of the Arnoldi process have been taken and h_(m+2,m+1)!=0h_{m+2, m+1} \neq 0. Furthermore, assume that {lambda_(i)^(GB)}_(i=1,dots,m+1)\left\{\lambda_{i}^{G B}\right\}_{i=1, \ldots, m+1} and { hat(lambda)_(i)^(GB)}_(i=1,dots,m+2)\left\{\hat{\lambda}_{i}^{G B}\right\}_{i=1, \ldots, m+2} are, respectively, the eigenvalues of the matrix pairs (P_(m),Q_(m))\left(P_{m}, Q_{m}\right) and (P_(m+1),Q_(m+1))\left(P_{m+1}, Q_{m+1}\right) arranged in decreasing order. Then, for any i <= m+1i \leq m+1,
and we may control the convergence of the iterates by Theorem 1 with the aid of the backward error. It is worth noting that we may use formulas (2) and (3) to evaluate the magnitude of the backward error in the Euclidean norm.
We obtain:
Theorem 3. Assume that the Newton-GMBACK iterates are well defined and converge to x^(**)x^{*}. If
{:(5)lambda_(k)^(GB)rarr0","quad" as "k rarr oo:}\begin{equation*}
\lambda_{k}^{G B} \rightarrow 0, \quad \text { as } k \rightarrow \infty \tag{5}
\end{equation*}
then they converge superlinearly.
Of course, we may also use the inexact Newton model, and control the convergence of the iterates with the aid of residuals, see [4].
FIGURE 1. Newton-GMBACK errors for the Bratu problem.
Bratu problem
Consider the nonlinear partial differential equation
over the unit square of R^(2)\mathbb{R}^{2}, with Dirichlet boundary conditions. As mentioned in [1], this is a standard problem, a simplified form of which is known as the Bratu problem. We have discretized by 5-point finite differences, respectively by central finite differences on a uniform mesh, obtaining a system of nonlinear equations of size N=(n-2)^(2)N=(n-2)^{2}, where nn is the number of mesh points in each direction. As in [1], we took ff such that the solution of the discretized problem to be the constant unity, and alpha=10,lambda=1\alpha=10, \lambda=1, the initial approximations in the inner iterations were zero. The runs were made on a HP Proliant 570 G4 server, using MATLAB 2007a.
We took N=16,384N=16,384 and considered first some standard iterations, with fixed subspace dimension, of size m=40m=40. We noticed that, starting from outer iteration k=9k=9, where ||F||\|F\| was of magnitude 1e-91 \mathrm{e}-9, lemma 2 did not hold in floating point arithmetic. As we can see in figure 1, the consequence is that the convergence of the errors of iterates is no longer monotone.
The remedy we propose is to check at each inner iteration step in GMBACK whether the size of lambda\lambda is decreasing, and to stop the iterations if the size is increasing. We obtain monotone behavior of the errors, and the plot in figure 1 is relevant for the runs we made.
ACKNOWLEDGMENTS
This research has been supported by MEdC under grant 2CEx06-11-96.
REFERENCES
P. N., Brown, and Y., Saad, Hybrid Krylov Methods for Nonlinear Systems of Equations, SIAM Journal on Scientific and Statistical Computing 11, pp. 450-481 (1990).
E. Cătinaş, Inexact perturbed Newton methods and applications to a class of Krylov solvers, J. Optim. Theory Appl. 108, pp. 543-570 (2001).
E. Cătinaş, The inexact, inexact perturbed and quasi-Newton methods are equivalent models, Math. Comp., 74 no. 249 , pp. 291-301 (2005).
R. S. Dembo, S. C. Eisenstat and T. Steihaug, Inexact Newton methods, SIAM J. Numer. Anal. 19, pp. 400-408 (1982).
J. E. Dennis, Jr. and J. J. Moré, A characterization of superlinear convergence and its application to quasi-Newton methods, Math. Comp. 28, pp. 549-560 (1974).
E. M. Kasenally, GMBACK: a generalised minimum backward error algorithm for nonsymmetric linear systems, SIAM J. Sci. Comput. 16, pp. 698-719 (1995).
J. M. Ortega and W. C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables, Academic Press, New York, 1970.