Abstract

We study the approximation of an eigenpair (an eigenvalue and a corresponding eigenvector) of a a linear operator T from X to X, X be a Banach space.

The eigenpair is regarded as a solution of a nonlinear system obtained by considering the usual definition plus a norming function and then applying the Chebyshev or the Newton method.

We take into account that the augmented mapping has the third Frechet derivative the null operator, and we give a semilocal convergence result, which ensures the r-order three (cubic r-order) of the iterates.

We consider next the finite dimensional case, and analyse the computation of an eigenpair of a matrix. We also compare in this case the efficiency of the Newton and Chebyshev method, and find out that the later is more efficient.

Authors

Emil Cătinaş
(Tiberiu Popoviciu Institute of Numerical Analysis, Romanian Academy)

Ion Păvăloiu
(Tiberiu Popoviciu Institute of Numerical Analysis, Romanian Academy)

Keywords

linear operator; eigenvalue; eigenvector; eigenpair; Newton method; Chebyshev method; semilocal convergence; r-convergence order.

Cite this paper as:

E. Cătinaş, I. Păvăloiu, On the Chebyshev method for approximating the eigenvalues of linear operators, Rev. Anal. Numér. Théor. Approx., 25 (1996) nos. 1-2, pp. 43-56.

PDF

Scanned paper: on the journal website.

Latex-pdf version of the paper. (soon)

About this paper

Print ISSN

2457-6794

Online ISSN

2501-059X

Google Scholar citations

[1] M.P. Anselone, L.B. Rall, The solution of characteristic value-vector problems by Newton method, Numer. Math. 11 (1968), pp. 38–45.
CrossRef

[2] P.G. Ciarlet, Introduction a l’analyse numerique matricielle et a l’optimisation, Mason, Paris, 1990.

[3] F. Chatelin, Valeurs propres de matrices, Mason, Paris, 1988.

[4] L. Collatz, Functionalanalysis und Numerische Mathematik, Springer-Verlag, Berlin, 1964.

[5] V.S. Kartısov, F.L. Iuhno, O nekotorıh Modifikatiah Metoda Niutona dlea Resenia Nelineinoi Spektralnoi Zadaci, J. Vıcisl. matem. i matem. fiz. (33) (1973) 9, pp. 1403–1409.

[6] I. Pavaloiu, Sur les procedes iteratifs a un ordre eleve de convergence, Mathematica (Cluj) 12 (35) (1970) 2, pp. 309–324.
paper

[7] J.F. Traub, Iterative Methods for the Solution of Equations, Prentice-Hall Inc., Englewood Cliffs, N. J., 1964.

Paper (preprint) in HTML form

On the Chebyshev method for approximating the eigenvalues of linear operators

On the Chebyshev method for approximating
the eigenvalues of linear operators

Emil Cătinaş and Ion Păvăloiu
(Cluj-Napoca)

1. Introduction

Approaches to the problem of approximating the eigenvalues of linear operators by the Newton method have been done in a series of papers ([1], [3], [4], [5]). There is a special interest in using the Newton method because the operatorial equation to be solved has a special form, as we shall see. We shall study in the following the convergence of the Chebyshev method attached to this problem and we shall apply the results obtained for the approximation of an eigenvalue and of a corresponding eigenvector for a matrix of real or complex numbers. It is known that the convergence order of Newton method is usually 2 and the convergence order of Chebyshev method is usually 3. Taking into account as well the number of operations made at each step, we obtain that the Chebyshev method is more efficient than the Newton method.

Let E be a Banach space over K, where K= or K=, and T:EE a linear operator. It is well known that the scalar λ is an eigenvalue of T if the equation

(1) Txλx=θ

has at least one solution x¯θ, where θ is the null element of the space E. The elements xθ that satisfy equation (1) are called eigenvectors of the operator T, corresponding to the eigenvalue λ.

For the simultaneously determination of the eigenvalues and eigenvectors of T we can proceed in the following way.

We attach to the equation (1) an equation of the form

(2) Gx=1

where G is a linear functional G:EK.

Consider the real Banach space F=E×K, with the norm given by

(3) u=max{x,|λ|},uF,u=(xλ), with xE and λK.

In this space we consider the operator f:FF given by

(4) f(xλ)=(TxλxGx1).

If we denote by θ1=(θ0) the null element of the space F, then the eigenvalues and the corresponding eigenvectors of the operator T are solutions of the equation

(5) f(u)=θ1.

Obviously, f is not a linear operator.

It can be easily seen that the first order Fréchet derivative of f has the following form [4]:

(6) f(u0)h=(Th1λ0h1λ1x0Gh1),

where u0=(x0λ0) and h=(h1λ1). For the second order derivative of f we obtain the expression

(7) f′′(u0)hk=(λ2h1λ1h20),

where k=(h2λ2).

The Fréchet derivatives of order higher than 2 of f are null.

Considering the above forms of the Fréchet derivatives of f, in the following we shall study the convergence of the Chebyshev method for the operators having the third order Fréchet derivative the null operator.

2. The convergence of Chebyshev method

The iterative Chebyshev method for solving equation (5) consists in the successive construction of the elements of the sequence (un)n0 given by

(8) un+1=unΓnf(un)12Γnf′′(un)(Γnf(un))2,n=0,1,,u0F,

where Γn=[f(un)]1.

Let u0F and δ>0,b>0 be two real numbers. Denote S={uF:uu0δ}. If

m2=supuSf′′(u),

then

supuSf(u) f(u0)+m2δ,
supuSf(u) f(u0)+δf(u0)+m2δ2=:m0.

Consider the numbers

(9) μ=12m22b4(1+14m2m0b2)ν=b(1+12m2m0b2)

With the above notations, the following theorem holds:

Theorem 1.

If the operator f is three times differentiable with f′′′(u)θ3 for all uS (θ3 being the 3-linear null operator) and if, moreover, there exist u0F, δ>0,b>0 such that the following relations hold

i. the operator f(u) has a bounded inverse for all uS, and

f(u)1b;

ii. the numbers μ and ν given by (9) satisfy the relations

ρ0=μf(u0)<1

and

νρ0μ(1ρ0)δ,

then the following properties hold:

j. (un)n0 given by (8) is convergent;

jj. if u¯=limuun, then u¯S and f(u¯)=θ1;

jjj. un+1unνρ03nμ,n=0,1,;

jv. u¯un νρ03nμ(1ρ03n),n=0,1,

Proof.

Denote by g:SF the following mapping:

(10) g(u)=Γ(u)f(u)12Γ(u)f′′(u)(Γ(u)f(u))2,

where Γ(u)=f(u)1.

It can be easily seen that for all uS the following identity holds

f(u)+f(u)g(u)+12f′′(u)g2(u)=
=12f′′(u)[f(u)1f(u),f(u)1f′′(u)(f(u)1f(u))2]+
+18f′′(u)[f(u)1f′′(u)(f(u)1f(u))2]2

whence we obtain

(11) f(u)+f(u)g(u)+12f′′(u)g2(u)12m22b4(1+14m0m2b2)f(u)3,

or

(12) f(u)+f(u)g(u)+12f′′(u)g2(u)μf(u)3, for all uS.

Similarly, by (10) and taking into account the notations we made, we get

(13) g(u)νf(u), for all uS.

Using the hypotheses of the theorem, the inequality (12) and the fact that f′′′(u)=θ3, we obtain the following inequality:

f(u1) f(u1)f(u0)f(u0)g(u0)12f′′(u0)g2(u0)
+f(u0)+f(u0)g(u0)+12f′′(u0)g2(u0)
μf(u0)3.

Since u1u0=g(u0), by (12) we have

u1u0νf(u0)=νμf(u0)μ<νρ0μ(1ρ0)δ,

whence it follows that u1S.

Suppose now that the following properties hold:

  • a)

    uiS, i=0,k¯;

  • b)

    f(ui)μf(ui1)3,i=1,k¯.

By the fact that ukS, using (12) it follows

(14) f(uk+1)μf(uk)3,

and from relation uk+1uk=g(uk)

(15) uk+1uk<νf(uk).

The inequalities b) and (14) lead us to

(16) f(ui)1μ(μf(u0))3i,i=1,k+1¯.

We have that uk+1S:

(17) uk+1u0i=1k+1uiui1i=1k+1νf(ui1)νμi=1k+1ρ03i1νρ0(1ρ0)μ.

Now we shall prove that the sequence (un)n0 is Cauchy. Indeed, for all m,n we have

(18) un+mun i=0m1un+i+1un+i
νi=0m1f(un+i)
νμi=0m1ρ03n+i
= νμρ03ni=0m1ρ03n+i3n
νρ03nμ(1ρ03n),

whence, taking into account that ρ0<1, it follows that (un)n0 converges. Let u¯=limnun. Then, for m in (17) it follows jv. The consequence jjj. follows from (15) and (16). ∎

3. The approximation of the eigenvalues and eigenvectors of matrices

In the following we shall apply the previously studied method to the approximation of eigenvalues and eigenvectors of matrices with elements real or complex numbers.

Let p and the matrix A=(aij)i,j=1,p¯, where aijK, i,j=1,p¯.

Using the above notations, we shall consider E=Kp and F=Kp×K. Any solution of equation

(19) f(xλ):=(Axλxxi01)=(θ0),

i0{1,,p} being fixed, where x=(x1,,xp)Kp and θ=(0,,0)Kp, will lead us to an eigenvalue of A and to a corresponding eigenvector. For the sake of simplicity denote λ=xp+1, so that equation 19 is represented by the system

(20) fi(x1,,xp,xp+1)=0,i=1,p+1¯

where

(21) fi(x1,,xp+1)=ai1x1++(aiixp+1)xi++aipxp,i=1,p¯

and

(22) fp+1(x)=xi01.

Denote by P the mapping P:Kp+1Kp+1 defined by relations (21) and (22).

Let x¯n=(x1n,,xp+1n)Kp+1. Then the first order Fréchet derivative of the operator P at x¯n has the following form

(23) P(x¯n)h=(a11xp+1na12a1i0a1px1na21a22xp+1na2i0a2px2nap1ap2api0appxp+1nxpn00100)(h1h2hphp+1),

where h¯=(h1,,hp+1)Kp+1. If we write k¯=(k1,,kp+1)Kp+1 then for the second order Fréchet derivative we get

(24) P′′(x¯n)k¯h¯=(kp+100k10kp+10k200kp+1kp0000)(h1h2hphp+1).

Denote by Γ(x¯n) the inverse of the matrix attached to the operator P(xn) and u¯n= Γ(x¯n)P(x¯n)=(u1n,u2n,,up+1n). Let v¯n=P′′(x¯n)(Γ(x¯n)P(x¯n))2=P′′(x¯n) u¯n2. We obtain the following representation

(25) v¯n=P′′(x¯n)u¯n2=(up+1n00u1n0up+1n0u2n00up+1nupn0000)(u1nu2nupnup+1n).

From this relation we can easily deduce the equalities

(26) vin=2up+1nuin,i=1,p¯;vp+1n=0.

Denoting w¯n=(w1n,w2n,,wp+1n)=Γ(xn)v¯n and supposing that x¯n is an approximation of the solution of system (20), then the next approximation x¯n+1 given by method (8) is obtained by

(27) x¯n+1=x¯nu¯n12w¯n,i=0,1,

Consider Kp with the norm of an element x=(x1,,xp) given by the equality

(28) x=max1ip{|xi|},

and consequently

(29) A=max1ipj=1p|aij|.

It can be easily seen that P′′(x¯n)=2, for all x¯nKp+1. Let x¯0Kp+1 be an initial approximation of the solution of system (20). Consider a real number r>0 and the set S¯={xKp+1xx0r}. Denote by

m¯0= P(x¯0)+rP(x¯0)+2r2,
μ¯= 2b¯4(1+12m¯0b¯2),
ν¯= b¯(1+m¯0b¯2),

where b¯=supxS¯Γ(x),Γ(x) being the inverse of the matrix attached to the operator P(x).

Taking into account the results obtained in Section 2, the following consequence holds:

Corollary 2.

If x¯0Kp+1 and r,r>0, are chosen such that the matrix attached to the operator P(x) is nonsingular for all xS¯, and the following inequalities hold:

ρ0<μ¯P(x¯0)<1
ν¯ρ¯0μ¯(1ρ¯0)r

then the following properties are true

  • j1.

    the sequence (x¯n)n0 generated by (27) is convergent;

  • jj1.

    if  x¯=limnx¯n, then P(x¯)=θ1=(0,,0)Kp+1;

  • jjj1.

    x¯n+1x¯nν¯ρ¯03nμ¯,n=0,1,;

  • jv1.

    x¯x¯nν¯ρ¯03nμ¯(1ρ¯03n),n=0,1,

Remark.

If the radius r of the ball S¯ is given and there exists P(x0)1 and 2rP(x0)1 <1, then

P(x)1P(x0)112rP(x0)1,

for all xS¯ and in the above Corollary, taking into account the proof of Theorem 1, we can take

b¯=P(x0)112rP(x0)1.

4. A comparison to the Newton method

Note that if in (27) we neglect the term 12w¯n, then we get the Newton method:

x¯n+1=x¯nu¯n=x¯nΓ(x¯n)F(x¯n),n=0,1, .

In order to compare the Newton method and the Chebyshev method, we shall consider that the linear systems which appear in both methods are solved by the Gauss method.

While the Newton method require at each step the solving of a system Ax=b, the Chebyshev method require the solving of two linear systems Ax=b,Ay=c with the same matrix A and the vector c depending on the solution x of the first system. So, we shall adapt the Gauss method in order to perform as few as possible multiplications and divisions. When comparing the two methods we shall neglect the number of addition and subtraction operations.

The solving of a given linear system Ax=b,AMm(K),b,xKm, (where we have denoted m=p+1) using the Gauss method consists in two stages. In the first stage, the given system is transformed into an equivalent one but with the matrix of the coefficients being upper triangular. In the second stage the unknowns (xi)i=1,m are determined by backward substitution.

The first stage. There are performed m1 steps, at each one vanishing the elements on the same column below the main diagonal.

We write the initial system in the form

(a111a1m1am11amm1)(x1xm)=(b11bm1).

Suppose that a1110,a111 being called the first pivote. The first line in the system is multiplied by α=a211a111 and is added to the second one, which becomes 0,a222,a232,,a2m2,b22, after performing m+1 multiplication or division (M/D) operations. After m1 such transformations the system becomes

(a111a121a1m10a222a2m20am22amm2)(x1x2xm)=(b11b22bm2).

Hence at the first step there were performed (m1)(m+1) M/D operations.

In the same manner, at the k-th step we have the system

(a111a121a1k1a1,k+11a1m10a222a2k2a2,k+12a2m200akkkak,k+1kakmk00ak+1,kkak+1,k+1kak+1,mk00amkkam,k+1kammk)(x1x2xkxk+1xm)=(b11b22bkkbk+1kbmk).

Supposing the k-th pivote akkk0 and performing (mk)(mk+2) M/D operations we get

(a111a121a1k1a1,k+11a1m10a222a2k2a2,k+12a2m200akkkak,k+1kak,mk000ak+1,k+1k+1ak+1,mk+1000am,k+1k+1ammk+1)(x1x2xkxk+1xm)=(b11b22bkkbk+1k+1bmk+1).

At each step k, the elements below the k-th pivote vanish, so they are not needed any more in the solving of the system.

The corresponding memory in the computer is used keeping in it the coefficients

ak+1,kkakkk,,amkkakkk,

which, of course, will be needed only for solving another system Ay=c, with c depending on x, the solution of Ax=b.

At the first stage there are performed

(m1)(m+1)++13=2m3+3m25m6M/Doperations.

The second stage. Given the system

(a111a121a1m10a222a2m200ammm)(x1x2xm)=(b11b22bmm)

the solution x is computed in the following way:

xm=bmm/ammm,xk=(bkk(ak,k+1kxk+1++akmkxm))/akkk,x1=(b11(a121x2++a1m1xm))/a111.

At this stage there are performed 1+2++m=m(m+1)2M/D operations. In the both stages, there are totally performed

m33+m2m3M/D operations.

In the case when we solve the systems Ax=b,Ay=c, where the vector c depends on the solution x, we first apply the Gauss method for the system Ax=b and at the first stage we keep below the main diagonal the coefficients by which the pivotes were multiplied.

Then we apply to the vector c the transformations performed to the vector b when solving Ax=b.

Denote c=(ci1)i=1,m.

At the first step

c22:=a21c11+c21cm2:=am1c11+cm1.

At the k-th step

ck+1k+1:=ak+1,kckk+ck+1kcmk+1:=amkckk+cmk.

At the m-th step

cmm:=am,m1cm1m1+cmm1.

There were performed m1+m2++1=m(m1)2M/Doperations.

Now the second stage of the Gauss method is applied to

(a111a1m10ammm)(y1ym)=(c11cmm).

In addition to the case of a single linear system, in this case were performed m(m1)2M/Doperations, getting

m33+32m256mM/D operations,

and taking into account (3.8) we add (m1) more M/D operations, obtaining

m33+32m2+m61 M/D operations.

Remark. At the first stage, if for some k we have akkk0, then an element ai0,kk0,i0{k+1,,m} must be find, and the lines i0 and k in A and b be swapped.

In order to avoid the error accumulations, a partial or total pivote strategy is recommended even if akkk0 for k=1,m1¯.

For partial pivoting, the pivote is chosen such that ai0,kk=maxi=k,m¯|aikk|.

The interchange of lines can be avoided by using a permutation vector p=(pi)i=1,m, which is first initialized so that pi=i. The elements in A and b are then referred to as aij:=ap(i),j and bi:=bp(i), and swapping the lines k and i0 is done by swapping the k-th and i0-th elements in p.

For the Chebyshev method, the use of the vector p cannot be avoided by the very interchanging of the lines, because we must keep track for the permutations made, in order to apply them in the same order to the vector c.

Moreover, we need two extra vectors t and q. In t we store the transpositions applied to the lines in Ax=b, which are then successively applied to q. At the first stage of the Gauss method, when the k-th pivote is ai0,kk and i0k, the k-th and i0-th elements in p are swapped, and we assign tk:=i0 to indicate that at the k-th step we applied to p the transposition (k,i0).

After computing the solution of Ax=b, we initialize the vector c by (25), the permutation vector q by qi:=i,i:=1,m¯, and then we successively apply the transforms operated to b, taking into account the eventual transpositions.

The algorithm for solving the second linear system in the Chebyshev method is as follows.

for k:=1tom1do

begin

if t[k]<>k

then   {at the k-th step the transposition}

begin  {(k,t[k]) has been applied to p}

auxi:=q[k];

q[k]:=q[t[k]];

q[t[k]]:=auxi;

end,

for i:=k+1 to m do c[q[i]]:=c[q[i]]+A[q[i],k]c[q[k]]

end;

for i:=m downto 1 do {the solution y is now computed}

begin

sum:=0;

for j:=i+1 to m do sum:=sum+A[p[i],j]y[j]; {now p=q}

y[i]:=(c[p[i]]sum)/A[p[i],i];

end.

We adopt as the efficiency measure of an iterative method M the number

E(M)=lnqs,

where q is the convergence order and s is the number of M/D operations needed at each step.

We obtain

E(N)=3ln2m3+3m2m

for the Newton method and

E(C)=6ln32m3+9m2+m6

for the Chebyshev method.

It can be easily seen that we have E(C)>E(N) for n2, i.e. the Chebyshev method is more efficient than the Newton method.

5. Numerical example

Consider the real matrix

A=(1111111111111111)

which has the following eigenvalues and eigenvectors

λ1,2,3=2x1=(1,1,0,0),x2=(1,0,1,0),x3=(1,0,0,1) andλ4=2,x4=(1,1,1,1).

Taking the initial value x0=(1,1.5,2,1.5,1), and applying the two methods we obtain the following results:

Newton method

n x1 x2 x3 x4 x5=λ
0 1.0 -1.500 000 000 0 -2.000 000 000 0 -1.500 000 000 0 -1.000 000 000 0
1 1.0 -0.900 000 000 0 -0.800 000 000 00 -0.900 000 000 00 -1.600 000 000 0
2 1.0 -1.012 500 000 0 -1.025 000 000 0 -1.012 500 000 0 -2.050 000 000 0
3 1.0 -1.000 152 439 0 -1.000 304 878 0 -1.000 152 439 0 -2.000 609 756 1
4 1.0 -1.000 000 023 2 -1.000 000 046 5 -1.000 000 023 2 -2.000 000 092 9
5 1.0 -1.000 000 000 0 -1.000 000 000 0 -1.000 000 000 0 -2.000 000 000 0

Chebyshev method

n x1 x2 x3 x4 x5=λ
0 1.0 -1.500 000 000 0 -2.000 000 000 0 -1.500 000 000 0 -1.000 000 000 0
1 1.0 -0.972 000 000 00 -0.944 000 000 00 -0.972 000 000 00 -1.888 000 000 0
2 1.0 -0.999 950 001 89 -0.999 900 003 77 -0.999 950 001 89 -1.999 800 007 5
3 1.0 -1.000 000 000 0 -1.000 000 000 0 -1.000 000 000 0 -2.000 000 000 0

References

Received March 7, 1996.

Tiberiu Popoviciu Institute of Numerical AnalysisRomanian AcademyP.O. Box 68 Cluj-Napoca 1RO-3400 Romania

1996

Related Posts