an eigenvalue algorithm based on norm-reducing transformations

148
An eigenvalue algorithm based on norm-reducing transformations Citation for published version (APA): Paardekooper, M. H. C. (1969). An eigenvalue algorithm based on norm-reducing transformations. Eindhoven: Technische Hogeschool Eindhoven. https://doi.org/10.6100/IR41102 DOI: 10.6100/IR41102 Document status and date: Published: 01/01/1969 Document Version: Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers) Please check the document version of this publication: • A submitted manuscript is the version of the article upon submission and before peer-review. There can be important differences between the submitted version and the official published version of record. People interested in the research are advised to contact the author for the final version of the publication, or visit the DOI to the publisher's website. • The final author version and the galley proof are versions of the publication after peer review. • The final published version features the final layout of the paper including the volume, issue and page numbers. Link to publication General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal. If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, please follow below link for the End User Agreement: www.tue.nl/taverne Take down policy If you believe that this document breaches copyright please contact us at: [email protected] providing details and we will investigate your claim. Download date: 12. Jun. 2020

Upload: others

Post on 11-Sep-2021

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: An eigenvalue algorithm based on norm-reducing transformations

An eigenvalue algorithm based on norm-reducingtransformationsCitation for published version (APA):Paardekooper, M. H. C. (1969). An eigenvalue algorithm based on norm-reducing transformations. Eindhoven:Technische Hogeschool Eindhoven. https://doi.org/10.6100/IR41102

DOI:10.6100/IR41102

Document status and date:Published: 01/01/1969

Document Version:Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers)

Please check the document version of this publication:

• A submitted manuscript is the version of the article upon submission and before peer-review. There can beimportant differences between the submitted version and the official published version of record. Peopleinterested in the research are advised to contact the author for the final version of the publication, or visit theDOI to the publisher's website.• The final author version and the galley proof are versions of the publication after peer review.• The final published version features the final layout of the paper including the volume, issue and pagenumbers.Link to publication

General rightsCopyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright ownersand it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights.

• Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal.

If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, pleasefollow below link for the End User Agreement:www.tue.nl/taverne

Take down policyIf you believe that this document breaches copyright please contact us at:[email protected] details and we will investigate your claim.

Download date: 12. Jun. 2020

Page 2: An eigenvalue algorithm based on norm-reducing transformations
Page 3: An eigenvalue algorithm based on norm-reducing transformations

AN EIGENV ALUE ALGORITHM BASED ON NORM-REDUCING

TRANSFORMATIONS

PROEFSCHRIFT

TER VERKRIJGING VAN DE GRAAD VAN DOCTOR IN DE TECHNISCHE WETENSCHAPPEN AAN DE TECHNISCHE HOGESCHOOL TE EINDHOVEN OPGEZAG VAN DE RECTOR MAGNIFICUS DR.IR. A.A.TH.M. VAN TRIER,HOOGLERAAR IN DE AFDELING DER ELEKTROTECHNIEK, VOOR EEN COMMISSIEUITDESENAATTEVERDEDIGENOPDINSDAG

2 DECEMBER 1969 TE 16.00 UUR.

DOOR

MICHAEL HUBERTUS CORNELIUS PAARDEKOOPER

GEBOREN TE ZOETERWOUDE

Page 4: An eigenvalue algorithm based on norm-reducing transformations

DIT PROEFSCHRIFT IS GOEDGEKEURD DOOR

DE PROMOTOR PROF. DR. G.W. VELTKAMP

Page 5: An eigenvalue algorithm based on norm-reducing transformations

Aan Wil Aan Evert, Wouter en Liesbeth

Page 6: An eigenvalue algorithm based on norm-reducing transformations

0. Introduction

0.0. Introductory remarks

0.1. Notations, definitions and elementary theorems

0.2. A survey of Jacobi-like

0. 3. Summary

1. Real Norm-Reducing Shears

1.0. Introduction

1.1. Row congruency and Euclidean parameters of a

shear

1. 2. transformations by real unimodular

shears

for the real unimodular norm-

1 .4. The particular case D = F 0,

1.5. The commutator in relation to shear transfor­

matioas

2. Complex Norm-Reducing Shears

2.0. Introduction

2. 1 • Row congruency and Euclidean oa.ra,llle c

shear

of a

2.2. The unimodular norm-reducing shear

transformation

for the complex unimodular norm-

reducing shears

2.4. The case D = F == 0

2.5. The commutator in relation to shear transfor­

mations

1

1

9 18

24

29

29

46

49

53 53

53

55

61

64

5

Page 7: An eigenvalue algorithm based on norm-reducing transformations

3. Convergence to Normality

3.0. Introduction

3.1. A lower bound for the optimal norm-reduction

b,r shear transformations

3.2, The convergence theorem

4• Jacobi-like Methods for almost Diagonalization of

almost Normal Matrices

4.0. Introduction

4.1. Almost diagonalization of a complex almost

normal matrix

4.2. Almost block diagonalization of a real almost

normal matrix

4·3· The real diagonali representative of

Jn,tm(A) 4.4. The complex diagonalizing representative of

5. Numerical stability and the norm-reducing process

5.0. Introduction

5.1. Input and output perturbations related to

rounding errors

5.2. Error analysis of similarity transformations

5.3. The general error analysis applied to shear

transformation

5.4. A numerically stable transformation by the dia­

gonalizing representative of >Sntm(A)

5.5. Diagonal dominance and shear transformations

References

Sa.menvatting

Curriculum Vitae.

6

69 69

73

76

76

77

90

94

97

1.01

101

104

112

118

125

128

139

145

Page 8: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 0

INTRODUCTION

0. 0. Introductory remarks

Since the rise of the program-stored digital computer it has been

possible to master effectively the bulk of work necessary to solve

numerically the algebraic eigenvalue problem, i.e. the approximate

calculation of the eigenvalues and eigenvectors of a linear trans­

formation represented by a given matrix. The advent of this appa­

ratus has stimulated the construction new algorithms for this

problem. As concerns the Hermitean eigenvalue problem we mention

the numerically stable methods of Givens and Householder.

But also for the non-Hermi tean problem several new methods are

proposed •. Since this problem can be very ill-conditioned, the

construction of the latter algorithms presents serious difficul­

ties. Inexact arithmetic, the reverse of the computer's speed,

makes therefore the problem mathematically interesting. Research

on the numerical solution of the non-Hermitean eigenvalue problem

is very active at present. At the moment it is not yet clear which

of the algorithms proposed by several authors is preferable. The

QR-algorithm, developed by Francis in 1961-1962, attracts much at­

tention and inspires confidence as to speed of convergence and ac­

curacy of the results.

The method for the non-Hermitean eigenvalue problem which we pre­

sent in this book, is of what is known as the Jacobi-like type,

i.e. an extension of the classical Jacobi-method to non-normal

matrices.

The Jacobi-algortihm is based on the use of rotations, the

original matrix A A being recursively transformed into matrices 0

7

Page 9: An eigenvalue algorithm based on norm-reducing transformations

A , A , •••• , which tend to a diagonal form. In each step of the

p~ode~s the plane rotation is chosen to minimize the sum of the

squares of the moduli of the non-diagonal elements. In principle,

each normal matrix A can be transformed into a diagonal form by

these unitary Jacobi-transformations; the Euclidean norm of the

matrix A is invariant under these transformations and for normal .!.

matrices this norm equals ( .E !A.I 2) 2 , where A, A~, .••• , A

J=1 J 1 "" n

are

For

the eigenvalues of A. 2 n 2

non-normal matrices 1fAIIE > .. E I A. I (!IAIIE ~J=1 J i,

called the Euclidean norm of A); hence it is not possible to

transform these matrices unitarily into diagonal form, and so the

Jacobi-method is fruitless to achieve this end.

In 1962, Eberlein [4] suggested the use of non-unitary plane

transformations in order to diminish the Euclidean norms of the

matrices in the sequence thus obtained. It is not impossible that

this may lead to diagonalization of non-normal matrices since [22]

inf f!T-1 A Tl!E 2

T regular

n E

j=1

In the first part of this thesis (chapter 1, 2, 3 and 5) we con­

struct and investigate an algorithm to normalize non-normal ma­

trices. In chapters one and two an algorithm is described to re-.. duce (in some sense optimally) the EUclidean norm of a real,res-

pectively complex, non-normal matrix qy a plane non-unitar,y trans­

formation. In chapter three we prove that the sequence-{~},

generated by the successive application of this algorithm, con­

verges to the class of normal matrices with the same eigenvalues

as A0

, and,finally, in chapter five we show that computation of

these plane non-unitary similarity transformations can be per­

formed in a stable way.

In the second part (chapter four) an algorthm is described b.y which,

using unitary plane transformations, an almost normal matrix- let

8

Page 10: An eigenvalue algorithm based on norm-reducing transformations

us say the result of our norm-reducing process - can be transformed

into almost diagonal form. F·rom the diagonal elements of this form

we may read approximations of the eigenvalues we have aimed at.

0.1. Notations, definitions and elementary theorems

Q. 1 •. 1 • We

and we

our preliminaries with the definition of a normal

a list of well-known theorems concerning these

matrices.

Let A be a linear transformation on R to R , where R n n n is a

unitary or a Euclidean space of dimension nand let~* be the ad-

of A .

Definit:ion 0.1. A (=>./hA ""AA*. Let A be the matrix representation of ..4 on some orthonormal basis

of Rn; the conjugate transpose A* of A is the matrix representation

of fi * on the same basis.

In the we deal with square matrices of order n over the field

of complex numbers, unless mentioned otherwise. The eigenvalues of

the matrix A will be denoted by A.. (j ""1,2, ••• , n) where A.=fJ..+iV.o J J J J

Defini ti.on 0. 2. A normal ~ A*A AA*.

Theorem 0.1. The matrix representation A of /cJ on an orthonorrnal

basis is normal if and only if JJ is normal ([21], p.56).

A matrix A is normal if and

similar to a diagonal matrix ([21], p.165).

if A is unitarily

A real matrix A is normal if and only if A is orthog-

similar to a matrix that is the direct sum of matrices of the

9

Page 11: An eigenvalue algorithm based on norm-reducing transformations

form (A.), where

( Re(.\)

- Im( A.)

A. is a real eigenvalue of A and of

Im(A.))

Re( A.)

2 x 2 matrices

where A is a complex eigenvalue of A. This direct sum is called

Murnaghan 1 s canonical form of A.

Theorem 0.4. Let A be a normal matrix. The real parts of the eigen­

values of A are eigenvalues of' the Hermitean part i(A + A *)of A,and

the imaginar,y parts of the eigenvalues of A are eigenvalues of the . * skew-Hermitean part i(A- A ) of A.

Proof.

U*AU = Since A is normal,

diag ( ll· +i v.) , where J J

there exists a unitar,y matrix U so that

IJ.. + iv. (j = 1, ••• , n) are the eigen-J J

values of A. We see that

U* A ; A* U = diag ( ll j) , j 1, ... , n

and

U* A- A* U d' (' ) . 2 ~ag ~ vj ' J 1 , 2, ••• , n. 0

0.1.2.

Theorem 0.5. (Schur's lemma). For any matrix A there exists a

unitar,y matrix U for which holds U*AU = T, where T is of upper

triangular form; T is diagonal if and on~ if A is normal ([21],

p. 158).

Theorem 0.6. For any real matrix A of order n with k pairs of

complex conjugate eigenvalues

1J. £ ± i V£ (V£ >0, 1 ..;;; £ ..;;; k, 0 ..;;; k ..;;; [~ ])

and n-2 k real eigenvalues llj(j = 2k+1, ••• ,n) there exists an

10

Page 12: An eigenvalue algorithm based on norm-reducing transformations

orthogonal matrix Q so that

triangular form

has the following block upper

2

( \)

1 u u1n f.l1 -a- 13 1

a IJ.1 3 "- 2n

1

'

-

IJ.2k+4

0 0

0 0

and the elements a£ of QTAQ are positive,

In the proof of Schur's lemma we have to add the part of

the inductive proof corresponding to a pair of complex

Let X ;!: i y1 be eigenvectors corresponding to 1

!l1 + i \) . Since \) -fo X and y are linearly and we 1 1 1 1 have

A [x : y1 J [x1 ~) ( ~1 -\!1) 1 ' ' 1 ll1

Then sin

:) (cos q> -A [x y1] . 1 cos Sln q>

Page 13: An eigenvalue algorithm based on norm-reducing transformations

Let

(cos {jl - sin • xos• =Lx,:y] 1 1 . Sln {jl COS cp Sln cp

(cos• - sin~) ( "' =lx:y] 1 1 1 sincp COS (jl V

. 1

u 1

x cos cp + y sin cp 1 1

u x sin cp + y cos cp • 2 1 1

Then A [ u u ] 1 2

and T

u u 1 2

sin •) ("' cos cp v

1

- v1 )~~s • -sin •)

j.J.1

ln <p COS<p

-v,) IJ.1

T If X y

1 1 0 then we take cp O, otherwise we determine cp by

T Taking either value of <p, u1

u2

0. Finally to obtain orthonor-

mal vectors, u and u have to be standardized to length one. 1 2

-1 -1 Let v

1 ·- l!u 11 u V := llu2

1!2

u2

• 1 2 1 2

Then !lv 1[ = llv 11 1 ' T 0 and V V

1 I 2 2 2 1 2

A [v1 V ] = [v V ] ( IJ.1 -v/cx) (0.1.1)

2 1 2 cxv IJ.1 ' 1

where a = l!u ll /llu IT • 2 2 1 2

Let v1 and v2 be the first two columns of an orthogonal matrix ~; ~,

let B .- p- AP 1 1 1 •

Then from (0.1,1) we derive

Page 14: An eigenvalue algorithm based on norm-reducing transformations

B P TAP = 1 1 1

with a1 = ·ex v > 0. 1 1

fl1

a 1

0

0

- v 2 /a 1 1 ~13 ~,n

111 ~23 ~2n 0 ~33 ~3n

0 ~n3 ~nn

For the rest, l.he formal inductive i:'roof of the theorem is anal-

ogous l.u l.hat of Schur's lemma. 0

Definition 0.3.

I!AIIE is called the E"L.tcJ.idean norm of A.

Theorem o. 7. The Euclidean norm IIAI!E of A is invariant relative

to unitary similarity transformations, i.e.

U*U = I =) I!U*AU11 = IIAII • E E

Theorem c.s~_ Let A be a complex matrix with eigenvalues

\ , !..2

, ••• , t..n. Then

n ~ L:

i=1 I t...l 2

~

Eq'J.ality holds in (0.1.2) if and only if A is normal.

(0.1.2)

Proof. This iE a direct consequence of the theorems 0.5 and

n . 1

Definition~L',(A) := (IIA!I: - L: jt...j 2)2 • i=1 ~.

6(A) i£ calJ.ed the departure-of normality of A [14]

Next we give some theorems about the departure of normality which,

1 3

Page 15: An eigenvalue algorithm based on norm-reducing transformations

as the Euclidean norm, is invariant relative to unitar,r trans~or­

mation of A.

Theorem 0.9. For any matrix A

inf o. T regular

The infimum is assumed if and only if A is diagonalizable [22].

Theorem 0.10. For each matrix A there exists a normal matrix N

which has the same eigenvalues as A and is. such that I!A-NffE .,;;L~(A).

Proof. According to Schur 1 s lemma there exists a unitary matrix U

for which holds

U*AU = diag ( 11.:.) + T, J

where T is strictly upper triangular.

Then A = N + UTU*, where N = U diag(X.)U*. J

This matrix N is normal and

0

Corblla;y. Let A= diag( X.), A. being the eigenvalues of A. Then J J

there exists a unitary matrix U for which holds

!IU*AU - All ..; /:::,(A).

Theorem 0.11. For any real matrix A of order n there exists a

real normal matrix N of order n for which holds that A a:nd N have

the same eigenvalues and llA-NilE .,;;; 6(A).

Proof. Let QTAQ be the block upper triangular matrix indicated

in theorem 0.6. Then QT AQ = M + T, where

14

Page 16: An eigenvalue algorithm based on norm-reducing transformations

2

0 V

1.!1 -v 0 V _ _!_

1 1 a 1

V 1.!1 a -v 0

1 1 1

Ilk -\: !

M= ' T= vk Ilk

0 ll2k+1 ,j 0

with U an upper triangular matrix of which

u2

. 1 2

. = 0 , i = 1 , 2, ••• , k. ~- ' ~

2

0 \l.k

vk --ak

~-vk 0

0

0

Thus QT AQ is the sum of a Murnagban canonical form M and a pertur-

bation matrix T. Then

T T A = QJVIQ + QTQ = N + P. (0.1.3)

The matrices A and N have the same eigenvalues and N is a normal

matrix. Since the Euclidean norm and the departure of normality

of a matrix are invariant relative to unitar.y similarity transfor­

mations, we have

(0.1.4)

where (M,T)E is the inner product of the matrices M and T (consid­

ered as elements of the n2 -space) corresponding to the Euclidean

norm. As we see from the matrices M and T, written in full above,

15

Page 17: An eigenvalue algorithm based on norm-reducing transformations

k l!A-r-rn~ = 1'12 (A)- 2 z:: (a[v,.e) 2 v/a,.e..:; t'1

2 (A). o .e=t

Corollary. Let M be Murnaghan 1s canonical form corresponding to

the eigenvalues A, A, ••• , A of the real matrix A. Then there 1 2 n

exists an orthogonal matrix Q for wh:i.ch holds

IIQT AQ - Mlf ..,; 1'1(A).

Definition 0.5. Let~ (k = 0,1, ••• ) be similar to A= A0

and

n(A) the class of normal matrices with the same eigenvalues as A. The sequence~ is said to converge to normality (or to converge

to ?1.(A)) if there exists a sequence {\;} where \: E '!/.(A) (k = 0,1, ••• ), so that

Theorem 0.12. Let {~} be a sequence of similar matrices. It con­

verges to normality if and only if lim ~(~) = 0. k- 00

Proof. The sufficieney follows immediately from theorem 0.10

(for real matrices from theorem 0.11).

The convergence to normality of {~} implies that there exists a

sequence {Nk}, NkE:il.(A) so that lf~-\;IIE -o. Then there exists a

sequence of unitary matrices {vk} for which holds that

Then n 2 _ * 2 _ 1 2 _ I (k)l2 I (k)la lll\:lfE - I!Vk ~UJIE - 11 A+ Ekl E - .~ A.+ e.. + .L: e~ 0

J=1 J J J Jl.t .,yv

16

Page 18: An eigenvalue algorithm based on norm-reducing transformations

Hence n

0.1. 3.

In contrast to 6(A) the measure of non-normality of the matrix A

which we define in this subsection, is effectively computable.

Definition 0.6.

commutator of A.

* * C(A) := A A -AA • The matrix C(A) is called the

Theorem 0.13. Let 6(A) be the departure of normality of the ma­

trix A and C(A) the commutator of this matrix.

Then [14] 1

62 (A),.; [(n3 -n)/12]2 [IC(A)IIE (0.1.5)

and if A I 0 then [5]

(0.1.6)

Corollar,y. Let {~} be a sequence of similar matrices. It con­

verges to normality if and only if C(~) - 0 •

Page 19: An eigenvalue algorithm based on norm-reducing transformations

Finally, we define here some notions which are used in the descrip­

tion of Jacobi-like algorithms.

Definition 0.7. A shear matrix Tn is a non-singular matrix which --- .-vm

differs from the unit matrix I on~ in one of its two-dimensional

submatrices. In that one submatrix the elements are t,e,e' t,em, tm.e

and t • The indices ,e and m, 1 ~ ,e <m ~ n are called the pivot-mm pair of T,em and the elements t,e,e' t.em' tm£ and tmm are called the

Jacobi.;..parameters of T.em• The class of shear matrices with pivot­

pair (,e,m) will be denoted by if .£m· [.em , ~m and U .£m are

the classes of shear matrices with pivots .£and m which are uni­

modular (i.e. jdet(T£m)1 = 1), orthogonal and unitary respectively.

Definition 0.8. The matrix

will be called the (..e,m)-restriction of A.

Definition 0.9. 1

S(A) := ( 2! la. ·12Y2. ifj J.J

S(A) will be called the departure of dia,gonal form of A

0. 2. A survey of Jacobi-Hke algorithms

In a Jacobi-like procedure for the computation of the eigenvalues

A1

, A2

, ••• , An of a matrix A of order n a sequence

A=A,A,A, 0 1 2

is constructed in which the matrices ~

18

(k) (aij ) are recursively

Page 20: An eigenvalue algorithm based on norm-reducing transformations

defined by the relation

-1 Ak+1 := Tk ~ Tk (k 0,1,2, ••• ).

The matrix Tk is a shear matrix with pivot-pair(..ek, ~) and Jacobi­

parameters

(k) (k) t = p

..ek,..ek k t ..ek,~ qk

(k) (k) t ~,..ek rk t

~·~ = sk.

The indices ..ek and ~, 1 ~ ,ek < ~ ~ n constitute the pivot-pair

of the k-th iteration of the Jacobi-like process. The choice of

the successive pivot-pairs (..ek' ~) is called the pivot-strategy

of the process. In several Jacobi-like processes the pivot-pairs

are selected in some cyclic order. We mention especially the se~

pivot-strategy indicated by the scheme:

(£ ' m ) 0 0 ( 1 '2)

[<'k·V1l (..ek+1'~+1) ~ (..ek+1 ,£k+2)

( 1, 2)

The method of Jacobi ([16], 1846) is

' if

,if

if

one

..ek < n-1, ~ <n

,ek < n-1, ~ n (0.2.1)

,ek n-1, ~ n.

of the few efficient

methods of solving the Hermitean problem which existed before 1950.

After its rediscovery in the late forties several modifications

and generalizations of this method have been proposed,

1. Hermitean matrices

In the Jacobi-procedure for the computation of the eigenvalues of

a Hermitean matrix the shear matrices Tk are unitary and the

Jacobi-parameters pk' qk' rk and sk are chosen to minimize the

* departure of diagonal form of ~+1 = Tk ~Tk. By minimizing this

19

Page 21: An eigenvalue algorithm based on norm-reducing transformations

departure the element a~k+1 ) of A 1

is annihilated. Therefore ""k'~ le+

the decrease of the departure of diagonal form equals 2ja(k) f2 •

~·~ a) In the classical Jacobi-process ([9], [12]) the pivot-pair

(tk, ~) is chosen so that

I a ( k) I = max ( I a~~) I ) . £k,~ i < j ~J

b) In the serial Jacobi-process the pivots are chosen in confor­

mity with rule ( 0. 2. 1 ) [ 13] •

c.) In the serial Jacobi-method with threshold t the pivots run

serially through all superdiagonal positions of the matrix,

except those for which /a~~)j < t ([24], [29]). l.J

If all off-diagonal elements of A(k) are smaller in modulus

than t, then the threshold t is lowered.

Each of these pivot-strategies gives rise to a convergent process:

if the values of the Jacobi-parameters for which a}k+1) = 0 are

chosen in a reasonable way, then lim \: = diag(A.)k'~ [s]. k-oo J

Moreover, the asymptotic convergence is quadratic ([17], [18],

[ 3 3]).

2. Normal matrices

In the extension of the Jacobi-procedure to the case of normal

matrices, as proposed by Goldstine and Horwitz [10], the shear

matrices Tk are unitary and for each k the Jacobi-parameters of

of Tk are chosen so as to minimize ~(\:+1 ) = ~ ja~~+1 )j 2 • if j l.J

Although at each step the decrease of the departure of diagonal

form is optimal,Voyevodin has exhibited a class of matrices, for

which, independently of the pivot-strategy, this process is sta­

tionary before the diagonal form is reached [32]. To prevent sta­

tionarity, Goldstine and Horwitz have modified their algorithm,

and they have shown that with this modified algorithm

20

Page 22: An eigenvalue algorithm based on norm-reducing transformations

lim k -+C'O

(Aj). Ruhe [25] has proved that if such modifica-

tions are superfluous, the convergence is quadratic.

3. Triangularization by unitary shears

Greenstadt [11](1955) and Lotkin [19] (1956) have generalized the

Jacobi-procedure to arbitrary matrices. Their algorithms,according

to a made by J. von Neumann, are based on Schur's lem-

ma (theorem 0.5): for each matrix A there exists a unitary ma­

trix U for which holds that U*AU is triangular. In both general­

izations the shear matrices Tk are unitary. Greenstadt determines

the Jacobi-parameters at the k-th stage of the process in such a

way that a(k+i) = 0, Lotkin, on the contrary, determines them so ,ek'~

that . Z. la~~+1 )1 2 is minimal. For some matrices, however, the ~ <J ~J

the sequences {~} generated by these methods are not convergent

whatever pivot-strategy is used([ 1 ], [2]).

4. Norm-reducing by non-unitary shears

In 1962 in her paper "A Jacobi-like Method for the Automatic Com­

putation of Eigenvalues and Eigenvectors of an Arbitrary Matrix''

[4] P.J. Eberlein introduces a norm-reducing process by transfor­

mations with non-unitary shears Tk, The underlying idea is that

(in conformity with theorem 0.9)

inf T regular

n z (0.2. 2)

Moreover, if and only if A is diagonalizable, there exists a non­

singular matrix T such that T-1AT is normal; in that case the in­

fimum in (0.2.2) is assumed. So the aim is to construct recurs­

ively a sequence A0

=A, A1

, A2

, ••• , where ~+i := T~1 ~ Tk, so

that um n~n~ k .... eo

n =r.IA·I~

j=1 J

In terms of definition 0.5 and according to theorem 0.12 this se­

quence {Ak} converges to normality.

21

Page 23: An eigenvalue algorithm based on norm-reducing transformations

a) In Eberlein 1s method ~+1 is produced from~ in two steps:

"' -1 -1 -1 ~ ~ := ~ ~~ , ~+1 := ~ sk ~ sk ~ (0.2.3)

with

~and ~unitary shears and Sk a norm-reducing non-unitary

shear, all having the same pivot-pair (tk'~). ~is chosen

such that C0 n c where ,..k,A;k mk,~

c

This pre-treatment facilitates the construction of a suit-·

able norm-reducing Sk. The Jacobi-parameters of Sk are chosen

in the following way:

(

Pk qk) (cos""' i e

1

"'sinh~ ' rk sk -ie -itj,ksinhcpk cosh<pk }

~ and <)lk real.

In order to minimize ~~~+1 11E , considered as functions of

and ~k' two simultaneous quartic equations have to be solved.

Since this is not easy, Eberlein gives an approximation

(~k' ~k) of the solution (~k' $k). The norm-reducing shears

Sk corresponding to (~k' $k), in combination with the plane

rotations ~ suffice - independently of the choice of the

unitary shear~- to obtain a sequence {Ak} which converges

to normality, provided the pivot-pairs (,ek'~) are chosen

appropriately.

b) In Rutishauser's norm-reducing algorithm [28]the transforma­

tion with Tk is also performed in two steps as in (0.2.3).

The unitary shear 1\: annihilates the element c,e of f"'V f"V *,......, ,.._.J ""'* k' ~ C = ~ ~ - ~ 1\· The non-unitary shear Sk is a diagonal

matrix, which scales-~ in the following manner.

If r~t t I ~ I~ I. ' then the lengths of the t-th column k' k ~'~

and the t-th row are made equal, else this operation is ap-

plied to the m-th column and the m-th row. Rutishauser states

22

Page 24: An eigenvalue algorithm based on norm-reducing transformations

that the sequence {~} obtained in this way converges to nor­

mality.

c) Voyevodin [ 31] proposes to use the following Jacobi-parameters.

-1 The parameter \:is chosen to minimize i[Tk A TklrE

The convergence to normality has been proved.

d) Osborne 1s equilibration [23], too, is based on the principle

of norm-reduction. The aim of Osborne's algorithm is to im­

prove the condition of the eigenvalue problem. In each step

of the process the Euclidean lengths of a certain row and its

corresponding column of the transformed matrix are made equal,

Osborne has proved that if A is irreducible, then there exists

a non-singular diagonal matrix ~ such that the diagonal ele­

ments of C(D-1AD) are zero. With sequential pivoting the pro-

cess mentioned above constructs iteratively such a

matrix D. The matrix D-1AD is called equilibrated,

diagonal

In the

class of matrices similar to A by diagonal transformation, -~ the eq,Jilibrated matrix D 'AD has a minimum Euclidean norm.

5. Diagonalization and combination of no:t'lll-reduction and

diagonalization

If !:J.(A) is small in relation to ILil,.'l~ then the matrix A is called

almost normal. According to the corollary of theorem 0.10 such an

almost normal matrix can be unitarily transformed into an almost

diagonal matrix. The "almost-diagonalization" of an almost normal

matrix is the second stage of a Ja9obi-like process for

arbitrary matrices and follows the process of norm-reduction. In

Eberlein 1s process the diag\)nalization is already promoted during

the norm-reduction stage. For that purpose the unitary shear~ in

(0,2.3)is chosen such that the departure of diagonal form of

~1 (s~1 ~ Sk) ~ is minimal. However, Voyevodin 's counterexample

23

Page 25: An eigenvalue algorithm based on norm-reducing transformations

[32] shows that the global convergence to diagonal form of the se­

q·.:ence f-\:}, n1:"':dned with these {~}, cannot be proved sin.:;e the

Eberletn algorithm is a generalization of the Goldstine-Horwitz

proced:;re. Ruhe [26] has shown that if the sequencE {.I\}, gener­

ated ty Eberlein 1 s norm-reducing diagonalizing algorithm, con­

verges to diagonal form, then the convergence is quadratic.

0. 3. Summary

In chapter we investigate the norm-reducing shear transformations

on the pivot pair (-e ,m), applied to a real matrix A. It is shown

that, in consequence of the invariance of the Euclidean norm of a

matrix under orthogonal transformations, each shear in a class of

w1~at will be called row congruent shears brings about the same norm

reduction. This class is determined by what will be called its Eu­

clidean parameters (x,y,z), x> 0, y > 0 (definition1.2). With the

Euclidean parameters of a shear Tim in such a class we find a simple

expression describing the Euclidean norm of the transformed matrix

(theorem 1.2). If the shears are restricted to be unimodu­

lar then :1 T;~AT );mll~ is a quadratic function of x, y and z defined

on the hyperooloid xy- "" 1. In theorem 1.4 it is shown that, ex­

cept for a number of particular cases, this quadratic attains its

infim~m on the hyperboloid for finite values of x, y and z.

:r:r: section 1. 3 we describe the algorithm to compute the Euclidean

(x,y,z) of the class'»1m(A) of row congruent unimodular

optimal norm-reduc shears (theorem 1.5). For the computation of

tLese parameters we have to determine a real root of a quartic

equation, which is uniquely localized by an inclusion theorem (lem­

ma 1.5).

TLe particular case that the infimum of the quadratic on the hyper­

b:;loid is not assumed for finite values of the Euclidean parameters

is fully described in section 1.4.

24

Page 26: An eigenvalue algorithm based on norm-reducing transformations

In section 1. 5 it is shmm that after optimal norm reduction by a

unimodular shear on tile pivot-pair (t,m) the commutator C 1 of the

transformed matrix has the properties c Jm = O, c tt = cr:rrn.

In chapter 2 we the complex norm-reducing shear trans-

formations on the pivot-pair (t,m) applied to a complex matrix A.

As in the real case, each shear of a class of what will be called

row congruent shears brings about the same norm reduction. This

class is again determined by its Euclidean parameters (x,y, z), where

now x and y are real and positive, but z is complex (definition

2. 2). the Euclideaz: parameters of a shear T ,em in such a class

we find an expression, less simple as in the real case, describing

lj _, 1[ ( ) 1 f h , T .£mAT tm: E theorem 2. 2 • In order to s imp i y this express ion, t e

matrix A is pre-treated by a unitary shear U.£m' so that

(U~~AUEm)m.£ = 0. In theorem 2,) it is shown that if A is a pre­

treated matrix (i.e. am.£ 0) and T t is a unimodular shear on the

pivot-pair (t,m), then IIT;~AT,emlli isma quadratic function of x, y,

z and;;, defined on xy-jzj 2 1, x> 0, y> 0, z complex,

As in the real case, this quadratic attains its infimum on

xy- I zl 2 = 1 for finite values of x, y and z unless the matrix A

satisfies particular conditions.

In section 2,) we ciescribe the algorithm to compute the Euclidean

parameters of the class irltm(A) of complex unimodular optimal norm-

reduc shears on the pivot-pair (t,m) (theorem 2.5).

TI1anks to the pre-treatment of the original matrix, it is possible

to transfer the algorithm for real matrices to the complex case.

In section 2.5 it is proved that for a real matrix A the Euclidean

parameters of the class of complex unimodular optimal norm-reducing

shears are real.

0.3. 3.

In chapter 3 the effect of consecutively applied unimodular optimal

25

Page 27: An eigenvalue algorithm based on norm-reducing transformations

norm-reducing shear transformations is investigated.

Let the pivot-pair (£,m) be chosen so that

C equal to A*A-AA*. Let T.R.m E~m(A) and A' = T~~ATtm' Then

theorem 3.2 gives a lower bound for the decrease of the Euclidean

norm effectuated by Ttm:

Our proof of this result is essentially the same as that of Eber­

lein [4]. Since WB make use of the Euclidean parameters (x,y,z) of

the shears involved, our calculation of Eberlein 's estimate for the

optimal decrease of the Euclidean norm is considerably more simple

than her ovm.

We use this estimate in the proof of the convergence theorem. Let

(,e ,m ),(.e ,m), ••• be a sequence of- pivot-pairs. Let A :=A and 1 1 2 2 0

~ := T:1

\: T 0 (k"' 0,1, •.. ), where T .e is a unimodular ""k'~ _, ""k'~ k'~

optimal norm-reducing shear on the pivot-pair (,ek'~). In theorem

0.3 we prove that if the pivot-strategy is so that for each k:

then the sequence {~} converges to normality in the sense of def­

inition 0.5.

0. 3 ·4.

As a consequence of the convergence theorem of chapter 3 we find

that for each e > 0 and each matrix A there exists an integer k and

a normal matrix N with the same eigenvalues as A, so that for the

matrix \: obtained after k norm-reducing similarity transformations

26

Page 28: An eigenvalue algorithm based on norm-reducing transformations

where ~(~) is the of normality of Ak.

Therefore, in 4 we consider a Jacobi-process which almost

diagonalizes an almost normal matrix A.

In the first part of this process t!:e Hermi tean part of A is almost

diagonalized. The resulting matrix, A1

(say), is shown to have ar:

almost block stru.cture, the Hermitean part of a

block being almost a multiple of the unit matrix (lemma 4.1). As a

consequence, it is shovm that the skew- Hermitean parts of these

diagonal blocks can be diagonalized by a second sequence of Jacobi

rotations without disturbing the "almost diagonal" character of the

Hermitean parts of the diagonal blocks (lemma 4. 2). Let be the

resulting matrix after this second half of our process. 'Ihe depar­

ture of diagonal form S (A2

) of this ultimate matrix proves to be

bounded by a function of

(i) the

(ii) the

of normality of the original matrix A;

of diagonal form of the Hermitean of the

matrix A, obtained after almost diagonalizing the Hermitean 1

part of A;

(iii) The departures of diagonal form of the skew-Hermi tean parts

of the diagonal blocks of

This function tends to zero if each of these quanti ties tends to

zero (theorem 4.2). If for real matrices we want to use only real transformations, a

somewhat more complicated result is obtainable. of

the symmetric part of a real almost normal matrix A results in an

almost block diagonal matrix which is an almost canonical

form, unless, if ~ ~ iv is a

of A, there exist yet other

orem 4.6).

of complex eigenvalues

(s) with real part(s) !J. (the-

Already the norm-reducing stage of the eigenvalue procedure,

diagonalization can be promoted by executing the norm-reducing shear

27

Page 29: An eigenvalue algorithm based on norm-reducing transformations

transformation with that element T.£m E m.£m(A) that, moreover, min­

imizes the departure of diagonal form of the transformed matrix.

This element will be called the diagonalizing representative of

~m (A).

TLe problem of the numerical stability of a single optimal norm­

reducing shear transformation, executed in floating point arithme­

tic, is considered in chapter 5. It proves to be possible to per­

form the transformation with the diagonalizing representative

TimE rtl:em(A)

mation e

in such a way that the actual result of the transfor­

T~~(A +F)Tim + G, where IIFIIE is small relatively to

is small relatively to llTi~ATimiiE' This result may contribute to explain the accuracy of the solution

of the eigenvalue problem which was observed during our numerical

experiments with procedures based on the algorithms described in

this thesis.

28

Page 30: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 1

REAL NORM-REDUCING SHEARS

I. 0. Introduction

In this way

one step of the norm-red:.wing Jacobi-like process, and we shall

determine tf,e norm-reducing unimodular shear similarity

transformation for the real case. Since for one transformation

the pivots £ and m are fixed, we shall omit the subscripts when

no ambiquity arises.

L L Row congruency and Euclidean parameters of a shear Let To be a shear rr,atrix with pivot-pair (,e,m) and Jacobi-

)'"m parameters p, q, r and s. So the (,e,m)-restriction of T hn is

(: :) . Let d := det(T£ ),

m_1 non-singular T£m

pivot-pair (£,m).

(1.1.1)

thus d ps-~r. Since T£m is supposed to be

exists. Tim- 1 is also a shear matrix with

The (i,m)-restriction of Tfm-1 is

(1.1.2)

In this chapter we asswne that the matrix A and the matrix T.£m'

by which A is transformed, are real matrices. In the description -1 of the Euclidean norm of A T im' we shall try to take into

aoco~~t the invariance of this norm under orthogonal transforma­

tion. In particular, if Q.£m is an orthogonal shear then we have T -1 -1

for each shear T n : IIQ, 1' n A •J' a QD liE liT 0 AT 0 11. Hence -vm .zm ,vm "'m ,m ,vm ,vm the optimal norm-redc1cing shear is determined except for an ortho-

gcnal factor Q n • xm

29

Page 31: An eigenvalue algorithm based on norm-reducing transformations

Definition 1.1. The matrices S and T will be called row congruent

if S = TQ for some orthogonal matrix Q.

Theorem 1 .1. S and T are row congruent if and only if SST TTT.

Proof, If S = TQ, with QQT

For the proof of the sufficiency of the condition, we make use of

the polar factorization ([20], page 74) of the matrices S and T.

Let S = PU, T =RV, where P and Rare positive semi-definite ma­

trices, U and V orthogonal m:ttrices .Since P and R are complet

determined by SST and TTT respectively and the latter are equal,

P R. So S = PU = RU = TV-1u. Hence S and•T are row congruent. D

The theorem shows that the class of matrices row congruent to T

is uniquely determined b,y the elements of ; they p~rametrize

the equivalence classes into which the full linear group of non­

singular matrices is d·ecomposed by row congruency.

Now we consider row congruency for shear matrices with pivot-pair

(t,m), If p, q, rand s are the Jacobi-parameters of Ttm then the

(t,m)-restriction of Ttm T~ is

pr + qs)

r2+ (1.1.3)

Pefinition 1.2. The quantities

x := x(Ttm) := p2+ q2

y := y(Ttm) := r2+ s2

z : z(Tfm) pr + qs

will be called the Euclidean parameters of Ttm'

According to theorem 1.1. the Euclidean parameters (x,y,z) of T1

m

determine the class of shears row congruent to T£m' 'rhis class

will be denoted asl't_gm(x,y,z).

Lemma 1 .1, The Euclidean parameters (x,y,z) of a shear Tfm

30

Page 32: An eigenvalue algorithm based on norm-reducing transformations

the inequalities

X > 0, y > o, (1.1.5)

Conversely, if x, y and z ( 1 .1 • 5) , then they determine the

class ~" (x,y, z) of shears on the pivot-pair Ce ,m). This class ,vill

has an upper and a lower representative, B£m and

, with (t,m) restrictions

yi y-iz) A ()

le_ and L£m = (1.1.6) 0 y2 X

a) From (1.1.5) and definition 1.2·we find for the (£,m)­

restriction of T£mT£;

(

+ q2

pr + qs

Since is non-singular T, T"T is ,c;ill hill

x > 0, y > 0, xy - z2 > 0.

b) If (x,y,z) satisfy (1.1.5) then the shear

(x+>~

(x+y+2'/ xy-z2 ) --i z

restriction

H£m

z

Cx,y,z). This shear has Euclidean parameters

representative of~" (x,y ,z). "''m

definite, and hence

with (£,m)-

) is the s;;rmmetric

c) From (1.1.6) we see immediately trJBt Bn E 1(," (x,y,z) and ,m .«ill

Corollary 1. For each shear Tlm E ~tm(x,y,z) there exist ortho­

gonal shears Q£m and Rim such that

T£m = B£m Q.£m = 1tm R£m'

where B£m and L£m are triangular shears with (£,m)-restrictions as

31

Page 33: An eigenvalue algorithm based on norm-reducing transformations

mentioned in (1.1,6).

Corollary 2. If T 0 E~" (x,y,~), then det2 (T") = xy­~ . ~

In the sequel we mostly use unimodular shears. Then the (t,m)­

restrictions of the triangular represr.::ntatives are

Bern (y-> y~z) and L;;m = ( )'

0 l 0 y2 X ""2z X

L 2. Similarity transformations by real unimodular shears In this section we shall consider the similarity transformation

by a real shear T£m with pivot-pair (£,m) and Jacobi-parameters

p, q, r and s.

Let

and

d := det(T;;m) = ps- qr

A '·. T-1 AT .em ,em'

(1.2.1)

(1.2.2)

The elements of Ar will be denoted by a!., i = 1,2, •• ,n,j=1, ••• ,n. ~J

Only the elements of A in the ;;-th and m-throws and columns are·

affected by the similarity transformation with T1m.

For the elements of A1 we find with (1.1.1) and (1.1.2)

and

a! J.ffi

p + r a aJ~= im' "'""

aJ,e (ps au"" qr amm + rs a;;m- pq am;;)/d

a' = fia - q2 a_ 0 + qs(a. 0 - amm)}/d :.em ' J!m uw "'"'

a'm:t {p2

ame- :t'a£m- pr(a;;;;- amm)}/d

a~ (ps amm- qr au- rs atm+ pq am;;)/d

a!. = a ... lJ lJ

otherwise.

(1.2.3)

(1.2.4)

Page 34: An eigenvalue algorithm based on norm-reducing transformations

2 I 2 order to simplify the formulae for IIAI!E and. IIA liE we introduce

the followir..g notation.

n c z: a. aik

i=1 ~j

if £,m

n ·- z: a ..

i=1 Jl

if'x,m

n cr : z: a~

i, j=1 l.j

ilt,m jf,e,m

I The same functions of the transformed matrix A

I i be denoted by Cjk' Rjk' e' and cr 1 respectively.

(1.2.5)

-1 T .£m AT .£m will

For convenience and for simplicity of the formulae we will not men­

tion the dependence of these parameters on A (resp.T,e;1 AT,em)'

,e and m.

We now find

(1.2.6)

Obviously, a is an invariant of A under similarity transformations

b;y shears with pivot-pair (,e,m). Since e = (71.~ 1 )) 2 + (A.~2 ))2 (where (11 (2\ hill hill

A., ' and A, 1 are the eigenvalues of the (,e,m)-restriction of A), ,,;m Lill

e is also such an invariant : e e 1 • In order to determine !lA 1 !1 2

E we (a 1 - a 1 f and C t + C I + R t + R 1 • tm mt U mm U mm

TheorEilm 1.2. I!T.£;1 A'r,emll~ is in terms of the functions

of A defined in (1.2.5) and the Euclidean parameters of T.£m'viz.

Page 35: An eigenvalue algorithm based on norm-reducing transformations

Proof. From (1.1.4), (1.2.3) and (1.2.4) we see

C1 + C1 = Le mm

and

n 2: { i=1

if'..e,m

n

;(.

)a. + 2(pr+qs)a. 0a. } 1m l-v 1m

(1.2. 7)

R 1 + R1 = ,e,e mm z i"'1

if'..e,m

- 2(pr+qs)a 0 .a . }/d2 "'1 m1

(Rmmx + R,e,eY - 2R,emz)/(xy-z2).

Since L ~

IIA 1 ll CJe + cr:nn + Rj,e + Rr:nn + (aJm- a~) + u + e,

formula (1.2.7) is obtained. 0

In order to determine inf

T..eJ ~m (liT~: AT ,em liE), the rational function

in the right-hand part of (1.2.7) has to be minimized in the r.alf-

cone

x > 0, y > 0, xy - z2 > o.

Since the determination of the values of x, y and z minimizing

this rational function, is rather complicated, we shall henceforth

restrict ourselves to unimodular shear matrices T;,m• With this

restriction on T 0 we may restate theorem 1.2 as : hm

Theorem 1.3. If T1m is a-unimodular shear with Euclidean para­

meters (x,y,z), then

(1.2.8)

34

Page 36: An eigenvalue algorithm based on norm-reducing transformations

where

· f(x,y,z) := ax + ~y + 2yz +(-XX+ ~y

with n z (a~i+a2 .), ~:=

i=1 ~ ffil

n 2: (a~.+a~ ),

i=1 hl lm

n y:= Z (a, 0a. -a

0.a .)

i=1 lh lm hl ml

,m if£,m if.£, n; (1.2.10) and

A := a£m' (1.2.11)

In order to determine inf (liT 1;1 AT );milE), we have to mi.nimize

T ..emE ~£m the function f(x,y,z) with side conditions

X> 0, y > o, xy- = 1. (1.2.12)

Definition 1.3. The subset ~:={(x,y,z);x>O,y>O,xy-z2 = 1} of

R will be called the positive sheet of the hyperboloid xy-z2 = 1. 3

Previous to presenting in section 1.3, an algorithm to compute

the values of x, y and z minimizing f on '1{ , we shall demonstrate

some properties of the coefficients of f and then we :?.hall estab­

lish a sufficient condition for which the infimum of f, on i.he

surface'J/ , is assumed for finite x, y and z.

In the we make use of the notations introduced in (1.2.10)

and (1.2.11) and, moreover, we define

D ·- a~ - - yv,

E 2 V + 4il.fh (1.2.13)

F 2 ·- - y

Lemma 1.2. The quantities a, ~, D, E and F defined in (1.2.10)

and (1.2.13) have the following properties

(i) o::;;. O, ~ :;;. O, F :;;. 0;

(ii) D2 + E F :;;. 0;

(iii) if E.< 0 then D 0 implies F = 0, o: = 0, ~ = 0;

(iv) iff E > 0 (E = O, E < 0), then the (,e,m)-restriction

of A has two different real (two equal real, two com­

plex conjugate) eigenvalues A~~) and A~!) respectively.

35

Page 37: An eigenvalue algorithm based on norm-reducing transformations

Proof, (i) a and pare non-negative since they are sums of

squares, In order to show F ~ 0, the elements in the ~ -th column

and the m-th row of A, not belonging to the (t,m)-restriction of

A, are considered as components of a vector in R2n_4

• In the

same way we consider the elements in the m-th column and the ~-th

row, not belonging to A£m' as components of a vector in R2n_4

From the inequality of Cauchy-Schwarz follows for these vectors

2 F = ap-y = n n n

[ ~ (a~ o+a2.) ][ ~ (a: +a~.)]-[ ~ (a. a. -a a; a .)J~o.

i=1 lh m1 i=1 1m ..vl i=1 u 1m hi. m 1

if't,m il,e,m i;h,m

(ii) If ap f 0 then

D2 + EF = (afl-pA-yv) 2 + ( }+4At.t) (o:~/) 2

o:p{v- ~ (o:~J-pA.) f+ a~? (o:IJ+0A.)2 ~ 0.

If a0 0, then F = o, thus D2+ EF = ~ 0.

(iii) If E < 0 and D = o, (i) and (ii) imply F = 0, hence a0=y2•

In order to prove o: = p 0 we make use of the fact that a!J-phyv.

This implies that (o:ll +0A.) 2 =(o:IJ-pfl.)2 +4a!34 "'¥ 2 ( l +4A.JJ.)

Hence E < 0 implies that y ~ 0 and O:!J.+pA.= O.Since ~-pA.=yv= O, we

also have: A.=O.Since E <0 implies :\j.!<o;t follows that o:=p= 0.

(iv) E = (a£.£-amm)2

+ 4.-emam.-e= (a££+amm)2

- 4(a£.£amm~ a£mam£)

( A(1) + i\(2))2- 4 i\(1) A.(2)= (;\.(1) _ i\(2))2 • • em £m £m .£m ' £m .£m

A

Hence iff E > 0 (E O, E < 0), A£m has two different real (two

equal rPal, two complex conjugate) eigenvalues A.~!) and 11.}!) re­

spectively. D

We shall now demonstrate that if D and F are not both equal to

zero, there exists a compact set Qc~such that the minimum of

f(x,y,z) on;( is assumed in the interior of Q. To prove this

theorem we need

Page 38: An eigenvalue algorithm based on norm-reducing transformations

Lemrrra 1.3. If D and F are not both equal to zero, then for

(x,y' x+y- implies f(x,y, z)- eo.

Proof. We start by remarking that

D and F wonld be equal to zero.

>0, for otherwise both

Let 'f2 be the subset of R defined by 3

xy- o, X ? 0 1 y ? 0.

Then~ c 1l. In 1:<. we have

f(x,y,z) = ~(x,y,z) + w2 (x,y,z),

where cp : we + i3Y + 2yz, w := -.\x + !J.Y + v z.

In fl. we find for the linear part cp of f:

2cp=( ex+~) (x+y)+(cx-~) (x-y)+4yz?( ex+~) (x+y)-1 ex-~ llx-y l-4lrzi

2 2 l 2 o:+~)(x+y )- {( o:-13) +4y }2 {(x-y) +4

2 2 2 l =(cx+~)(x+y)-{(cx+p) -4F {(x+y) -4(xy-z) }2

?{x+y) {cx+f3-V(cx+~) 2-4F2 }?0.

IfF> then in~ and a fortiori on:;(, cp ?a(x+y), where 1l >0.

Thus if F > 0, then on{('

2 f = cp + w - co for x+y - =.

Now we conside:J? the particular case of F 0, D I 0. Then in;fl.:

2cp :;, ( {x+y-V(x+y) 2- 4(xy-z2

) }.

Hence ondf : cp >0 and in 14: cp?O.

(x,y, z)E 1< and cp 0 implies cxx+py = -2yz, hence

(cxx+py)2

= 4/ = 4cxpz2

..; 4cx(3xy.

Hence in~ cp 0 if and only if (x,y,z) = t((3,cx,-y), t ~ o. Along this line,~, the plane cxx+py +2yz = 0 is tangent to the

boundary a k of 11,. On t we find w = tD. Hence., since D I o, in tz - {(o,o,o)} : cp + I wl>o.

Let E be the intersection of the plane x+y = 1 and 12 • Since C is

a compact set and cp + lwi>O in t, continuous function cp+lwl

attains on't_ a positive minimum, say o(o>O). From the linear

homogeneity of cp and w it now follows that in 1<, : cp + lw r:;;;. o(x+y).

37

Page 39: An eigenvalue algorithm based on norm-reducing transformations

A fortiori cp+lw I~ o(x+y) on:(. Consequently

cp+w2~ cp+ lw I - i ~ o(x+y)- Hence also in the case F O, D f O,

on ~ q~+w2- oo for x+y .... oo • D

Theorem 1.4. If D and F are not both equal to zero, then the

infimum of f(x,y, on{ is assumed for finite x,y and z.

Proof. Lemma 1.3 asserts the existence of a. YJ.umber M > 0 such

that on~f(x,y,z) >f(1,1,0) if x+y >M. The theorem now follows

from the continuity of f and the fact that the subset of 1{ for

which x+y .;;; M is a ccmpact set. o

In the next section we make also use of 1

Lemma 1.4. (i) inf (ax+f3y+2yz) (x,y, z)E'/(

2F2 • The infimum is assumed

for finite x,y ,and z if and only if F f 0 or a + f3 = 0.

) inf ,,) -A.x-+tJ.y+vz)2

(x,y,z)Eq max(O,-E). The infimum is

assumed for finite x,y and z if and only if E I 0 o~ A. = ~·

Proof. (i)If F > 0 then, as we have seen in the proof of lemma 1.3,

on?{ a:x-r(3y+2y~~ .... oo for x1~ - 00• Hence, in this case the infimum

on :t{ of the non-negative function ax+f3y+2y z is assumed for finite

x, y and z. Using a Lagrange multiplier we easily find that 1

~n(ax+f3y+2yz) = 2F2r The coordinates of this unique stationary

1

point on:{are F'"'2 (f3,a,-y). If a+f3=0, then a=f3=y=0; then

ax+f3y+2yz = 0 for each (x,y,z)EJf.

If F = o, cx+f3fO,then on-:( ax+f3y+2yz ~ (a:+f3){x+y-V(x+y) 2 -4}~ 0.

We now consider the curve r c:(, d.efined in the following way :

:X:= -(a-f3)t+{1+(a+f3) 2 t 2 }~, y=(a-p)t+{1 +(a:+f3)2 t 2 }~, , t > o. On this curve r , using the fact that F = 0, we find

a:x+f3y+2yz=(a:-rp) {V 1+(a+f3) 2 t 2 -(a+f3 )t} .... 0 for t - 00•

d. Hence ~ (ax+f3y+2yz) = 2F2

(ii) If E > 0, then the plane -A.x~+vz=O intersects~. The point

with coordinates

Page 40: An eigenvalue algorithm based on norm-reducing transformations

:x:= . E+21l(J.L-A.) _ E-2A(u-A.) {E2+E(A.-!.!.)2}f' y- {E2+E(A.-!.!.)2}f

is an element of this intersection.

If E ,.. 0 then Afl ,..o and so we can apply, with appropriate modifi­

cations, the reasoning of (i)

~ (-A:x:+!.!.Y+vz) 2 = -E. D

Definition 1 .4. The class of unimodular row-congru.ent shears

for which the Euclidean norm of T 0-1A T

0 is minimal, will be

...vm ..vm

m

call~d the class of minimizing shears corresponding to A, £ and m,

and will be denoted by ~,em (A),

The Euclidean parameters (:x:,y,z) of the shears in ~,em (A) minimize

f on"!(. If D and F are not both to zero, then theorem 1.4. shows that mt .tm(A) is non-void, The particular case D = F = 0 will

be discussed in section 1.4.

1. 3. An algorithm for the real unimodular norm-reducing shears ars

In this section we suppose that D and F are not both equal to zero.

According to theorem 1.4.this is a sufficient condition for

f(:x:,y,z) to attain a minimum on~. We apply Lagrange 1s method of

multipliers to determine the point (x,y, z) on<?t where f is statio­

nery. We consider

g(x,y,z;p):=ax+py+2yz+(-A.x+uy+vz) 2+p(xy-z2 -1) (1.3.1)

In the stationary of f on'?( the partial derivatives ·Of

g(:x:,y,z;p) with respect to :x:,y,z and p are zero. Hence x,y,z and

satisfy

a - 2A.w + PY 0 (1.3.2)

€Sy p + 2/.l.w + pX 0 (1.3.3)

1 2gz = y + \'W - pz 0 (1.3.4)

and

=xy- - 1 0 (1.3.5)

39

Page 41: An eigenvalue algorithm based on norm-reducing transformations

where

w :~ -Ax + ~ + vz. (1.3.6)

We now eliminate x, y and z from the equations (1.3.2) to (1.3.6)

incl. Since px -0-211w , pY = -a+2J.. w and pz = y+vw, (1.3.5)g:Lves

On the other hand, we multiply (1.3.2), {1.3.3) and {1.3.4) by jl,

-A. and v respectively and add results. Then we find

ajJ - rn. - Y v + P w - (l +4 All) w ~ o.

With the notation (1.2,13), the equations

become 2 ?

p + E w- - 2Dw- F 0

( 0-E)w+D 0.

(1.3.8)

(1.3.7) and (1.3.8)

Elimination of w from (1.3.9) and (1.3.10) gives

(1.3.11)

The value of the Lagrangean multiplier p corresponding to the

minimum of f on~ will be called the feasible multiplier. This

multiplier satisfies (1.3.11). The next lemmas make it possible to

locate the feasible multiplier among the zeros of the quartic

equation,

Lemma 1.5. The Lagrangean multiplier p, corresponding to a

stationar,y point of f on~, satisfies the inequalities

p ~ - + min(O,E),

p < min (o,E)

(1.3.12)

(1.3.13)

Proof. We multiply (1.3.2), (1 .3.3) and (1.3.4) by x, y and z

respectively and add results. Then we find

o:x+t3y+2y z+2 ( -A.x-11J.y+v z )2 + 2 p 0.

According to lemma 1.4

~ (ax+t3y+2y z)

i~ ( -AX+iJY+ vz)2

40

max(O,-E).

Page 42: An eigenvalue algorithm based on norm-reducing transformations

Therefore

p 1 2 ],_

""2(a:x-t(3y+2y z)-( -l..x-ljly+vz) ~ -F2 +min(O,E). (1. 3.12)

If F I O, then (1.3.13) follows immediately from (1.3.12), which

for the present case proves the theorem.For the case F = O, D I 0

we have to show that E < 0 implies p < E and that E ;:;, 0 implies

p <0. Now let F 0, If E > O,then (1.3.12) shows p ~O.Since p 0

would imply w = D O, as is seen from (1.3.9) and (1.3.10), we

have p <0. If E ~ 0, then (1.3.12) implies p~E. Now p = E again

implies D = 0, as is seen from (1.3.10), hence p <E. This proves

(1.3.13). D

Ler:nma 1.6. +E 2 -2D F=O The equations { w w - have one and (p-E)w + D = 0

and only one solution (p,w) for which holds

p < min(O ,E).

For this root holds 1

-(F+rf /EY2 ~p~- if E > o,

if E = O,

(1.3.14)

(1.3.15)

(1.3.16)

E- 1 nl I E I P if E< o. c 1. 3 .n) Proof. We investigate the intersection of the quadratics (1.3.9)

and (1.3.10). To that end we distinguish three cases.

I E > 0. In this case the graph of (1.3.9) in the ~,w)-plane is

an ellipse with centre p 0, w = D/E. The half-axis parallel 1

to axis w = 0 is of length (F+ n2/E)2 • This ellipse intersects 1

the a:xis w 0 in the points having as coordinates p =£F2 , w=O.

The graph of (1.3.10) in the (p,w)-plane is an hyperbola with

asymptotes p = E and w = o. The hyperbola passes through the

centre cf the ellipse. In figure 1 the quadratics are sketched

for the case D > o. If D < 0 then the appropriate sketch is

obtained from figure 1 by reflection with respect to w = 0.

If D = 0, then the hyperbola degenerates into the lines p =E

41

Page 43: An eigenvalue algorithm based on norm-reducing transformations

42

and w = o. If Il I 0 then the quadratics have a unique point

of intersection S (see figure 1) in the half-plane p < o. The

-coordinate of S satisfies the inequalities l. l.

-(F+D 2/E)2 < p<- F2 ..;; O.

p

E

D,E> 0.

fig. 1.

If Il = 0 and F I 0, the degenerate hyperbola and the ellipse

have a unique point of intersection in the half-plane p < 0. . l.

The coordinates of this point are p = - F2 , w = o.

Page 44: An eigenvalue algorithm based on norm-reducing transformations

II -E o. In this particular case the graph of (1.3.9) in the

(p,w)-plane is a parabola which intersect the axis w = 0 in le

the po].nt having as coordinates p = :±:. p2, w 0. If D I 0,

then the parabola and the hyperbola (graph of (1.3.10)) r~ve a unique point of intersection S in the half-plane p < 0. The

p -coordinate of this point S satisfies the inequality .1.

p < F-2 .;;;; 0.

p

E=O I D>O .

fig. 2.

If D = 0 and F I o, then the parabola and the degenerate

hyperbola have a unique point of intersection in the half -r;lane 1

p < 0. The coordinates of this point are p = - F2 , w = o.

43

Page 45: An eigenvalue algorithm based on norm-reducing transformations

III E < 0. In this case the graph of (1.3.9) in the (p,w)-plane

is a r.yperbola with centre p "' 0, w D/E. This hyperboia

intersects the axis w 0 in the points having as coordinates ~ ~

44

p .:!:. F2, w "' o, and its asymptotes are p .:!:. I El 2 (w-D/E).

These asymptotes and the graph of (1.3.10) have a unique

of intersection in the half-plane p < E. The coordinates 1 1

of this point are p = D! !E!-2 , w "' (D) !E!2 •

We conclude from lemma 1.2 (iii) that D r O, forD= 0 would

F = o. Hence the

is not degenerated.

of (1.3.10), being an hyperbola,

As we see from figure 3 the byperbolae have a unique point of

intersection S in the half-plane p < E. The p -coordinate of S

satisfies the inequality

< p <E.

p

0>0 . E <0 .

fig. 3.

Page 46: An eigenvalue algorithm based on norm-reducing transformations

If F I 0 1 we can sharpen the upper bound of the p -coordinate

of s. For that purpose we :1

of p "" E - F2 and the :1

consider T,the point of intersection

of (1.3.10). The coordinates ofT

are p = E - F2 , w D Simple calculation show that the :1

left-hand part of ( 1 • 3. 9) in T equals (E-2F2) (D2 +EF) /F < 0, .l.

whereas in the point with the coordinates p = E-F2, w 0 the

.l. same function has the value -2EF2 > o. Hence the

of S is smaller than that of T if F >0. That means

:1 :1

E-IJI t.Eil-2 E;p ,;; E-F2 • D

From the lemmas 1.5 and 1.6 we that there exists one and

one Lagrangean multiplier that corresponds to a stationar,y

point of f on t( . With theorem 1 we find that in this unique

stationar,y point the minimum of f on ?I( is reached. Using the fea­

sible multiplier p, i.e. the root of (1.).11) which satisfies

(1.3.13), we find with (1.).2), (1.3.3) and (1.3.4) that f(x,y,z)

is minimal on ~ in the point

x= 2f.!D - @~p-E) p(p-E ' (1.3.18)

y -2XD - f 12:E) p(p-E (1.3.19)

z -vD + 1 12:E) p(p-E

. (1.3.20)

We summarize the results of the lemmas in

Theorem 1.5. If the quantities D and corresponding to the ma-

trix A and the pivot-pair (£,m) are not both equal to zero, then

the Euclidean parameters x, y and z of ~m(A) (being the values

of x, y and z which minimize f on :t ) may be computed from the

formulae (1.3.18), (1.3.19) and (1.3.20), where p is the unique

root of the quartic equation (1.3.11) for which root holds

p < min(O

45

Page 47: An eigenvalue algorithm based on norm-reducing transformations

1. 4: The particular case D = F , 0

In this section we investigate the properties of the

shears in the case that the functions D and F of the

matrix A and the pivot-pair (2,m) (see (1.2.13)) are to zero.

Theorem 1.6. Let D = F = 0. Then

}t'{o:x-tpy+2yz+(-A.x-tp.y+vz?} = max (0,-E).

This infimum is assumed for finite (x,y, z):: d( if and

a = 0 A [3 = 0 A (E .j 0 V A.= 1J.).

Proof. We two cases.

I a = [3 = 0 . Then y 0.

if

For this case the theorem has already been proved in lemma

1.4 (ii).

The infimum is assumed for the following values of x, y and z:

x= E+2u(u-A.) 1

, y= E-2A.(u-A.)1

, z• - (u-A.)v 1

,if E>O

{:Ff' +E(A.-tJ.f r2 {Ef +E(A.-tJ.l ra {:Ff'+E(A.-tJ.)2 }2

X= 1 y = 1 ' z = 0 if 0 1 1

X = 2j Ej-211-l.r , z =lEI-2 vsign(A.)if E < 0.

II a I 0 V [3 I o. According to lemma 1.2 (iii) this situation does not occur if

E < O.

Since D = F = 0, the line which the plane ax+fly+2y z 0

is tangent to the cone xy-~ O,ooinoides with the intersection

of the planes a x-tf3y+2y z 0 and - A.X-f].J.y+v z = 0. This line, whi oh

we shall denote by l , is in parameter form by

(x,y,z) = 2t([3,a:,-y ), t ~ o. Now we describe a curve r on:( of

which t:l is the asymptote. 1 1

r : x=- ( a-[3 )+{ 1 +(a+fl )2 t 2 }~' y=(a-[3 )t+{ 1 +(o:+fl i t 2 rr' z=-2yt' ~0. On this cure we find, using the fact that D = F = 0,

o:x+[3y+2yz = (o:+P)N1+(o:+[3)2 t 2-(o:+[3)t}- 0 fort- eo

Page 48: An eigenvalue algorithm based on norm-reducing transformations

and

Since on <1( ax+py+2y z > 0,

\Pf{ax+py+~z+(-Ax~y+vz) 2 } = o,

but this infimum is not assumed for finite x, y and z. ::l

Remark 1 • In the case D = F = 0 the

(p-E)2 (p2 (2p-E) = 0 has solutions p O,O,E,E. Thus the for­

mulae (1.3.18), (1.3.19) and (1.3.20) for the Euclidean parameters

x, y and z of the optimal norm-reducing shears are not usable.

Remark 2. If a = 0 A ~ = o, then each affected element of A not

belonging to A£m equals zero. Hence the investigation may be con­

fined to the (l,m)-restriction of A. If a 0 A p = 0 A E > 0,

f on//( determines a

class of shear similarity transformations. Each transformation of

this class symmetrizes the (l,m)-restriction of A, i.e. a£m a~.

If a 0 A p 0 A E = 0 t\ A. t ll, then the infimum of f on !( is

not assumed. This situation is connected with the defeotness of

Alm: this matrix of order two has two equal real eigenvalues and

is not symmetric.

If a 0 A p = 0 A E < 0, then the Euclidean parameters ~

(x,y,z) = rEI 42(2riJ.r, 2IA.I·, A.)) correspond to shears T£m'

such that the (l,m)-restriction of Tt~1A Ttm has Thfuxnaghan 1 s

canonical form:

where A.(1 ) is a complex tm

Remark 3. The particular case a + p f o, D = F = 0 will be il­

lustrated by an example ([3], pag. 272).

47

Page 49: An eigenvalue algorithm based on norm-reducing transformations

A

31 6

2

5

3C

7 2

5

-3 3

13

16

If (x,m) = (1,2), then a= p y 54, A= 6, ~ = 30, v = 24. Thus

D F 0, E = 1 296. Since a + p I 0, D = F = 0,

~ (flT1 ,;1

A T12fl~ = o + e = 1868.

This irSimum is not assumed as is seen from the solution

(x,y,z) = t(1,1,-1) of the equations a:x+py+2yz = o, -Ax+p.Y+vz = 0,

thus of the equations x + y + 2z o, -x+5y+4z = o. This line ,(3 :

(x,y,z) = t(1,1,-1) is generator of the as,ymptotic cone ~-z2 0

which for large (x,y,z) approximates the hyperboloid ~-z2 1 •

.l is an asymptote of the curve r c :(, defined by 2 _l. I

x y = {1+(108t) }2 , z -108t , (t ~ o).

Now we consider the uppertriangular shear T1 2 with Euclidean

' parameters (x,y,z) E r • Thus

A " (y~ ey4) T1,2

Then -1 6(y+z) (5y-z):Y1 1 1.

31-6zy -3(y+z)? -4(y+zr?

6y-1 +6 -1 _l.

7 zy 3y-2 -1 -=A( t). T1,2AT1,2 1 1

2y""2 2 (y+z )y ""2 13 -8

l 1

5Y""2 5 (y+z )y ""2 16 -3

For (x(t), y(t), z(t) r we find

1 ( -1) ~ ( -2) ;r(t}+ z(t2 0 (t-3/2) YTtT = 0 t ' y t = -1 + 0 t ~y(t)

Page 50: An eigenvalue algorithm based on norm-reducing transformations

and [y(t)+ z(t);~~)(t)- z(t)]

Hence

37 0 0 0

c 1 0 0 1 A(t)

0 0 13 -8 +O(t'"'2) , for t-oo

0 0 16 -3

;;;_;;;;;.::::;;,:;.::._.;I.:. From the theorems 1 .4 and 1.6 it follows tbat on 1r(' the

infimlUll of f(x,y,z) o:x-!f3y+2yz+(-AX"'J.!Y+vz? is assumed iff

D ~ 0 V F ~ 0 V ( a = 0 fl p = 0 f\ (E ,) 0 V A. ll ) ) ; then the minimum

is also the unique stationary point unless ex = 0 fl p = 0 f\ E > 0

(then the minimum value is assumed on the intersection of~ and

the plane -A.x-fjly+vz 0) or a = 0 fl p == 0 fl E = 0 fl A = ll (then f

is zero everywhere on:{ ) • The infimum of f(x,y,z) is not assumed onf'iff

D == 0 fl F = 0 fl (a I 0 V f3 I 0 V (E = 0 f\ A I ll ) ) • In this case f

bas not even a stationary point on 'f' .

1. 5. The commutator in relation to shear transformations

In this section we demonstrate that a stationary value of

1!T "-1

A T £ If~ 2 , considered as function of the fuclidean pararne1>e1:s ,~:;m m ~~- _

1 of T 0 , corresponds to annihilating a part of C(T 0 A T0 ) • .vm .vm .vm

Theorem 1.7. Let f(x,y,z) be the function in the right-hand part

of ( 1. 2. 7), thus f(x,y, z) = liT ,e~1A T .£milE 2 ~ (x,y, z) being the

Euclidean of T 0 • Let C(T-1A T0

)= (c!.). For the 1 .vm ,tm .vm lJ

elements of C (T ~' A T ..em)holds c };(' 1 =c lm 0 if and if the

Euclidean parameters x, y and z ofT~ are such that f is station­

ary in (x,y, z).

Proof. We assume that p, q, rand s are the Jacobi-parameters of

the shear T~. ~calculating the elements of c(T~1 AT~) we

find

49

Page 51: An eigenvalue algorithm based on norm-reducing transformations

c' mm

2c 1 £m

Since p2

2pq

= f X

2rs

we obtain c t-t

2pq

pr

qs

ps+qr

+ f y + f z

pr

qs

ps+qr

c' mm c1m 0 if and only if fx f y

f = o. 0 z

We would remark that the elements of the (.t,m)-restriction of C(A)

are obtained from (1.5.1) by substituting p = s = 1, q = r o. If C(A) = (cij), then we find c U fx(1,1,0), cmm fy(1,1,0)

and c-tm ifz(1,1,0). Thus the element of the (t,m)-restriction of

C(A) are the components of grad fin (x,y,z) = (1,1,0).

Now we restrict ourselves again to unimodular shears Tim" We in­

vestigate the properties of C(T;1 AT ) for the case ToE: ?'Jl;n (A), Nm m Nm Nm i.e. the Euclidean parameters (x,y,z) of T..tm are such that

f(x,y, z) o:x+~y+2yz +( -Ax+f.ty+vz) 2 is minimal on If.

Lemma 1 • 7. Let T £m be a unimodular shear w:i th Euclidean para-

meters (x,y,z). I,et C(T~1 A T 0 ) (c! .). For the elements of

-1 kill lJ C(T£m A Ttm) holds c£..e- c~ ctm = 0 if and if f is station-ary on/{ in (x,y,z).

Proof. We introduce new variables: t and w

X = t + w, y = t - w.

Since on~ xy - z2 = 1, x > o, y > 0, t!( is now given by

2 ~ . t = (1+w )2

• With the variables wand z we find a new expression

( ) -1 2 kw,z foriiT..em AT-tmHE:

50

Page 52: An eigenvalue algorithm based on norm-reducing transformations

Partial differentiation of k(w,z) gives

kw(w,z)= (a~) ~ +a-0+2{(~-A)t-(~+A)w+vz} {U~A w-(~+A) },

Simple calculations produce the relations between these

partial derivatives of k(w,z) and the elements of the (t,m)­

restriotion of C(T£~1 A T£m).

(

1z-(p2 +s2 -q2 -:I- )

pq-rs (

c' - c1

) ££ mm

2c 1 £m

pr-qs) (kw)

ps+qr k z

where p,q,r and s are the Jacobi-parameters of Ttm•

Since 1z-(p2 +s2 -q2 ) (ps+qr) -(pq-rs) (pr-qs) = t(ps-qr )(p2 +q2 +:I- +s2 )=t,

we conclude that cl£- c~ = elm = 0 iff kw = kz = o, i.e. iff f is stationa:cy on#( • D

Remark. If C(A) = (c .. ), then, by substituting ~J

( 1 • 5 • 3) , we find c" n.. c = k ( 0, 0) , c n = ik ( 0 , 0) •

Theorem 1 .8.

inf I! T..emE ~,em

NN mm W ,vill Z

Let C(A) = (c .. ). Then lJ

-1 A T_emiiE = I!A[E

if and only if c - c = c = o. ££ mm £m

, q=r=O in

Proof. If et£- cmm = c£m= o, then, as is seen from lemma 1.7,

f(x,y,z) is stationa:cy oncf'in (x,y,z) = (1,1,0).

There, as observed after theorem 1.6, f is a minirrum; hence

min (l!T£~1 A Ttmf!E) = lfAI[E •

T£mE t tm

Conversely, if inf (ffT£;1

AT£m!IE) [AI!E' thenf(x,y,z) is TtmE ~,em.

51

Page 53: An eigenvalue algorithm based on norm-reducing transformations

stationary on -:( in ( x,y, z) =( 1 , 1 , 0). From ~emma 1. 7 it now follows

that C 00 - c = c = o. 0 "'"' mm tm

If inf. nT -1

A T "m \lE = ffAffE' then it follows from Remark 4 T re ,em N

,em'- ,em

Remark.

after theorem 1.4 that Ttm Otm unless

a = 0 A~ 0 A(E > 0 V(E 0 A X=~)).

Finally we shall prove that if a matrix is non-normal, then for

some pivot-pair (t,m) et£- cmm and c£m are not both equal to zero.

Theorem 1.9. Let C(A) =(c .. ) be the commutator of A. A matrix A lJ

is normal if and only if for each pivot-pair (t,m) c,e,e- cmm= ctm = 0.

Proof. The necessity of the condition follows immediately from

the definition of a normal matrix. If for each (t,m) ct,e= cmm and

c£m = 0, then obviously C(A) = ai. However, tr(C(A)) = 0, hence

C(A) = 0 and then A is normal. 0

Page 54: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 2

COMPLEX NORM-REDUCING SHEARS

2. 0. Introduction

In this chapter we shall consider the norm-reducing process

for complex matrices. An algorithm analogous to that for the real

case as described in the preceding chapter, may be used to obtain

the optimal unimodular norm-reducing shear transformations. We make

use of the formulae (1.2.3) and (1.2.4) for the elements of the

transformed matrix

t -1 A Ttm A Ttm

but now Ttm and A are complex matrices.

2. 1. Row congruency and Euclidean parameters of a shear

Let Ttm be a complex shear with pivot-pair (t,m); p, q, rand s

are the Jacobi-parameters ofT. , thus . ..,m

(: :) (2.1.1)

and the (.e,m)-restriction of T£~1 is

( ~ y) .=.!:. :2. d d

(2.1.2)

where d:= det(T.em) = ps - qr.

As in the preceding chapter, we shall investigate the consequences

of the invariance of the Euclidean norm under unitary similarity

transformation.

53

Page 55: An eigenvalue algorithm based on norm-reducing transformations

Definition 2.1. The matrices Sand Twill be called row congruent

if S TQ for some unita~ matrix Q.

Analogously to the preceding chapter we can prove

Theorem 2.1. The matrices S and T are row congruent if and only

if SS* = TT*.

The row congruency is now considered for shear matrices with pivot­

pair (t,m). If p, q, rand s are the Jacobi-parameters of T$m,

then the (t,m)-restriction of T$mT£; is

( 1:1' _· lql'

pr+qs

Definition 2.2. The quantities

x := x(T£m) : JpJ2 + JqJ2

Y := y(T£m) := JrJ2 + JsJ2

z := z(T ) := pr+qs tm

will be called the Euclidean parameters of T$m'

(2.1.3)

(2.1.4)

LeiiliilB. 2.1. The Euclidean parameters (x,y,z) of a shear T$m satis-

fy the inequalities

X > 0, y > 0, (2.1.5)

Conversely, if x, y and z satisfy (2.1.5) then they determine the

class ae$m(x,y,z) of shears on the pivot-pair (t,m).

This class has an upper and a lower triangular representative

B$m and L$m respectively, with (t,m)-xestrictions

and Lim= (2,1.6)

54

Page 56: An eigenvalue algorithm based on norm-reducing transformations

Corolla:cy 1. For each shear T tmE t£ ,em (:x:,y, z) there exist unitary

shears Qtm and Rtm such that

where Btm and Ltm are triangular shears with (.£, -restrictions as

mentioned in (2.1.6).

Corollary 2. If TtmE ~ tm(x,y,z), then !det(T,em)l 2 == :x;v-lz! 2

In the .sequel we mostly use unimodular shears. Then :x;v-!zl 2 = 1

:: ~(ey~~,m)re;tric)ti: :~t:e(~~lar :)presentatives are

0 .) :x: 2 z x 2 •

2. 2. The complex unimodular norm-reducing shear transformation

In this section we shall consider the similarity transformation by

a complex shear Ttm with pivot-pair (t,m) and Jacobi parameters p,

q, r and s.

Let and

d := det(T,em) = ps-qr

I -1 A': T,em AT.£m.

(2.2.1)

(2.2.2)

The elements of A' are denoted b,y a!., i 1 ,2, ••• ,n, j 1,2, ••• ,n. ~J

We make use of the formulae (1.2.3) and (1.2.4) for the elements

of the transformed matrix.

In order to simplify the formulae for IIAII~ and IIA' 11~ we introduce

the following notations:

n

cjk := E a .. aik i=1 ~J

i,lt,m

n Rjk := E a.; ajk

i=1 Jl. (2.2.4)

i,lt,m

55

Page 57: An eigenvalue algorithm based on norm-reducing transformations

e ·- laul2 + 1amm12

n (2.2.5) a ·- 2:! la .. j2

. . 1 ~J l.,J= if.£,m jf..e,m

The corresponding functions of T..e:1 A T..em will be denoted by

C 'jk, R' jk, e 1 and o 1 • For convenience, we do not mention the

dependence of these parameters on A, Since .£ and m are fixed

during one step of the process, these indices are omitted where

no ambiguity arises. We now find

(2.2.6)

Theorem 2.2. I!T ..e:1 A T ..em!!~ is expressible in tenus of the

functions of A defined in (2.2.3), (2.2.4) and (2.2.5) and the

Euclidean parameters x, y and z of Ttm' namely

-1 R x + R ,eY-2 Re(R z) I!T.£mAT,emiiE=C,e_ex+CJ+2Re(C.£mz)+ mm :.e .£m +

:x;v-lz12 + la,emy+(a££-amm)zj jam.£x-(a..e..e-amm)zl

2- !z(a..e..e-amm)j

2 +

:x;v-1 z 12

With (1.2.3) we find

n 2:!

i=1 if£,m

+ (pr + qs)ai,taim+ (pr + qs) ai.£ aim }

56

(2.2.7)

Page 58: An eigenvalue algorithm based on norm-reducing transformations

+ qs)ao.a .+(pr + qs)an. a . "'~ nu "'~ m~ }

= {R X + mm

After some

( 1.2.4)

IPs - qrl2

- 2Re(R1mz) }/(xy- lzl 2).

~~·u~~.c but time-consuming calculations we find from

I a~,.el 2 +1 alm12

+1 a~1 2 +1 a~l 2 + la,.emY + (a££- amm)zl2 +

xy- lzl2

larntx-(a££- amm)il2-lz(a11- amm)1

2- 2Re(z2a1mam£)

+ • xy- lzl2

Since

I!T -1 AT 112 = C' +C' +R' +R' +la' r2+1a' 12+1a' 12+1a' 12+ a £m £m E U mm t . .e mm ££ ..em m£ I · mm '

we obtain (2.2.7). 0

We now have an expression of x,y and z describing I[T - 1 A T n 00 • tm ,.,mE This function has to be minimized in the domain x > 0, y > 0,

xy- lzl o. I~ order to simplify the variational problem we

the matrix A a pre-treatment. We transform A with a shear

Utm so that (U1m A U..em)m£ = o. Since the shears with pivot-pair

(..e,m) constitute a group, this pre-treatment causes no loss of

in the norm-reducing process. Let A be the ~~a-,c~c,~

matrix, then A = U1m A U£m" Let T ,em = U£m T £m" Then -1 rv "" -1 ""' "' rv

Ttm A T..em = T..em A Ttm and Ttm T1m U..em T£m T1m· U1m·

Consequently, if (x,y,z) and (x,y,z) are the Euclidean parameters

of T£m and T,em respectively, then

57

Page 59: An eigenvalue algorithm based on norm-reducing transformations

From the non-singularit,y of U~m it follows that the pre-treatment

corresponds to a non-singular transformation of the Euclidean ,..., -1 ""

parameters. Hence, liT~ A T JJm!lE' consi~ered a;:_ a function of

(x,y, i), is stationar.v if and only if liT ,e;1 A T ,em liE' considered

as a function of (:x:,y,z), is stationar,y in the corresponding point,

Since~- lzl 2 = 1 is equivalent to i y- fZI 2 1 (as is seen by . -~ -taking determinants in (2.2.7a)), !IT.em A T.em!IE' considered""'as a

function of the Euclidean parameters of a unimodular shear T,e ,

is stationar.v if and only if 1fT .e;1

AT .emi!E' considered as a f~c­tion of the Euclidean parameters of the corresponding unimodular

shear Ttm' is stationar,y.

The Jacobi-parameters (p,-r,r,p), where !PI unitar.y shear u)Jm have to be such that am.e = o, or,

(1.2.4), am£ p2

-(a.e.e- amm)pr- a.em r2 = o. This

1

1, of the

to

p : r = a ,e[ a mm .:t {a ,e[ a mm+ 4a £m am£ }13 2allli (2.2.7b)

The sign in pis chosen in such a way that p r is maximum.in

modulus. The relation (2.2.7b) which we have derived for p and r

is used to c(ompute a unitar.yi:h~ar)U£m with (JJ,m)-restriction

,. cos cp -e SJ.nq> •

u.em = -ie e sin cp cos q>

In the sequel we suppose that A is the result of such n~o-,c~~'~

with UJJm' Thus am£ = 0, Then theorem 2.2. can be re-stated as

If am£ = 0 then

!IT ~;1 AT mD~ = CUx + CJ + 2 Re(C.emz) +

+ Rp,j! + R~- 2 Re(R,emz)+ ia,emY + (ap,[ amm)zl2

~ - jzl2 + a + e.

In order to determine inf (!IT - 1 A T "m1I!E), the rational

T E Cf £m "' . Jlm £m

Page 60: An eigenvalue algorithm based on norm-reducing transformations

function in the right-hand side of (2.2.7b) has to be minimized

in the half-cone

X > 0, y > 0,

As in chapter one, we shall restrict ourselves to unimodular

shears T$m' With this restriction on Ttm we may re-state theorem

2. 2. a. as

Theorem 2.3. If amt= 0 and T$m is a unimodular shear with

Euclidean parameters (x,y,z), then

(2.2.8)

where

(2.2.9)

with

n n (X ·- l:: Clau12+ jami12) ~:= .E (I a im12 + I at i 12 ) ·-

i=1 i=1 if-t ,m irft,m

n (aiiim- a,e' a . )

(2.2.10) y := .E

i=1 ~ m1

i~.£,m and

ll == atm ' \1 := at.£ -a mm (2.2.11)

In order to determine int (ffT_e;1A T mffE), we have to minimize

T£mEv..em

the function f with side conditions

X> 0, y > 0, (2.2.12)

Definition 2.3. The subset#(:= {(x,y,z); x>O,y>O,:x;y-j z! 2 1}

of (x,y,z)-space (x and y real, z complex) will be called the

positive sheet of the surface xy- I z! 2 = 1.

In the sequal we make use of the notations introduced in (2.2.10)

and (2.2.11) and, moreover, we define

59

Page 61: An eigenvalue algorithm based on norm-reducing transformations

D ·- aj..l. .. yv

E ·- I vl2 (2.2.13)

F :=a~ - lrl2

Analogous~ to chapter one we can prove

Lemma 2.2. The quantities a, ~ , E and F defined in (2.2.13)

have the following properties

(i) a ~ 0, ~ ~ o, F ~ 0,

(ii) ~ = lx~~)- x~!)l 2 ,where x~:) and 1~!) are the eigenvalues of

A_em·

We shall now staroe a sufficient condition for the requirement that

the infimum of f on the surfacer is assumed for finite x, y and z.

Lemma 2.3. If D and Fare not both equal to zero, then for

(x,y, z)E: df, x + y- oo implies f(x,y, z) - oo •

The proof of this lemma is analogous to that of lemma 1.3.

Consequently we are able to state

Theorem 2.4. If D and F are not both equal to zero, then the

infimum of f(x,y,z) on~ is assumed for finite' x, y and z.

Analogous~ to.lemma 1.4 we can prove _;!,_

Lemma 2 • 4. ( i) 2F2• The infimum is inf (ax + ~y + 2Re(yz))

f' assumed for finite x, y and z if and only if F I 0 or a = ~ = 0.

(ii) ~ (lilY + vzl) = o. The infimum is assumed for

finite x, y and z if and only if E f 0 or ll - o.

Definition 2.4. The class of complex unimodular row congruent -1 shears T

1m for which-the Euclidean norm of T

1m A T1m is minimum

will be called the class of minimizing shears, corresponding to A,

£ and m and will be_ denoted by ~£m(A).

The EUclidean parameters (x,y,z) of the shears in ?n (A) minimize £m f on :t(. If D and F are not both equal to zero then theorem 2.4

60

Page 62: An eigenvalue algorithm based on norm-reducing transformations

shows that 17Z £m(A) is non-void. The particular case D = F = 0 will

be discussed in section 2.4.

2. 3. An algorithm for the complex unimodular norm-reducing shears

In this section we suppose that the functions D and F of the pre­

treated matrix A are not both .equal to zero. According to theorem

2.4 this is a sufficient condition for f(x,y,z) to attain a mini­

.mum ono/'. We apply Lagrange 's method of multipliers to determine

the point(s) (x,y, z) on r Where f is stationary.

We consider

g(x,y,z;p):= ax + py+ 2Re(yz)+ I~Y + vzl 2 + p(xy-lzl 2 -1). (2.3.1)

Let u := Re(z), v := Im(z).

In the stationary points off on~the partial derivatives of g with

respect to x, y, u, v and p are zero. Hence x, y, u + iv) aad p

satisfy

and

gx a + pY = o,

gy ~ + px + 2j~j 2y + 2 Re(~vz) = O,

i~) = y + ~~y + (jvj 2- p)z = 0

gp = xy - I zl2- 1 = 0.

Let w be defined by

w := ~y + vz. (2.3.6)

We now eliminate x, y and z from the equations (2.3.2) through

(2.3.6). Since px = - ·p - 2j ~~ 2Y- 2Re(~vz), py=-rx and pz=Y+IJ.vY+! vj2 z,

(2.3.5)

On the other hand we multiply (2.3.2) and (2.3.4) by~ and -v re­

spectively and add the results. Then we find

61

Page 63: An eigenvalue algorithm based on norm-reducing transformations

(2.3.8)

With the notation of (2.2.13) the equations(2.3.7) and (2.3.8) be-

come

p2 + Eiwl 2 - 2 Re(Dw)- F = O,

(p - E)w + D = O,

p real, w complex, (2.3.10) complex, E :;:;. 0.

Elimination of w from (2.3.9) and (2.3.10) gives

(p - E) 2(p2 -F) + IDI 2 (2p- E) = o. (2.3.11)

The value of the Lagrangean multiplier p corresponding to the mini­

mum of f on f will be called the feasible multiplier. This multi­

plier satisfies (2.3.11). The following lemmas make it possible to

locate the feasible multiplier among the zeros of the quartic equa­

tion.

Lemma 2.5. The Lagrangean multiplier p corresponding to a statio­

nary point of f ono/ is negative and satisfies the inequality

p <E; -

Proof. We multiply (2.3.2) and (2.3.3) by x and y respectively

and we take the real part of (2.3.4) which is previously multiplied

by 2z. Adding these results, we obtain, using (2.3.5),

a:x: + f3Y + 2 Re(yz) + 2 lilY+ vzl 2 + 2p = 0,

According to lemma 2.4, .J,_

~ (ax + f3Y + 2 Re(yz)) = 2 F2,

Therefore 1

p = -t( a:x: + !3Y + 2 (Re ( yz) ) - lilY + vz !2 < - i 2• (2.3.13)

If F > 0 then indeed p < 0. In order to prove p < 0 also in the

case F O, D r O, we assume p = O, F = O. Then, according to

(2.3.13) also 1JY + vz = w = 0 and we conclude from (2.3,10) D = 0,

Then F = O, D r 0 implies p < O. 0

62

Page 64: An eigenvalue algorithm based on norm-reducing transformations

Analogously to lemma 1.6 we state

Lemma 2.6. The equations (2.3.9) and (2.3.10) have one and only

one solution (p,w) for which p is negative. For this holds 1 1

-(F + ID I2/E)2 ,.;;;; pE;; - F2 ' if E > 0

1

p .;;;; - F2 , if E = 0.

_;!,.

(2.3.15)

(2.3.16)

Proof. If D = 0 then (p ,w) = (-F2 ,0) is a solution of the equa-

tions (2.3.9) and (2.3.10). This is the unique solution with neg­

ative p. If D I 0 we substitute in (2.3.9) and (2.3.10) 1

i arg D w = w e • Then we obtain the following equations for p and

w':

p 2

+ E I w 1 [

2 - 21 D [Re ( w' ) - F 0

(p-E) w1 + [D[ = 0. (2.3.17a)

(2.3.17b)

Since E ~ 0, we derive from (2.3.17b) that in the restriction

p < o,w' is real and positive. Thus for the case p < o,the qua­dratic equations become

p2 + E(w 1

)2

- 2[Diw'- F 0

(p-E) w' +ID[ = 0.

Consequently, the reasoning applied in the proof of lemma 1.6 can

be applied to prove the present lemma. D

From the lemmas 2.5 and 2.6 we conclude that there exists one and

only one Lagrangean multiplier that corresponds to a stationary

point of f on :If. With theorem 2.4 we find that in this unique

stationary point the minimum of f on~ is reached. Using the fea­

sible multiplier, i.e. the negative root of (2.3.11) which satis­

fies (2.3.12), we find with (2.3.2), (2.3.3) and (2.3.4) that

f(x,y, z) is .minimum on :fin the point

X=

y =

2 Re(~D)- §(p-E) p p-E)

ex --p

(2.3.18)

Page 65: An eigenvalue algorithm based on norm-reducing transformations

z z - vn + x~p-E) . p(p-E

We summarize the results of the preceding lemmas in

Theorem 2.5. If the quantities D and F, corresponding to a pre­

treated matrix A (arnt 0) and pivot-pair (,e,m), are not both

equal to ~ero, then the Euclidean parameters x, y and z of ~,em(A)

may be computed from the formulae (2.3.18), (2.3.19) and (2.3.20),

where p is the unique negative root of the quartic equation

(2.3.11).

It is now clear that on account of the pre-treatment performed on

the original matrix it is possible to carry over the algorithm

for real matrices to the complex case. Without this pre-treatment

the analogy with the real algorithm is lost and then it is much

more difficult to obtain the Euclidean parameters of the unimodu­

lar shears Ttm for which IIT,e~1 A T.eJI E is minimum.

2. 4. The particular case D F = 0

Analogously to the real case we investigate the properties of the

optimal norm-reducing shear similarity transformations in the

particular case that D = F = o.

Theorem 2.6. Let D = F = 0. Then

}1ft {a:x + !3Y + 2 Re( yz) + !IJY + vzl 2} = 0.

This infimum is assumed for finite (x,y, z) E cf' if and only if

ex = 0 11 ~ = 0 II(E > 0 V 11 = 0).

Proof. We distinguish two cases.

I. ex ~ = o. Then y = 0. For this case the theorem has already

been proved in lemma 2.4 (ii). The infimum is assumed for the

following values of x, y and z :

Page 66: An eigenvalue algorithm based on norm-reducing transformations

E + 21~1 2 E

x= (Ff+ El~f2)f' y= (E2+ Ej~l2)2 t z= 1!. E - V (E2+ El~12>* t

if E > 0.

X= 1 ' y = 1 ' z = o, if E ~ = O.

On r we find,

ax+~y+2Re(yz)

and

the assumptions about D and F:

(a+~){V1+(a+~) 2 t 2 -(a+~)t}-o fort- eo

~ + vz ~JV1+(a+~) 2 t 2 - (a+~)t}- 0 fort ..... oo.

Since on.( ax+~y+2Re(yz)+ ~~ + vzl 2 > 0 (the intersection of

the planes ~y + vz = 0 and ax + ~y + 2 Re( yz) = 0 is generator

of the cone xy - r z r 2 = 0)'

~ {ax + ~y + 2 Re(yz)+ ~~ +vz[2

} = 0

and this infimum is not assumed for finite x, y and z. D

Remark. From the theorems 2.4 and 2.6 ii; follows that onlf the

infimum of f(x,y,z) = ax + ~y + 2 Re(yz)+ j~y + vz[ 2 is assumed

iff D I 0 V F I 0 \I{ a = 0 A ~ = 0 A (E I 0 V l.t = 0)); then the min­

imum is also the unique stationary point unless a 0 A p= 0 A E >0

(then the minimum value is assumed on the intersection of~ and

the plane ~ + vz = o) or a = 0 A p = 0 A E = 0 A l.t 0 (then f

is zero everywhere onJ( ) • The infimum of f(x,y,z) is not assumed on~ iff

D = 0 A F 0 A (a I 0 V p I 0 V (E = 0 A ~ :j 0)). In this case f

even has no stationary points on f'.

2. 5. The commutator in relation to shear transformations

As in the real case, the reaching of a stationary point of

'I!T ;::;1A T J::mliE' considered as a function of the Euclidean para-

65

Page 67: An eigenvalue algorithm based on norm-reducing transformations

meters of T0

, corresponds to the annihilating of a part of ' ,vm

C(T,e~1 A T£m). Without proof (which is entirely similar to the

proof of theorem 1.7),we state

Theorem 2. 7. Let T ,em be a shear wi t:i Euclidean parameters x, y

and z = u + iv. Let f(x,y,z) be the function in the right-hand

( ) ,, ) I -1 2 side of 2.2.7, i.e. flx,y,z IT ,em A T,emiiE• 1 Let C(T 0 T 0 ) = (c! .). For the elements of C(T,- A T,.J holds

Nm ,vm lJ Nm hlil

ct,e = c~ = c~m = o, if and only if for the Euclidean parameters

(x,y,z) of T,em holds

af af = af ax (x,y,z) ay (x,y,z) au (x,y,z)

= o. av (x,y, z)

Now we restrict ourselves again to unimodular shears T,em•

We shall investigate the properties of C(T,e~1 A Ttm) when

T £m E: 11b,em (A), i.e. for the Euclidean parameters (x,y, z) of T £m

holds f(x,y,z) is minimum on:t( in (x,y,z).

Lemma 2.7. Let T0 be a unimodular shear with Euclidean para-"'m 1 ,

meters (x,y,z). Let C(T0

- A T0

) = (c! .). For the elements of _

1 ,.,m "'m lJ

C(T,em A T,em) holds c£;;- cr:un = c£m 0 if and only if f is sta-

tionary on:!( in (x,y,z).

We first consider a pre-treated matrixA (for which

a me =' 0) and we introduce the new variables t and w:

X = t + W, y t - W •

Since on xy-1 zl 2 = 1 , x > 0, y > 0, :r' is now by ..!.

t = (1+vf+l z[' 2 )2

• With the new variables wand z which para-

metrize :f', we find from (2.2.9) a new expression k(w,z) for

1fT -1fT 1! 2 : ,em ,em E

f(x,y, k(w,z):= (a-ttJ)t+(a-~)w+2 Re(yz)+!!l(t-w)+vz[ 2•

From simple calculations, we find the relations between

the elements of C(T,e;1A T,em) and the partial derivatives of

k(w,z):

66

Page 68: An eigenvalue algorithm based on norm-reducing transformations

o' -c' IPI 2+1 si

2-[ql

2-frl

2 - - k .e.e mm pr-qs pr-qs w

-0 lm =i - (2.5.3) pq-rs ps qr k +ik u V

c' - - -pq-rs qr ps k -ik me u V

where u = Re(z), v = Im(z); p,q,r and s are the Jacobi-parameters

of Tx,m• Since the determinant of the coefficient matrix in (2.5.3) equals i-1 ps-qrf 2 (I pl' 2 +f qf +! rj 2 +I sf2

) = (x+y) /2 =J 0, we find

o 1 - c 1 = c 1 = 0 if and if k = k = k = 0, thus iff .. et mm ..em w u v

f(x,y,z) is stationar,r on~. It is easy to see that (ou- omm) 2 + 4lo,.em[ 2 is an invariant of

C(A) under unitar,r shear similarity transformation of A. Hence, "" if A is a general matrix and A the result of the pre-treating of

A, the quantities o£..e - c~ and ctm corresponding to A are zero

iff the same quantities corresponding to A are zero. . ~ -1 "'

On the other hand we have seen that 11T ,em A T ,em liE is stationar,r

I ("' ) I . -1rv under the restriction det T .em = 1 ~ff liT ..em A T .emlfE is

stationar,r under the restriction jdet(T..em)l 1 (see section 2.2). Hence we find also for a general matrix A: c!n - c 1 = c 1 = 0 iff

""""' mm £m f(x,y,z) is stationar,r ondf'. 0

Remark. Let C(A) = (ci)' T tm E t)l,m and k(w,z) ftTx,:1A Ttm!IE,

where w = x-y. Then it is a simple matter to prove that

c - c = akj ; c = t (ak + i ak) J

)1,)1, mm aw (w,z)=(O,O) )l,m au av (w,z)=(D,O).

As in chapter one, a direct consequence of lemma 2.7 is

Theorem 2.8. Let C(A) = (c .. ). Then inf (ftT "-1

A T ,_ftE)=IIAIIE lJ T E t ...,m """'

J,m J,m

if and only if c££ - cmm = c,em = o.

Finally we. prove that for a real matrix A the Euclidean parameters

of the complex optimal norm-reducing unimodular shears are real.

67

Page 69: An eigenvalue algorithm based on norm-reducing transformations

Theorem 2.9. Let ~tm(A) be the class of complex unimodular op-

timal norm-reducing shears. If A is a real matrix, then the

Euclidean parameters of the row-congruent shears in 1n ,em (A) are

real.

Proof. As we have seen in lemma 1.7 a real optimal

unimodular shear T£m (determined except for row congruency) trans­

forms the matrix A in such a way that if C(T~1 A T£m) (oij), then c£~- c~ = c~ = 0. But then, as we have seen in theorem 2.8,

it is not possible to decrease the Euclidean norm of T~1A T,em by

a complex unimodular shear transformation on the pivot-pair (~,m).

Thus the real unimodular optimal norm-reducing shears are also op­

timal ones of the complex algorithm. D

68

If A is a real matrix, then

( I!T ~1 A T £m HE) = min

T£mEl£m T ,em complex

(liT ~1 A T ,emiiE).

Page 70: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 3

CONVERGENCE TO NORMALITY

3. 0. Introduction

In this chapter we shall prove that the unimodular

optimal norm-reducing shear transformations described in the pre­

ceding chapter, and using a well-chosen pivot-strategy, a sequence

{Ak} of similar- matrices is obtained which converges to normality

in the sense of definition 0.5. Our proof of the convergence theo­

rem is essentially the same as that of Eberlein [4]. Since we make

use of the Euclidean parameters of the norm-reducing shears, our

calculation of Eberlein 1s estimate for the decrease of the Euclid-

ean norm(see theorem 3.2)is considerably more than her own.

3.1. A lower bound for the optimal norm-reduction by shear transformations

In order to prepare the proof of the convergence theorem we would

call to mind some properties of unimodular optimal norm-reducing

shears as indicated in sections 1.5 and 2.5. Ex:cept for row con­

gruency, a shear T n is uniquelY determined. by its Euclidean para-• A;ll

meters x, y and z of which x and y are real and z is complex. In-

stead of (x,y,z) we introduce the real variables t, u, v and w :

X = t + w, y t - w, z u + iv. (3.1.1)

The conditions x > 0, y > 0, XY -[zj 2 >0 which are necessa:cy and

sufficient to characterize a shear (see lemma 2.1) are equivalent 1

to the condition t >(u2 + v2 + w2fl,

The conditions x > 0, y > 0, :xy -l'z[ 2 = 1, which characterize uni­

modular shears are equivalent to

(3.1.2)

We shall consider u, v and w as independent variables, determining,

·except for row congruency, a unimodular shear T,em•

69

Page 71: An eigenvalue algorithm based on norm-reducing transformations

Let 1

k(u,v,w' A) :=HT~ A Thnll~.

The determination of the unimodular shears Thn such that

is equivalent to that of (u , v , w ) for which holds 0 0 0

k(u ,v ,w 'A)= inf k(u,v,w;A). 0 0 0 ( ) u,v,w E R

3 Let P be the point in the (u,v,w)-space with coordinates

u = v·= w = 0. This point P corresponds to the Euclidean para1ne1ceJ:s

x = y = 1, z = 0 of the class Uhn of unitar.y shears on the pivot­

pair (t,m). Let (c .. ) = C(A). we have seen in the remark after :I.J

lemma 2.7

akl aw p

akt + i akl c ,e[ cmm .au P av P

In the proof of the convergence theorem we make use of a gradient

method to approximate the minimum of k(u,v,w;A). In order to sim­

plify the application of this method it is advantageous to have

c hn = o. For unitar.y shears U Jkn holds uzmc(A)U hn = C(UlnA U tm).

Hence the (t,m)-element of C(UlmA Ulrn) can be annihilated by

transformation of A with the shear which diagonalizes the

(,e,m)-restriction of the Hermi tean matrix C(A). Let A 1= UtmA U,em•

Then the u and v components of grad k( u, v, w;A ') in u = v = w = 0

method that is used for the

approximation of the minimum of k(u,v,w;A 1 ) it suffices to vary

along the w-axis. This corresponds to considering only those shears

that are row congruent to a diagonal matrix DJkn = diag(dj), where

d. = 1 for j I£ A j I m and dn d = 1. J N m

Since ItA nE = l!A I nE we easily find

70

Page 72: An eigenvalue algorithm based on norm-reducing transformations

We .now introduce a new variable -r : 1,_

1 +'!; 2

(1--r) ' -1 < ... < 1.

Further we define

Then /"L I EO K, /NI' EO M, L + 2N == -t( c }.e - c~-r).

With this new variable " and these notations we find

__ 2 _ (L+2N) 1i+(K+4M) , 2-(L-2N) , 3 +

1-1:'2 1--t2 (3.1.5)

If K =M= o, each affected non-diagonal element of A1 is equal to

zero, hence o = 0 and mJJn (At) = 1{.£ m"

If L + 2N o, then et.£- c~= 0. Since elm also equals zero, we

conclude from theorem 2.8 ?n.£m(A 1 ) 1t.£m' In both cases norm­

reduction is impossible on the pivot-pair (,e,m).

We now consider the case L + 2N f 0. Then K + 4 M > 0.

It is easy to see that

~ 0 ~ = -2 (1 + 2N) , "-r==O

A gradient approximation of the minimum of b('>) is obtained by

taking as a trial value of 'r:

'T 0

_ d 0 1 1 1

d 2

0 r = -i L + 2N d-. _0 d 2 0 K + 4M

... - 't' '>=

(3.1.6)

(It follows from (3.1.4) that I 't" rEO ) Using this value T we 0 0

obtain frcim (3.1.5)

71

Page 73: An eigenvalue algorithm based on norm-reducing transformations

-21: t(L + 2N)-(L - 2N)-r 2 2K" 4 o ________ ..;..o + _..;;o_

1-'r2 -'];2 (1-1:2)2 0 0 0

2K 4

=i(L+2N)2

[ 1 +t (6N-L)(L,+2N) J + "o K + 4M (K + 4M)2- i(L + 2N)2 (1 _ 1:2)2

0

Since [ (6N-L)(L+2N) [ <(6M+K) (K+2M) = K2+8Kllll+12M2..: (K+4M) 2 and

[L+2N [ ..:K+4M we find

1 ( c ~,e - c~) 2

3 a'+ ~ 1 + 4 ( I a k I I a~,e [2)

Since c 1n = o, we have (c },e - c~) 2 = (c U - cmm) 2 + 4[c fm'2

Moreover, a' + ~ 1 + 4( [a_km[2

+ [a'm,/) < 411A' ~~ = 4[AH~ •

Hence ( _ ) 2 + 4 r [ 2 1 c ,e,e cmm c fm 0 :;, - ___ __;;;____; __ ___;=--

12 ll Al!2 • E

We summarize the results in

Theorem 3.1. For each pivot-pair (i,m) there exists a unimodular

shear T im such that

(3.1.8)

The pivot strategy, which must guarantee that the Euclidean norm

decreases in a sufficient degree for convergence to normality, will

be derived from the lower bound (3.1.8) of this decrease. Therefore

we need

Lemma 3.1. There exists a pivot-pair (i,m) such that

(c 1,g- cmm)2

+ 4[cim[2

:;, n(t-1) IIC(A)II~.

72

(:~.1.9)

Page 74: An eigenvalue algorithm based on norm-reducing transformations

Proof. We have

2:: ( .)2 ifj J

n

n 2(n-1) )::: - 2 c .. c .. .

i=1 ll JJ

But since r, c .. ll

0 (i.e. tr (A*A AA*) = o)' i=1

n n 2: c.

i=1 :.i I: +

i=1

'A1lence for n ?-> 2 n

(c .. - c .. )2 = 2n 2~ ~-l J ,j i=1

c .. c .. o. ll JJ

n ?->4 2:: c~.

i=1 Jl

Fence L: {(c .. - c .. ) 2 + 4lc. ·l 2 f :;:,411CII~. i~j ll JJ lJ

If the pivcts ();,m) are chosen in such a way that

then (3.1.9) holds, as follows from (3.1.10). D

A consequence of theorem 3.1 and lemma 3.1 is

(3.1.10)

Theorem 3.2. Let A be ann x n matrix and C(A) = A*A -AA*. Then for each pivot-pair (t,m) for which (3.1.9) holds there exists

a unimodular shear 'I' tm such that

/!c (A)[[_} l!J (3.1.12)

3. 2. The convergence theorem

We now retvrn to the optimal norm-reducing unimodular shears intro­

duced in the chapters one and two.

Lemma 3.2. If the pivot-pair (t,m) is chosen in accordance with

(3.1:9) and T.em is an element of the class mtm(A) of optimal

norm-reducing unimodular shears constructed by the algorithm de­

scribed in the preceding chapters, then

73

Page 75: An eigenvalue algorithm based on norm-reducing transformations

Proof. The estimate (3.1.12) of the decrease of the Euclidean

nonn by transformation with an approximation of the optimal norm-

reducing shears holds a fortiori for the optimal one. 0

Theorem 3. 3. Let the sequence {~}, starting fro~ A0

= A, be

constructed recursively by

k 1,2, ...•.

where (~k'~) is the pivot-pair according to rule (3.1.9)

and T~k'~ is an element of the class ~tk'~(~_1 ) of optimal

norm-reducing unimodular shears constructed by the algorithm de­

scribed in the preceding chapters, Then the sequence {~}converges

to normality.

Proof. {lf~l!~} i~ a monotonically decreasing sequence of numbers,

bounded below by E ~~[ 2 (see theorem 0,8). Therefore, i=1

3n(n-1)

we have

I!C(~_1 )11~

IIAII~

Thus also !IC(~)IIE - 0 fork- oo, From the corollary of theorem

0.13 follows that {Ak} converges to normality. 0

Corolla;y. The sequence Df departures of {~(~)},

corresponding to the sequence {Ak} of theorem 3.3, converges to

zero.

74

Page 76: An eigenvalue algorithm based on norm-reducing transformations

Proof. Consequence of theorem 0.12. 0

Finally we prove an analogous theorem for the real norm-reducing

algorithm described in chapter one.

Theorem 3.4. Let A = A0

be a real matrix and let {~} be the

sequence of real matrices constructed according to the recursive

relation

,k=1,2, •••• ~ := T .£:~~ Ak-1 T ,ek'~ where the pivot-pair (,ek,mk) is

and T 0 is an element of the "'k'~

selected according to rule ( 3.1 .11)

class Zt,ek'~(Ak_1 ) of real opti-

mal norm-reducing unimodular shear computed with the algorithm

described in chapter one. Then {~} converges to normality.

Proof. In the case of a real matrix the approximate optimal

norm-reducing shear transformation of lemma 3.2 produces a real

matrix. From the corollary to theorem 2.9 we conclude that lemma

3.2 still holds for this real transformation. Hence the proof of

theorem 3.3 can be copied. 0

75

Page 77: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 4

JACOBI-LIKE METHODS FOR ALMOST DIAGONALIZATION OF ALMOST NORMAL MATRICES

4. 0. I ntroduction

During the numerical realization of the Jacobi-process for

Eermitean matrices the Hermitean character is usually preserved

by storing and computing only the elements in the upper trangle

and the corresponding elements in the lower triangle to

b.e thGir complex conjugates.

It is generally not pos~dble to preserve in such a way the normal

character of a normal matrix: al~eady the floating-point repre­

senti1hon of a normal matrix A differs not only from A, but gen­

erally it will even not be normal. Therefore, algorithms for ex­

actly normal matrices are mainly of theoretical interest. Their

scope of applicability has to be extended to a larger class,

namely to the class of almost normal matrices.

In section 4.1. we shall describe an algorithm by which, using

unitary shears, a complex almost normal matrix is transformed in­

to almost diagonal form. The ultimately obtainable departure of

cliagonal form ·depends on the departure of normality which is in­

variant under unitary transformation.

In section 4.2. we shall formulate a real variant of this process.

In this case our aim is to reach an almost bl?ck-diagonal matrix,

being a perturbed Murnaghan canonical form. This proves to be

possible, unless the matrix has, with a pair of complex conjugate

eigenvalues, yet other eigenvalues with the same real part.

In the sections 4.3 and 4.4 we shall indicate which element in

the class ~~m(A) of real (respectively complex) optimal norm­

reducing shears may be preferably used in a norm-reducing trans­

formation on the pivot-pair (~,m). As we have seen in the chapters

one and two, the optimal norm-reducing shears with pivot-pair(t,m)

Page 78: An eigenvalue algorithm based on norm-reducing transformations

are determined up to row congruency, i.e. if T0 E ln" (A) , ,vill ,vill

E m,1

(A), where U o is any ...,m "'m

,m) • As we have seen in chapter

gence to normality of the sequence {~} is

T n E lnv (Ak 1) that ~k'~ Nk'~ -

element

transformation

k 1. 2 ••••

shear with

the oonver­

of the

is used in the

Since usually norm-reduction is only the first of an algo-

rithm for the computation of the eigenvalues

able to make use of the degree(s) of freedom

~ g .~ (Ak_1 ) in order to promote the

of A, it is reason-

within the class

of the k k .

matr1ces in the sequence {~}. Therefore,

use the element from ~n (A 1

) that '"k'~ -K-

it seems appropriate to

minimizes the departure

of diagonal form of the transformed matrix ~·

4. l. Almost diagonalization of a complex almost normal matrix ix

We introduce the following notations

H :== t(A+A*), ,D .­

A := i(A-A*), K .-

diag(Re(a .. ) ) , JJ

,2, ••• ,n, F := H-D. (4.1.1)

diag(i Im(a .. )), j JJ

1,2, ••• ,n, G:=Z-K.(4.1.2)

If A is normal and the Hermitean part H of A is diagonal, i.e.

H = D, then DZ = ZD. Consequently, = 0 if d .. f d ..• Hence J.J. JJ

a normal matrix of which the Hermi tean part has diagonal form, is

a direct sum of diagonal blocks. Each diagonal block is the sum

of a real multiple of the unit matrix and a skew-Hermitean matrix.

In the sequel we shall

almost diagonal Hermitean parts. Of these matrices the diagonal

elements can be grouped into elements within a cluster

baNing almost equai real parts. In order to describe this phenom­

enon we introduce the notion of a ,;-partition of a complex matrix.

77

Page 79: An eigenvalue algorithm based on norm-reducing transformations

Let us permute the rows and columns of A so that the real parts of

the diagonal elements are non-increasing. Without loss of generality

we may assume this was true originally, i.e. if 1 "' j < k < n, then

Re(a .. ) ;;. Re(akk.). JJ

Let

A

A 11

A 21

A 12

be a partition of A with square diagonal blocks. Let~ ;;. 0.-

This partition is called a ~-partition of A lithe real parts of suc­

cessive diagonal elements to the same diagonal block dif-

fer not more than~, whereas the differences of the real of

diagonal elements occurring in different diagonal blocks are greater

than this value,.. Or, eg_uivale~tly: Re(a .. - a. 1

. 1

) "'~if and ll l+ ,l+

only if a .. and a. 1

. 1 belong to the same block. ll l+ ,l+

Let the dimensions of the diagonal blocks be n, n , ••• , n. Then 1 2 K k

L: j=1

n.=n. The real parts of the diagonal elements belonging to the J

j-th diagonal block do not di'ffer more than (n .-1 h. Let for this J

~-partition the g_uanti ty 1: * be defined by

: = min (Re (a .. - a . . ) ) , . . ll JJ l<J

and a .. belong to different diagonal JJ 2 2 2

Since (-r *) "'ma.:x[ Re (a . . - a .. ) ] "' 2f!AIIE' . . ll JJ l' J

where

is only one block, then we take 1:* = +=.

78

blocks. Evidently ],_

"*"'22 IIAIIE. If there

Page 80: An eigenvalue algorithm based on norm-reducing transformations

Let U be a block diagonal unitar.y matrix with the same

A, thus

u

Let

u 11

0

0

A' 11

A' 21

A1 := U*AU =

Then

A! . = Ut. A. . U .. lJ ll lJ JJ

0

u 22

0

A' 12

AI 22

i,j

0

0

A' 1k I k

1,2, .•• ,k.

tion as

(4.1.6)

Analogously to (4.1.1) and (4.1.2) we denote qy H1 , D', F 1 , Z1 , K'

and G' the corresponding matrices derived from A1 • These matrices

corresponding to A and to A' are partitioned as A.

Using these notations we start to estimate the non-diagonal blocks

of z.

Lemma 4.1. according to (4.1.3) with some

given non-negative value of ~· D, F and Z are defined according to

(4.1.1) and (4.1.2), ~*according to (4.1.4). Then for the elements

zij of Z not belonging to a diagonal block holds l

izij'"'" 2""2 [ll(A)+ 2(l!ZliE+ ll(A)) (!IF!IE+ ll(A))/ -.*] , (4.1.8)

79

Page 81: An eigenvalue algorithm based on norm-reducing transformations

where 6(A) is the departure of normality of A.

11breover,

2: !IZ n IIE2

tfm --vm

Proof. In with theorem 0.10 we can write A = N +

where the normal matrix N has the same eigenvalues as A and

!! PilE l:!. (A). Then

D + F = N + N*)+ t(P + P*), Z = t(N - N*)+ t(P - P*).

Let Q : P + P*), R := t(P- P*),

Since N is a normal matrix, the Hermitean and skew-Hermitean parts

D -+ F - Q and Z - R respectively,. commute:

(D + F - Q) (z - R) = (z - R) (D + F - Q) •

Hence

DZ - ZD = DR - RD + ( Z - R)(F - Q) - (F - Q)( Z - R). (4.1.10)

From this matrix identity we find the following identities for the

(i,j)-elements: n

(d .. -d .. )z. ·"" 2:: Yz.k- r.k)(fk.- qk.)-(f.k- q.k)(z. .- rk.)} + ll JJ lJ k=1~ l l J J l l kJ J

(4.1.11)

With the triangle inequality and that of Cauchy-Schwarz we derive

from (4.1.11) the following inequality:

2t 2t 2t 2t I d .. -d .. jjz .. j <~> {(Biz.kl ) +(Bjr.kl ) }{(~Irk./ ) +(~jqk. I ) } + ll JJ lJ k l k l k J k. J

2t 2t 2t 2t + { ( Z If. k I ) +( Z:: I q. k I ) }{ ( E I zk. I ) +( E Irk. I ) }+Id .. -d .. 11 r .. 1. k l k l k J k J ll JJ lJ

Since F and Q are Hermitean, and Z and Rare skew-Hermitean we can

majorize the right-hand part in the way : 1 ' 1

jdii-djjll zij j..;2'"'2/dii-djj I I!RIIE+2. 2'"'2(11 ZI!E+I!RIIE) (IIFIIE+[Q~).

Since ITQ~ = nt(P + P*)I!E..;;fiP!IE = l:!.(A) and similarly [RilE<~> l:!.(A},

80

Page 82: An eigenvalue algorithm based on norm-reducing transformations

we have -t .d..

ldii-djj J lzij 1,..;2 /dii-djjl 6(A)+ 22 (llZJIE+ 6(A))(IIFIIE+ 6(A)).

If (i,j) is the index-pair of an element not belonging to a diago­

nal block, Id .. -d. ·I > 1:*, hence (n .• 1.8) is obtained. ll JJ ' In order to prove (4.1.9) we start from (4.1.11):

z .. = r .. + [(z- R)(F - Q)-(F - Q)(Z - R)] .. /(d .. -d .. ). lJ lJ lJ ll JJ

Hence, if z .. does not belong to a diagonal block: lJ

lzij I ,..; lrij I+ l[(z-R)(F-Q)-(F-Q)(Z-R) ljl/-c*· (4.1.12)

We now take squares, and in the left-hand side.we sum over the non­

diagonal parts of A, whereas in the right-hand side we sum over all

i,j. Then we obtain

2: UZn ITE2 ,.;11jRI+i(Z-R)(F -Q~ /r:* + I(F-Q)(z -R)I/r*ll~ tiro "'m

,.;3liRIT~ + 6(-r*)- 2 1lZ-RI!~ ffF-Qll~

<e;;3t.2(A) + 6{'r*)-2 (nznE+ t.(A)J (\!FilE+ t.(A)) 2•

(Here I RI denotes the matrix with elements I r .. 1 etc.). D lJ

The qualitative interpretation of (4.1.9) is very simple. If !!FifE

and 6(A) are small relative~ to 1:* and t.(A) is small relatively to

nAHE' then t~m nz~mll~ is small relative~ to IIAII~ (note that

!lznE ""!!AI! E). Hence an almost normal matrix with a Hermitea~ part of almost dia­

gonal form is almost a direct sum of diagonal blocks. Each diagonal

block is the sum of a Hermitean matrix which is an almost multiple

of the unit matrix and a skew-Hermitean matrix.

We shall now prove a lemma which is useful in dealing with diagonal

blocks of the above type. To such a block we want to apply a uni­

tary _transformation to diagonalize the skew-Hermitean part of the

block. The lemma asserts that after this transformation the

Hermitean part of the block is still an almost multiple of the

unit matrix.

81

Page 83: An eigenvalue algorithm based on norm-reducing transformations

Lemma 4.2. Let H be a Hermitean matrix and

t : max ( j h .. - h .. !) • . . n JJ ~,J

(4.1.13)

For each unitary matrix U the departures of u..l."•"'v·''"" ... form of H and

U*HU satisfy the inequality

(4.1.14)

If a:= i(max (h .. ) + min (h .. )), then I o:- h. ·I os;; it for j JJ j JJ JJ

each j. Let F be defined by F : == H - diag(h .. ) • Since JJ

H =a I·+ diag(h .. - a) + F, for each unitary U we find JJ *

U*rlli o:I +U* [diag(h .. -a) +F]U. Hence we have for S(UHU) JJ

S2 (U*HU) os;; 11U* [ diag(hjj- o:)+F]U!I~ os;; z n t 2 + !IF!!~ • D

With the preceding lemmas we are able to estimate the departure of

diagonal form s(A') of the matrix A'= U*AU (described in (4.1.6))

in terms of l'l(A), the non·-d~agonal part of the Hermitean pal.'t of

the non-diagonal parts of the skew-Hermi tean parts of the diagonal

blocks of A1 , and the quantities~ and~* belonging to the parti-

tion.

Theorem 4.1. Let A be a complex matrix for which holds

Re(a .. ;;;. Re(a .. ) if 1 os;; i < j ...:; n. Let A be partitioned acc;rding n JJ

to (4.1.3) for some given non-negative value of~, i.e. and

a. 1

. 1

belong to the same diagonal block iff Re(a .. -a. 1

.+1

) os;;-;. J.+ ,~+ J.J. J.+ ,J. Let U be a unitary block diagonal matrix having the same block

structure as A. Let A 1 := U*AU. The matrices F, G and Z are defined

according to (4.1.1) and (4.1.2). Let~* be defined according to

( 4 • 1. 4) • Then

s 2 (A')< IIF!I~+ 3l'l2 (A)+ 6h·*)~ (!!ZfiE+ l'l(A) )2 (!!FilE+ n(A) )2 +

k + ~ {I!G!.!!~ +in .(n .- 1 ) 2 ~2 }

j=1 JJ J J

82

Page 84: An eigenvalue algorithm based on norm-reducing transformations

where n. is the dimension of the j-th diagonal block A .. of A. J JJ

Proof. We start by rellll;l,I'.ICI.r.lg that S2 (A) "' S2 (H)+ (z). Hence k k

(A 1 )= 4, ff-(A!.)+ I!A! .!IE2 = ~ {S2 (H!.)+S2 (Z!.)}+ ~ !IA!.IIE2 •

j=1 JJ ::LJ j=1 JJ JJ ifj ::LJ

Since H'. . = F! . + D t . and Z! . = K~ . + G~ . , JJ JJ JJ JJ JJ JJ

k c A,) = ~ c !IF ! . nE2 + n G! . rrE2 ) + 2:; RA! .11 E~ •

j=1 JJ JJ ifj J.J

We now apply lemma 4.2 to the Hermitean part H! J

parts of the diagonal elements of A .. differ JJ

whence I!F!.I! 2E.,.;}-n.(n. -1?-r2 + [[F .. I!E2 • JJ J J JJ

Thus k ( 2 2) k .J..._ )2 2 E !IF! .HE+ I!G! .liE ..;;; ~ {411.(n.-1 -r + l!F ..

JJ JJ j=1 J J JJ

(4.1.16)

of A! .• The real JJ

more than (n.-1)-r, J

Since Ai_j = uri Aij Ujj' we have I!AJ}E

of lemma 4.1 to estimate ~ !lA .. IIE2 •

IIA .. lfE. We make now use ::LJ

J "J ' irj ~

l!A .. IIE2

= ~ (!IF .. I!E2 + I(Z •. IIE2

) ::LJ ifj 1J 1J

11~+ ~L12 (A)+ 6(-r*)-2 (I!ZIIE+Cl(A) )2 (I!FIIE + ll(A) )2

(4.1.18) With (4.1.17) and (4.1 .18) we find from (4.1.16)

+ ~· {l!G!.II~ +in.(n.-1)2-l}. D j=1 JJ J J

83

Page 85: An eigenvalue algorithm based on norm-reducing transformations

We conclude from (4.1.15) that S(A') will be small

HAnE if the following requirements are fulfilled:

1 • 11(A) << !lA!~

2. IIF1!E << llA!lE

3· 't" << IIA!IE

4· !IF liE + 11(A) << -.*

5. k

l!G'. -llE2 << I!AIIE 2. L!

j=1 JJ

to

If the conditions 1 and 2 are fulfilled then obviously there exists

a ,;-partition so that the requirements 3 and 4 are statisfied. This

'_ist of requirements for the smallness of s(A 1 ) /!lA liE may be used as

a lode-star for a safe (but in many cases unnecessarily pessimistic)

choice of the value of -r which deten"tlines the -r-parti tion of A. The

value of -r is to be chosen so that the terms

6(T*)- 2 (!lZIIIE + 6(A)) 2 (!!F!!E + ll(A)) 2 and t t 2 ~ n.(n.-1) 2 are both

j=1 J J small relatively to liAl!E. In the following corollary to theorem 4.1

we mention a value of -r which fulfills this requirement.

where F is defined according to (4.1.1). Using this value of 't" in

the of A and u, and using the same notations as in

theorem 4.1, we find that the departure of diagonalform of A'= U*AU

satisfies the inequality

Proof. We derive from (4.1.19)

·h*)-2< 1:-2 = (I!FIIE + ll(A))-1 KAn;1

• (4.1.21)

From (4.1 .15) and (4.1.21) we conclude that (4.1.20) holds. D

Page 86: An eigenvalue algorithm based on norm-reducing transformations

In order to

write this

a qualitative interpretation of (4.1.20) we re­

in the following manner:

k k 2.: n.(n.-1)2 +-1-

J J IIAII~ E I!G~. lli .

j=1 JJ (4.1.22)

~(A) << IIA!IE' IIF''IE << IIAII'IE and U is such that Hence, if k 2 Z !IG~ .liE

j=1 JJ << IIA'I/~, then S(A ')<< IIAIIE· This means that if A is al-

most normal, the Hermitean part of A almost diagonal, and ~ chosen

to (4.1.19), then A1 ·almost iff the skew-

Hermitean parts of the diagonal blocks of AI are almost diagonal.

We would remark that the estimates (4.1.20) and (4.1.22) for S(A')

are very crude, and that for the reasons.

Within a diagonal block of our Re(aii- ai+i,i+1 ) ~ ~' whereas I Re( a .. - a .. ) I ~ 1:* > 1: if and a .. belong to different

ll JJ JJ u~'~g,;r~'~ blocks of the partition. For the majority of (almost nor-

mal) matrices (namely, those whose have well-separated

real there exists a 1: of the order of !!FilE + ~(A), for which

-c* (of the corresponding 1:-parti tion) is of the order of I!A!fE •

Then we find from (4.1.15) that (A 1 ) is of the order of k 2

ftFjj~ +62(A) + Z IIG~ .!lE • For these matrices the value of ,. given j ==1 J J

in (4.1.19) is a severe over-estimate of the

smallest value of 6( 1:*)-2 (IIZIIE+ !.I( A) )2 (!lFII

value of 1: gjving the

( ) ) 2 1 2 k ( )·2 1". A + 4 -r 2.: n. n.-1 j=1 J J

and a severe under-estimate of the 1:* corresponding to the latter·

value of 1:. Hence for these matrices taking both for 1: and 1:* the

value given by (4.1.19) means severely overestimating of the

hand side of (4.1.15).

The estimate obtained for S(A') suggest an algorithm to transform

an arbitrary matrix into almost diagonal form.

85

Page 87: An eigenvalue algorithm based on norm-reducing transformations

Algorithm. Let A(o) be an arbitrar,y complex matrix of the order n.

This matrix has been transformed into almost diagonal form in the

following way.

(i)

(ii)

We apply the norm-reducing process of theorem 3.3 to the

matrix A(o). Thus the matrix A(o) is transformed into anal­

most normal matrix A(1 ), i.e. li(A(1 )).,.e: , where e: is some 1 1

prescribed positive number.

B,y applying the classical Jacobi-process we transform A( 1 )

unitarily into a matrix A(2 ) of which the Hermitean part is

almost diagonal, i.e. )+A(2)))E;;e: , where e: is some 2 2

prescribed positive number.

(iii)We permute the rows and columns of A( 2 ) in such a way that of

the resulting matrix A(3) the real parts of the diagonal ele­

ments are arranged in non-increas~ order.

(iv) We choose such a value of ~ that for the corresponding~­

partition the quantity

f(~)

is m1n1mum. This can be done by tr,ying the values

0 Re(a(3)- a(3 )) Re(a(3 )- a(3 )) ••• Re(a( 3 ) - a(3 )). ' 11 22' 22 33' ' e--1,n-1 nn

With the value of -r 0btained in this way, A( 3 ) is .. -partitioned.

A(3) A(3) A(3) 11 1 2 1 k

86

Page 88: An eigenvalue algorithm based on norm-reducing transformations

where 1 .;;; k ,..; n.

(v) We apply the classical Jacobi-process to the diagonal blocks

of A( 3 ) (i.e. A( 3 ) is transformed with a block dia­

gonal matrix u< 4 ) , having the same block structure as A ( 3 ) )

in such a way t~t the skew-Hermitean parts of the diagonal

blocks A~~), A~~), ••• ,A(~ of the transformed matrix

A( 4 )::: u(4 )*A( 3 ) U(4)are almost diagonal, i.e.

~ s2(t(A(4 )- A(4 )*)),..; s 2, where e: is some

·=1 3 3 ~ositive number.

The question whether and in what sense A( 4) is an almost diagonal

matrix is answered by the following theorem,

Theorem 4.2. For each e: > 0, we are able to prescribe in the al-

gorithm sketched above, such values of e: , e: , 't' and e: that

s(14))o;;;e:. () () 1 2 3 6{11 t(A 2 - A 2 *)11 + e: }2 (e + e: T !

Take 1: = [ E 1 1 2 J • ! n(n-1 )2

. * With this value of 1: we have, s~nce 1: > 1:,

6(1:*)~{!1t(A(2)_ A(2)*)!fE + e:1 }2 (e1 + e)2 =

= (~2 n)t (n-1)(8 + e: ){!lt(A(2 )- A( 2 )i1E + e L 1 2 1 '

k 2 2 For the term 2:: n.(n.- 1) 1: in (4.1.15) we find

j=1 J J

k t 2:: n. (n .-1 )2"2

,..; i n(n-1 )21:

2 = j=1 J J

• (~2 (n-1)(e + e: ){flt(A( 2)- A( 2)*)11E + e } • 1 2 1

Hence we find from (4.1.15) (since ITA(2 )11E= !!A( 1)llE"' UA(o)IIE)

ff(A(4 )).;;; 3e 2 +e:2 +e:3 +2(~2 n)t(n-1)(e +e ){IIA( 2 )11E +e:} 12 3 12 . 1

(4.1.23)

Page 89: An eigenvalue algorithm based on norm-reducing transformations

By taking appropriate values of e , e and e it is obviously pos-

sible to obtain S2 (A (4 \..; e2

1 0

2 3

Remark. From theorem 0.13 follows a criterion by which effective-

ly can be tested whether ~(A) ~ e • 1

n3 - n .1. .1. For if ( -.:r2) 4

11 c(A)IIE 2 ~ e1

, then a fortiori ~(A)~ e • 1

We shall now demonstrate that the ~-partition of an almost normal

matrix with almost diagonal Hermitean part corresponds to a par­

tition of the eigenvalues with regard to their real parts. To that

end we need

Theorem 4.3. Let i(A +A*) = D + F, where D = diag(Re(a .. )). JJ

Suppose Re(a .. )?Re(a .. ) if 1 ~ i < j ~ n. If 11 + iv, 1.12

+ iv , ••• , ~~ JJ 1 1 2

" + iv (". ? 11 • 1

) are the eigenvalues of A then ""n n ,...~ ~+ ·

n ]2 ( 2 E [Re(a .. ) - ll· ~ IIFIIE +MA)) • j=1 JJ J

Proof. As is stated in theorem 0.11, the matrix A is the sum of

a normal matrix A having the same eigenvalues as A, and a matrix P,

so that \I PilE = 1~A - N11E = ~(A)

For the Hermi tean part i(A + A*) of A we have

i(A + A*),= D + F = i (N + N*) + i(P + P*).

Hence D - i(N + N*) = i(P + P*) -F. (4.1.24)

Since, according to theorem 0.4, the real parts of the eigenvalues

of A are the eigenvalues of the Hermitean par~ + N*) of the

normal matrix N, we find from (4.1.24) with the Wielandt-Hoffman­

theorem

n :1.

( E [Re(ajj)- llj]2

)2 ~ lli(P + P*)- FIIE~I!PIIE+ !!FilE= IIFffE+~(A). 0

j=1

We conclude from theorem 4.3 that the real parts of the diagonal

elements of an almost normal matrix A, with almost diagonal

Hermitean part i(A +A*), may be considered as approximations of

the real parts of the eigenvalues of A.

88

Page 90: An eigenvalue algorithm based on norm-reducing transformations

We first consider matrices having eigenvalues with well-separated

real parts ll ~ ll ~ ll ~ •• ~ lln• Again we assume that 1 2 3

Re( a .. ) ~Re( a .. ) if i < j. n JJ

Theorem 4·4• Let A be a complex matrix with eigenvalues

~=t..tj+ivj

if i < j.

( j = 1 , 2, ••• , n). Let Re (a .. ) ~ Re (a .. ) and ll.; ~ ll . l.l. JJ ~ J

Let d : = min ( lt..ti- ll j l') ll . fll '

J. J

e :=~(A) + s(t(A +A*)).

(4.1.25)

(4.1.26)

If e <id then there exists a partition of A, which is a ~-par­

tition with~= 2e and ~* > d- 2e.

Proof. According to theorem 4.3, for the real part of the diag-

onal elements of A holds

n ( E [Re (a .. )- ll. J2

J J J (4.1.27)

We now consider in the complex plane the vertical strips

I£ = { z ; I Re ( z)- ll i I .,;;;; e:} , i = 1 , 2 , ••• , n. ( 4.1 • 28)

If f.!£ -f llm' then since e~, the strips I ,e and Im are disjoint. As a

consequence of (4.1.27) the diagonal elements of A are elements of

n U I,.e and the number of diagonal elements belonging to the strip

i=1

I , equals the multiplicity of ll 0 in the sequence f.! , ll , ••• , f.l • "' "' 1 2 n

Since e <id, the (horizontal)distance of the diagonal elements of

A belonging to different strips is greater than d - 2e: > 2e:. The

(horizontal) distance of the diagonal elements belonging to the

same strip is not greater than 2e. Hence for the ~-partition of A

with 1: = 2e holds ~* > d - 2e. [J

We can derive an analogous result for matrices with eigenvalues

clustered in vertical strips of the complex (eigenvalue) plane.

Page 91: An eigenvalue algorithm based on norm-reducing transformations

,.

Theorem 4.5. Let A be a complex matrix with eigenvalues A.=~.+iv. J J J

(j 1 ,2, ••• ,n). Let Re( a .. ) ;,Re( a .. ) and ~· :;;, ~· if i < j. Let ll J J J. J

these eigenvalues cluster around k distinct vertical lines z = mh

(t = 1,2,,,.,k), and let o be the ;vidth of these clusters, i.e.

Let d ·- min [m.- m"l' 1<j<t<k J h

e := t:,(A) + S(t(A +A*)).

If c: + o < then there exists a of A which is a ,_

with 1: = 2 ( e + o) and '* > d - 2 ( c; + a).

From theorem 4.3, (4.1.31) and (4.1. ) we deduce that

the diagonal elements of A are elements of the vertical strips

I..r;={z; (z)-m..el"'e:+o}, .R=1,2, ••• ,k. Ifd>2(c:+o),

then the strips are disjoint and the number of elements

of A in the strip equals the number of eigenvalues of A in this

Consequently since b + e < the (horizontal) distance of the

diagonal elements of A belonging to different strips is greater

than d - 2(6 + e)> 2(o + e) whereas the (horizontal) distance of

the diagonal elements belonging to the same is not

than 2(o +e). Hence for the ,-partition of A with,= 2(o + e:) holds '*>d.- 2(o + e). o

4. 2. Almost block diagonalization of a real almost normal matrix

An~logously to the preceding section we now investigate a real A

of which the diagonal elements are assumed to be non-increasing.

Let this matrix be 1"-partitioned with some value of ' : aii and

ai+1 ,i+1 belong to the same diagonal block iff - ai+1 ,i+1 ..,. T.

We assume that we have in this partition of A k diagonal blocks k

Thus L: n. i=1 l

with dimensions n1

, n.

90

Page 92: An eigenvalue algorithm based on norm-reducing transformations

Let the indices of the diagonal elemenvs of the blocks run

from p. up to and including q .• If ~ ~

:==tea +a pi,pi qi,qi) (4.2.1)

then for each diagonal element a££ in the block holds

(4.2.2)

We now consider the following decomposition of A

A= M+ G + L.

In (4.2.3) M is block diagonal with diagonal blocks

(I~ the n .. x n. unit matrix) and G is block .L ~ ~

blocks

i = 1,2, ••• ,k. Gii = !(Aii - AiiT),

If for each i (1 .;;; i .;;; k) holds n . .;;; 2, then M + G is a mur.t'll:~gr:~an l

canonical form. If, moreover, lfLIIE is small relatively to llAUE ,

then A may be said to have an almost Murnaghan canonical form •

In order to estimate ITLIIE we investigate the c-parti tioned matrix.

A=

We would recall that

and a. . belong to the l~

'* = min faii- ajjl' blocks of the

of lemma 4.1 •

a a ,;;:: ~ whenever a ii- i+1,i+1-' i+1,i+1

same diagonal block, and that by definition

where and a .. belong to different diagonal JJ

Without proof we state the real analoguE

Lemma 4.3 •.

(4.2.5) with

Let the real matrix A be '1"-partitioned according to

some value of "· Then

91

Page 93: An eigenvalue algorithm based on norm-reducing transformations

i

k 2:

where

With lemrr.a 4.3 we are able to prove

Let the real matrix A of order n have uvl<-.J..acoJ.''""""J.."-4";

elements: a .. 3 a .. if 1 ,..; i < j ..:; n. Let this matrix be ll JJ

• If A is decomposed according to (4.2.3), then

k + t E

i==1

where H = t(A +AT), Z == t(A- AT) and n. (i 1,2, ••• ,k) is the J.

dimension of the diagonal block A .. of A. ll

We start by remarking that

k [[L = s 2(H) + 1:: fiZ .. [~ + 2:: 2:: (a if" ai)2 •

if'j lJ i==1

According to (4.2.2) (a 1;;- ai) 2 ..:; - 1 ) 2

• Hence the con-

tribution of the diagonal elements of 1 to 1~1~ is not greater .!';

than i=1

n . ( n. - 1 ) 2 • Thus J. J.

k nL1f~,..;s 2(H)+ L: IIZ.I[2

E+!-r2

2:: if'j J.j i=1

With lemma 4.3 we find

2': llz .. fl ~ ,..; 3 t? (A) + 6 ( 't*)- 2 (nz [~ + l'l( A) ) 2 ( S (H) + 6( A)) 2 • ( 4. 2. 9 )

if'j lJ

With '(4.2.8) and (4.2.9) we obtain (4.2.7). D

We conclude from (4.2. 7) that ffLI~ will be small relatively to

nAITE if the following requirements are fulfilled :

92

Page 94: An eigenvalue algorithm based on norm-reducing transformations

1. /1, (A)<< 11 A !lE

2. S(H) << \1 All., ]<,

3. T << 1\Al\E

4· S(H) + n(A) << -r*.

If the condition 1 and 2 are fulfilled, then obvious~y there exists

such a 't"-partition that the requirements 3 and 4 are satisfied. A

suitable value of T is mentioned in the following 1 1

Corollary. Let -. :=- (s(H) +n (A))2 ITAII; • (4.2.10)

Using this value of ,; in the -r -partition of A, we find that the

Euclidean norm of L =- A - ~1 - G satisfies the inequality

ITLIIE EO ff (H)+ 3/1,2

(A)+ 611AII; {IIZIII.J+ n(A)f (s(M)+ n(A)) +

k 2 +tllA.liE{S(H)+n(A)} ~ n.(n.-1).

j=1 J J

Proof. From (4.2.10) we derive

(4.2.11)

C..*)-2 EO T- 2 = {s(M) +/I, (A) r 1 nAr~ t. (4.2.12)

From (4.2.7) and (4.2.12) we conclude that (4,2.11) holds. o

From (4.2.11) ,we conclude that if we take for 1: the value described

in (4.2.10), the perturbation Lis small relatively to A if

l1(A)<< IIAIIE (i.e. A almost normal) and S(H)<< IIAIIE (i.e. the sym­

metric part of A almost diagonal).

Analogously to the complex case, described in the preceding section,

the estimate of I!LIIE given in theorem 4.6 suggests an algorithm by

which an arbitrary real matrix may be transformed into almost block

diagonal form.

In the Erst stage of this algorithm the origir.al matrix is trans­

formed into an almost normal matrix by means of the norm-reducing

method. Finally, the almost normal matrix is transformed by Jacobi

iterations, in order to diagonalize the symmetri.c part of the al­

most normal matrix, Let the resulting matrix be

Page 95: An eigenvalue algorithm based on norm-reducing transformations

A ( 1 ) = M( 1 ) + G ( 1 ) + 1 ( 1 ) • ( 4. 2.13)

In tcis formula we consider 1(1 ) as a perturbation which can be

estimated b,y means of (4.2.1). M( 1 )+ G(1 ) is a block diagonal nor­

mal matrix, each diagonal block M\ 1) + G ~ 1) consisting of a multi­

ple M\1.) = ex. I of the n. x n. ~it ma~~ix and a skew symmetric .JJ J n. J J .

matrix G~ 1). J JJ

If a diagonal block has dimension~ 2, its eigenvalues can be

readily computed. If all diagonal blocks of A(1 ) have dimensions

~ 2_, M( 1 ) + G(1 ) is a so-called Murnaghan canonical form.

If fo; some diagonal block A~ 1.) n. > 2 and the skew-s;ymmetric part ( ) JJ J .

G. 1. of that block is not almost zero, the situation is less favor­JJ

able. It is obvious that all real parts of the eigenvalues of the

blocks M~ t_) + G ~ t_) are ex. ( j = 1 , 2, ••• , k). From oonsidera tions sim-JJ JJ J

ilar to those which led to theorem 4. 5 it follows that this sit-

uation can occur only if the original matrix has three or more

eigenvalues with almost the same real part and among them at least

one complex conjugate pair. It is a well-known fact that a:nY skew­

symmetric matrix can be orthogonally transformed into its Murnaghan

canonical form (see theorem 0.3). However, for d.imension > 2 at

this moment no general algorithm is available to achieve this

transformation with the aid of a sequence of Jacobi rotations.

In this case we advise a transfer to complex arithmetic and use of

the algorithm for complex matrices described in the preceding sec­

tion.

4. 3. The real diagonalizing representative of7Jkrn(Al

In section 4.2 we considered the Jacobi process applied to the sym­

metric part of an almost no:r.mal matrix. In fact, this process ge­

nerates a sequence of real matrices which converges to Murnaghan 1s

canonical form, unless with a pair of complex conjugate eigen­

values A.. + iu. the matrix A has still other eigenvalues with tbe J - ... J

94

Page 96: An eigenvalue algorithm based on norm-reducing transformations

same real part A·• B~t the Jaoobi process applied to the ~JW~etric J

part H .. + of a real matrix A may be considered from an-

other point of view. As a matter of fact, a Jacobi transformation

with Q.£ minimizes not only S(Q~ H Q.£ ) but also the departure of m T m a

diagonal form of Q£mA Q~m' as is stated in

Lemma 4. 4. Let A be a real matrix and Q£m an orthogonal shear.

Then Qtm minimizes S(Q~mA Q£m) if and only if Qtm minimizes

s( QtmH Qtm) •

Proof. It is easy to d.emonstrate that

s2(QT HQ ) + s2(Q:m lm .t:m "'

T 82 (QT H Q ) + g2 (A-

2A ) •

.t:m lm 0

Remark. It is well known that Qtm minimizes S(Q~~ Qtm) if and

only if the (.t:,m)-restriction of Q~mH Qtm is diagonal.

Lemma 4.4, in conjunction with theorem 4.6, suggests which element

of the class ?n1m(A) of unimodular optimal norm-reducing shears is

suitable for a numerical norm-reducing process, if we want to com­

bine diagonalization and norm-reduction. It seems reaso:r..able to

select from~. (A) the element Tn that diagonalizes the (t,m) -,vm /Jffi

restriction of the symmetric part of A. For this T tm lntm(A)holds,

as we have seen in lemma 4·4

S(T.-1A T. ) = min S(P"m-

1A P.m) •

..vm "'m P E m, (A) "" "" tm tm

This unimodular norm-reducing shear Ttm will be called the

diagonalizing representative of nttm(A).

The diagonalizing representative Ttm of nz.t:m(A) can be obtained

from the upper-triangular representative Btm of m.,.t:m(A) by multi­

plying this shear with an appropriate orthogonal shear Qtm :

95

Page 97: An eigenvalue algorithm based on norm-reducing transformations

Let A' := T;:~1 A T.em· Then Qtm is determined from the relation

a lm + a~ 0 . (4.3.2) The (.e,m)-restriction of B:1A B0 is

.Aim Aim

a.emY- amjy-1

+(a u-amm) z )

-1 amm+ zy am.£

where y and z are computed with the algorithm of theorem 1.5. Let the (t,m)-restriction of Q.£m be

(

co. s cp - sin cp) Q£m

sJ.n cp cos cp

Using (1.2.4) and (4.3.2) we find the angle of rotation cp of Q£m

from

where w == - amtx + a tmY + (a.er amm)z.

The same diagonalizing shear T£m may also be constructed as the

product of the lower triangular representative 1£m of ~£m(A) and

an orthogonal shear R.em: T.em = L.em R.em· Then

(

)-'- 0 J (cos 4 ~ sin 4) ZX

2 X SJ.n (j; COS (j;

The rotation parameter q; of R£m has to be determined by the equa­

tion (4.3.2).

Page 98: An eigenvalue algorithm based on norm-reducing transformations

Similarly to (4.3.4) we find for the rotation parameter q; of the

orthogonal factor Rtm

(1-z 2)a + i'a - xz(a - a ) tan 2~ = ------~~~m~--~mt~_,~~£~~--~mm==­

x(a~£- amm)+ 2za£m

- X!N + 2a £m

x(au- amm)+ 2za£m ' ~ < \jJ..;; n/4

where w = - am£x + a£my + (at£- amm)z.

In the next chapter we shall explain that for reasons of numerical

stability it is advisable to derive the diagonalizing representa­

tive To of ~o (A) in the case of x..;; y from B with (4.3.4), kill kill £m

and in the case x > y from 1 in accordance with (4.3.5) and £m

(4.3.6).

4. 4. The complex diagonalizing representative ofn"Ztm(Al

For a complex matrix, too, it makes sense to perform simultaneous­

ly the normalization process (described in chapter two) and a

Jacobi-like diagonalization process. ThereTor we have to select

an appropriate element from 7Jt,£m (A), the class of unimodular op­

timal norm-reducing shears. In contrast with the analogous real

problem the unitary shear which diagonlizes the (£,m)-restriction

of the Hermitean part differs from the unitary shear that r;:ini­

mizes the departure of diagonal form.

We propose to select, as has been suggesced by Goldstine and Hor­

witz for normal matrices [10], the transformation shear T E:7n (A) £m £ID

in such a way that the departure of diagonal form is minimized~.e.

S(T:1A To ) = min S(P:1A P ).

kill kill p E: m (A) km £m £m £m

The shear T£ru will be called the diagonalizing representative of

m t~(A). In this section we describe the construction of the diagonalizing

representative of ~tm(A).

97

Page 99: An eigenvalue algorithm based on norm-reducing transformations

As .we have seen in theorem 2.2

r: Cla! 12 + ja! [2 + ra'.[ 2 + [a'.f 2

) = ccx + PY + 2Re(yz). • .n U J.m .EJ. mJ. J."'"' ,m

Hence this sum is independent of the choice of the representative

from the class of row congruent matrices with Euclidean parameters

(x,y,z). Consequently, for the determination ofT" E m (A) for "'m ;;m

which (4.4.1) holds, we may confine ourselves to minimizing "-1 ....._ A

S(P..em A P,em) with ?n,em(A). Let A be a pre-treated matrix,i.e.

am£ = 0 and let (x,y,z) be the Euclidean parameters of the row

cong!'l<ent shears of ""in ,em (A).

The representative of ~..em(A) will be considered as

the product of the upper-triangular rep~esentative B of ~ (A) ,em ..em and an appropriate unitary shear Q. • Hence, T := B. Q •

"'m ..em "'m :em The (..e ,m)-restriction of B..em is

) Let the (..e,m)-restriction of ~m be

Q = ( cos cp -e -:ie sin cp)

:em i8 e sin cp cos cp

Now the nni tary factor Q..em has to be

(4.4.1). This means that we have to

S(Qf B:1

A B0

Q0 ) = min

obtained from relation

select Q..em in such a way that

S(P* B-1 A P0 m) • .Em .em "' "'m "'m ,m "'m p E u

J:m J:m

We easily find that the (.e,m)-restriction of B;~ A BJ:m equals

( ao.U wamm) (4.4.4)

where w := y a..em + z(a;;;;- amm). (4.4.5) Then the (.e,m)-restriction of A' := Q* B-1 A B Q is :em J:m .em .em

98

Page 100: An eigenvalue algorithm based on norm-reducing transformations

a.£.£- VSlll <p 1. -ie . 2

A' = .£m (

. 2 --zve sJ..n2 cp + wco s ;:p )

(4.4.6) 1. iS .

""2Ve Sln2cp 2iS . 2 - we s1n ;:p

• 2 1. iS ·n2 amm+ vs1n ~ - ~e Sl ;:p ,

where v := au- amm. (4.4.7) A I

If f(cp,e) := 82 (A.£m) , we have

f(<p,S) = l'-ivei8sin2cp - <pl 2 + j'-ive-i8 sin2cp + wcos2 cp[ 2

= i[vi 2 sin2 2cp + lwl 2 (cos4(jl + sin4

<p)- !Re(vwei8 )sin4q>

The parameters <p and e of Q.£m are those for which f(cp,8) is a min­

imum. The minimum value of f proves to be ij'wf. Hence we have

Let A be a pre-treated matrix, (x,y,z) the Euclid­

ean parameters of~ (A), and T0 the diagonalizing representative ,em ,.,m of ~,em(A).Let further A' = Ti~ A T,em, v = a,e,e- amm and

w a,emY + (a,e,e - amm)z. "I

Then S2 (A,em) = ijwj 2• (4.4.8)

The shear T,em is the product of the upper-triangular representative

B,em of ~,em(A) and the unitary shear defined in (4.4.3) : = B ,em Q,em .For the parameters cp 8 of the unitary shear Q.em

holds

<p = 0 if w = o, (jl n/4

s o e = o if \) = o, w ~ o, (4.4-9)

-tan2cp ie

: ( 0 < cp < n/ 4) , e I:~ I if wv ~ 0.

" ' non-diagonal elements of A,em holds alm = i wand For the

a~ - iw if v = o, while a~ = - iW'v2/f vl.-2 if v ~ o.

We would remark that the pre-treatment of the

causing the (t,m)-restriction of the

..~.~-'-!S-'-lta...t. matrix,

matrix to have

upper form, renders the construction of the diagonal-

99

Page 101: An eigenvalue algorithm based on norm-reducing transformations

representative T.em of 1?'1-.e m(A) much :nore

described by Goldstine [10].

than that

In the next chapter we shall explain tt~t it is advisable to con­

st~ot the diagonalizing representative Ttm of ~.em(A) in the

case x >y from the lower triangular representative Ltm of ~tm(A~

We then makP. use of the following matrix

(

{x

" B • tm z

{x

L.em

Hence T ,em = B tm Qtm 1 tm R tm Qtm , where Qtm and R ,em are unitary

shears.

Unlike the algorithm described in theorem 4.2., the Goldstine­

Horwitz algorithm for the diagonalization of normal matrices,based

on minimizing the departure of diagonal form, does not guarantee

convergence to diagonal form. Voyevodin and [32] have con-

structed a class of non-diagonal normal matrices for which the

Goldstine-Horwitz algorithm becomes stationary before diagonal

form is reached. For matrices of this class holds for each pivot-

(,e,m)

laPm12 + 1am,el2 = lti2(Atm).

Hence we cannot expect that for all matrices A the sequence {~},

which is constructed with the representatives of

nv1

~ (Ak_1

) and converges to normality, also converges to k' k

diagonal form.

100

Page 102: An eigenvalue algorithm based on norm-reducing transformations

CHAPTER 5

NUMERICAL STABILITY AND THE NORM-REDUCING PROCESS

5. 0. Introduction

It goes without saying that each author in the field of er-

ror analysis has to acknowledge the impact of the enormous amount

of results obtained by Wilkinson in a series of papers culminating

in the detailed expositions entitled "Rounding Errors in Algebraic

Processes" [34] and nThe Algebraic Eigenvalue Problem" [35] • In

the latter book Wilkinson gives general error analyses which apply

to almost all the numerical methods for the eigenvalue problem that

are investigated in that book~ These analyses enable Wilkinson to

assess the numerical stability of several algorithms.

For our norm-reducing algorithm it is required to ascertain that

the numerical error associated with the actual computation of an

approximation of a similarity transformation, is bounded in some

appropriate sense.

In [35] Wilkinson has given a general err6r analysis of eigenvalue

techniques based on similarity transformations.

Let A A , A , A , •.. , ~ be the successive transforms of 0 0 ' 1 2 --k

A computed with inexact arithmetic. Owing to the rounding errors 0

ih the similarity transformations, A (p = 1,2, ••• ,k)is onzyalmost p

similar to A0

• The resulting matrix ~ , obtained after k steps ,

can be considered to be a perturbed exact transform of A 0

Ak = Pk-1 Ao pk + G(k) (5.0.1)

Wilkinson shows that 'it also may be advisable to consider~ as an

exact transform of the original matrix with some perturbation :

~= (5.0.2)

Wilkinson ([34], page 125) explains that "Bounds for G(k)are like­

to be usefui a posteriori when we have computed the eigensystans

of ~ (at least approximately) and have estimates for the conditions

101

Page 103: An eigenvalue algorithm based on norm-reducing transformations

of ·its eigenvalues. A priori bounds for F(k), on the other hand,

enable us to asses the algorithm" (viz., with respect to its numer'­

ical stability).

Useful a priori error bounds for F(k) and G(k) have been found for

quite large classes of transformations based on the use of unitary

transformations. "The great numerical stability of algorithms in­

volving unitary transformations springs from the fact that , apart

from rounding errors, the 2-norm and Euclidean norm are invariant,

so that there can be no progressive increase in the size of

the elements in the successive transformed matrices. This is im­

portant because the current rounding errors at any stage are essen­

tially proportional to the size of the elements of the transformed

matrix' ([34], page 162). Therefore, our almost diagonalization(by

unitary shear transformations) of an almost normal matrix can be

performed in a numerically stable way.

For non-unitar,y transformations the situation is less favourable;

useful a priori bo1mds for F(k) and G(k) do not exist when A is a

general matrix. Wilkinson remarks that"···· when we use a trans­

formation matrix with large elements, the rounding errors made in

the transformation are equivalent to large perturbations in the

original matrix'' •

As we have seen in the chapters one and two, the elements of opti­

mal norm..: reducing shears may be very large. Hence it is not a priori

certain that our norm-reducing algorithm will be numericallystable.

In the sequel of this chapter, we shall prove a "local stability"

result, namely, that it is possible to perform the transformation

with Tk in the k-th step of the norm-reducing process in such a

way that - -1 c ) ~ = Tk ~-1 + Fk-1 Tk + Gk '

where both IJFk_1 !1E/II~-1 1lE and HGk!IE/II~IIE are small{irrespective

of the condition number of Tk).

102

Page 104: An eigenvalue algorithm based on norm-reducing transformations

On the basis of this result we would venture the following

tative argument, from which a global result would follow.

Let ak_1 be the of Ik_1, the spectrum of ~-1 + hence also of - Gk' and ak the spectrum of Ik. Then (dependent

on the condition of the eigenprcblem of ~-1 ), ok~ differs little

from ak_1 iff 11Fk_111/lf\:_1 [[ is small, and (dependent on the condi­

tion of the eigenproblem of Ik) ok differs little from ak-t iff

ftGkll/1!\:l! is small. Further we remark that, since we are normal­

izing, the condition of the k-th eigenproblem improves as k in­

creases (at least in a global sense)o Therefore, we feel that our

local result described above explains, essentially,the a

observed of our norm-reducing algorithm.

The possibility of performing the transformation by Tk in a stable

way can be sketched in the following manner.

As we have seen, the decrease of the departure of normality a

nonn-reducing shear transformation with pivot-pair Ce,m) does not

depend on the choice of a representative Ttm from the class ~m(A)

used in the transformation

A'

So it makes sense to try to find an element S"m of ~ (A) for ..v ,em

which the floating-point computation of A1 is numerically stable in

the backward sense, i.e.

A' st~1 (A +F) s..em (5.0.5)

with I!F!IE/ I!AI[Esmall. Elements from ~ ..em(A) with this

will be called stable representatives of ~tm(A). In the sectiam

4.3 and 4·4 we have argued, however, that it is advisable to use

the diagonalizing representative T..em ·Of »v..em(A). But· since the

elements of nttm(A) are row congruent, there exists a shear

U..em so that Ttm Stm Utm • Thus the computation can be performed

in a stable way as follows

-1 A11 ::o S.8m A S.8m , A 1 (5.0.6)

103

Page 105: An eigenvalue algorithm based on norm-reducing transformations

Then for

A"

and forA'

Hence

A'

we have, conformally to (5.0.5)

-1 s-tm (A + F) stm

U -:1

A" TJ + G tm tm '

(5.0.7)

where 11 FHEis small relatively to !1 1\I!Eand ff q!Eis small relatively

to I[Ai !lg

In order to determine the stable representatives of ~tm(A) we

first investigate the characteristics of the rounding errors.made

in matrix multiplications (sections 5.1 and 5.2). In section 5.3

the res~lts of this general error analysis will be applied to shear

similarity transformations. In section 5·4 we shall shOw that ei-·

ther the upper triangular or the lower triangular representative.

of JnJ £ m(A) is a stable representative of this class and that this

is independent of the largeness of the Euclidean parameters of

m 1 m(A). Finally in section 5.5 it will be shown that the diago­

nalizing representative TA; m of J/ttm(A) itself is a stable repre­

sentative if

Hence in this case (which occurs if A is almost diagonal and a ' 11

and a are well separated) the decomposition mentioned in {5.0.6) mm -1

is not necessar.y in order to compute Ttm A Ttm in a ·stable way •

5. L Input and output perturbations related to rounding errors

5.1.0. Following Wilkinson ([34], pag, 4) we use the symbol ft to

indicate that the computed results are obtained with floating-point

arithmetic.

104

Page 106: An eigenvalue algorithm based on norm-reducing transformations

If the floating-point representation of a number has a binaxy t­

mantissa, then we assume that the rounding errors in the sunr

mation and multiplication are such that

f,e(a+b)

fi(a*b)

(a+b) (1+e ) 1

(a*b) (1 )

(5.1.1)

(5.1.2)

For floating SLilllmation and multiplication of two cor:tplex numbers

we can analogous description of the rounding errors, using the

properties (5.1.1) and (5.1.2).

Let z 1

Then

a + i b , z 2

c + i d.

f1(z +z ) = fi(a+c)+ i fi(b+d) 1 2

(a+c)(1+o )+ i(b+d)(1+o ), 1 2

-t with ~ 2 • i 1,2.

'I'hus for the modulus of the error we find

Jn(z + z)- (z + z )[ 2 = (a+c) 2 o 2 + (b+d)o 2 ~2-2t Jz + J2 • 1 2' 1 2 1 2 1

So there exists a complex number e , Je J~2-t, for which holds 1 1

Le(z + z ) 1 2

(z + z ) (1 + e ). 1 2 1

(5.1.3)

E ' 2

z 1

we can derive that there exists a complex

<2 -t+ 3/2 (1+2-t-1) for which holds

z z (1 +e ) • 1 2 2'

number

So with cor:tplex arithmetic the number of significant bits is essen­

tially one and a half less than that for real arithmetic.

5.1.1. The given d~scription of the rounding errors for real and

complex floating-point arithmetic enables us to give error bounds

for floating-point matrix computations. We state the following re­

sults wi tr~ou t proof (Wi lkinson [ 34], page 8 3) •

Let fi(A,x) denote the floating-point product of the matrix A and

the vector x.

Then for

e := fi(A,x) - Ax

105

Page 107: An eigenvalue algorithm based on norm-reducing transformations

we have

(5.1.6)

The elements of the vector jej and the matrix /A/ are the moduli

of the elements of e and A respectively. For real single-length

arithmetic • -t ( ) n =n2 5.1.7

whereas for complex arithmetic

n ~ n 2-t+3/2 (5.1.8)

Here and in the sequel we shall use the symbol ~ with the follovling

meaning. If a ~ b, then

a b (1 + 0(2-t)) , (t '-+ oo),

Similarly we use the symbol ~ in the following sense: if a ~ b

then there exists a number b*, with b* ~ b, for which holds that

a .-; b*.

These symbols enable us to avoid second order terms which tend to

obscure the fundamentally simple results. With the quantity n• de­

fined in (5.1.7) and (5.1.8), for real and complex arithmetic res­

pectively, we are able to estimate the rounding errors in the

floating-point product of matrices.

Let ft(A,B) denote the floating-point product of the matrices A

and B. Then for

E := .te (A, B) - AB

we have

(5.1.10) In the rest of this chapter we consider floating-point computatiom

having the characteristics given in this subsection.

5.1.2. We now investigate the multiplication of a non-singular ma­

trix A and a vector x:

(5.1.11)

106

Page 108: An eigenvalue algorithm based on norm-reducing transformations

e := fi(A,x) - Ax. (5.1.12)

Let f and g be any vectors

A f + g = e, (5.1.13)

then we can write

U(A,x) = A(x +f) +g. (5.1.14)

The vector fin (5.1.13) will be called the input or per-

.:::..:.;;_:....;;.;..;;. or forward perturbation corres-

ponding to the error e in this process. The

ward) error analysis (Wilkinson [;5]) corresponds to the

tion of the errors for which the output (respectively input) per­

turbation is taken zero.

Definition 5.1. p(A,x) is a stability bound for the

point computation of Ax if there exist vectors f and g for which

holds

and

llgfl2

.,.; 1') p(A,x) !1Axlf2

,

and which satisfy (5.1.13).

(5.1.15)

(5.1.16)

We now consider several ways in which the error e may be distrib-

uted over a forward part g and a backward part f in accordance

with the rule e = A

5.1.3. -1 f =A e. Then (5.1.6) implies

I fl .,.; I A -1

1 I el .,.; 1'J I A -1

1 I AI I xl •

Thus

llf!l .,.; ") C 0 (A) llxll , 2 N 2

where

c1 (A) := IIIA-1

1 ]AI 1!2

(5.1.17)

(5.1.18)

(5.1.19)

107

Page 109: An eigenvalue algorithm based on norm-reducing transformations

The quantity c£(A) will be called the left-condition number of A.

(5.1.18) shows that for all x, c,e(A) is a stability bound.for the

computation of Ax.

It is important to observe that c,t(A) is invariant relatively to

row-scaling of A. In fact, for non-singular diagonal matrices, D,

holds

JA-1

1 IAI = jA-1

D-1

1 IDAJ.

Hence

C n (A) ~ inf C ( I DA j) ' x D diagonal 2

(5 .1 • 20)

where c (A) = llAII IIA-11! 2 2 2

5.1.4. Forward error analysis •. We now take f = 0, thus g e.

Then (5.1.6) implies

I gj ~ 1J I Aj I xl ~ 11 I AI I A -11 I Axj • (5 .1 • 21)

Thus

11 gll ~ 11 c (A) 11 Axll , 2 r 2

(5 .1.

where

c c A) = = 111 AI 1 A-1

1 n • r 2 (5.1.23)

The quantity cr(A) will be called the right-condition number of A.

(5.1.22) shows that for all x, c (A) is a stability bound for the r computation of Ax.

c (A) is invariant relatively to column-scaling of A, since for r

non-singular diagonal matrices, D, holds

Hence

c (A) ~ inf· c (jADj). r 2 D diagonal

Mixed forward and backward error analysis. We shall dennn-

strate tr~t there is a non-singular matrix A and a vector x for

r:hich all stability bounds p(A,x) are very large. That means that

despite an opt.imal distribution of the error e in a backward part

f and a forward part g, these backward and forward perturbations

1 08

Page 110: An eigenvalue algorithm based on norm-reducing transformations

are not small relatively to x and Ax respectively.

Let

where f and g are such that

Af + g e U(A,x)- Ax).

(5.1.25)

evidently, M is a stability bound and, moreover, for all sla.­

bounds p(A,x) holds

then

'I.' he

M..: p(A,x).

if p0

(A,x) is the infimum of all stability bounds p(A,x) ,

M.;;; p .;;; M. 0

~~~r~r1g~)~n multipliers we find easily

(llxll 2 A AT + IIAxll 2 I)-1 e TJ -2

2 2

2 lf(llxll 2 A AT + I[Axl 2 I)-1 I!

2 2 2

-2 T)

(5.1.26)

(5.1.27)

(5.1.28)

in this estimate of M holds if and only if e is an

of A corresponding to the smallest, eigenvalue

nA-1!1 - 2 of A 2

On the basis of this result it is easy to construct for each cor­

arithmetic a non-singular matrix A and a vector x

so that all

of Ax are of

for the floating-point computations

• Therefore the matrix A and the vector x ha:ve

to the following conditions:

(i) c1 (A) and cr(A) is large;

(ii). xis such that 11Axll /l!xll is near IIA-1

11 -1

; 2 2 2

(iii) e is to be near an eigenvector of A AT corresponding to the

smallest ' viz. IIA - 1 n -2 ' of A AT. 2

109

Page 111: An eigenvalue algorithm based on norm-reducing transformations

Let us now consider a correctly rounding floating-point arithmetic

with a mantissa of t bits.

A possible choice of A and x is

( 1-2-k 1+2-k) ' A=

1+2-k

·ll!here k := (t+3) : 2.

Then we find

by

( 1 +2-k \ X= -k) -1-2

Ax = 2-k,..l ( 1 +2-k) (-~ ) , e = 2-.2k+l ( ~~) We observe that, indeed, e is an eige~vector of A AT, correspond­

ing to tr.e ei.genvalue 2-2k+2 of A AT. With (5 .1 .28) vm find . _2 -2 -2k-1 ( -k) -2 • -1 -k-t

lVl = Y) 2 1 +2 • Thus M = 11 2 • Hence it folJ.ows

from (5.1.26) that for all stability bounds p(A,x) holds

( ) . -1 -k-1 • -t p A,x, :;;;. 17 2 • Since 'f) = 2.2 we find

( ) • t-k-2 t ! 2 -3 p A,x :;;;. 2 = 2 • •

5.1.6. ~.~atrix factorization. In the algorithm for the corr.puta-

tion of Ax to be examined in this subsection we suppose to have at

our disposal two matrices, tl:.e exact product UT of these matrices

equals A. The algorithm consists of two steps in which the factors

T and U of the factorization are used consecutively

b := f£ (T,x) c := ft (u, b).

Let e := b - Tx, g := c - Ub.

Then, i~ conformity with (5.1.6)

For the ultimate result of the process we have

c = Ub + g = U T (x + T-1e) +g.

(5.1.29)

(5.1.30)

(5.1.31)

. -1 . We now consider f := T e as a backward perturbation ar~d the error

110

Page 112: An eigenvalue algorithm based on norm-reducing transformations

g, generated in the second step of the process as a forward pert~

bation. Thus

c = A (x + f) + g.

From (5.1.31) we find for f

I fl = j T-1ej ~ I T-11 I ej ~ 1J [T-11 JTI lx I· Thus

llfll ~ TJ c n(T) llxll • (5.1.32) 2 h 2

For the forward perturbation g we find

I gl ~ 11 /UI jb I = TJ j·Jj jT(x + f) I ~ 1'J ju I lu-11 jA (x + f) 1-

Thus l!g!l2 ~ n, cr(u) !fA(x + f)ff

2, or

ngn ,;; n c (u) IJAxH • 2 r 2

We conclude from (5.1.32) and (5.1.33) that

max (c _/T) , cju)) J.

(5.1.33)

(5.1.34)

is a stability bound fer the computation of Ax vri th the algorithm

(5.1.29). Obviously, if both c;;(A) and cr(A) are very large, it

makes sense to try to find a factorization of A = UT such that the

stability bound mentioned in (5.1.34), is not largeo In the next

subsection we shall mention a recipe for the factorization so that

for each A: max (c_/T), cr(U)) ~n.

5.1.7. Principal factorization. For each matrix A the princ~

factorization

A U 11 V

exists. In (5.1.35) U and V are unitary and 11 is a diagonal matrb4

The diagonal elements of 11 are the principal values of A.

Let T := 11 V. If we have at our disposal the matrices U and T,then

it is possible to perform the algorithm (5.1.29) for the computa­

tion of Ax.

111

Page 113: An eigenvalue algorithm based on norm-reducing transformations

For the factors U and T we have

and, since the left condition munber is ii:variant relatively to

::cow-scaling, also

) = c_e(v) .;;; n.

:ience it follows from (5.1.34) that n is a stability bound for

the comp1~tation of Ax if the principal faCtorization is used.

We observe that this factorization would only be practicable for

n .;;; 2. In section 5.4 we shall mention for n = 2 an even more sim­

ple fao~crization.

5. 2. Error analysis of similarity transformations

5.2.0. We will now (;Onsider algorithms for the computation of tne

similarity transformation A1 = P-1A P of A with P. Let be the

r-?sul t of the algorithm by which P-1 is computed. Let

, f,e(A,P))

and

If F and G be any matrices satisfying

P-1 F P + G = E

then Yre can write

A' = P-1 (A + F} P +G.

(5.2.1)

(5.2.2)

The mtrix F in (5.2.3) will be called the input or baclnvard per­

turbatio:!1, the matrix G the output or forward perturbation corre­

sponding to the error E generated in this process.

The backward (forward) error analysis corresponds to that descrip­

tion of the errors for which the output(input )perturoat:Lon is taken

zero.

11

Page 114: An eigenvalue algorithm based on norm-reducing transformations

Def:ini tion 5. 2. p (A, p) is a stahli ty bound for the floating -

point computation of P-1A P if there exjst matrices F and.G for

which holds

and

11 E.,.; 0 p (A,

and which satisfy

IIP-1 A Pll E

(5.2.;;).

(5.2.6)

In the algorithms to be considered we distinguish two parts. In tbe

first part a numerical approximation P-1 of P-1 is ccmputed,in the

last an approximation of p=1 A P. This resu.l t is called A 1 •

5.2.1. Backward error anal;ysis. -1 We now take G = 0 and F=P E P •

We may write

-1 -1 ( ) P = P I + K • 1

K equals the residual matrix P P-1 - I, Let 1

IlK 11 .,.; k Tl· 1 E 1

(5.2.7)

(5.2.8)

If P-1 equals the correctly rounded P-

1, then l!K 11 .,.;TJk n-ic (P) ,

1E r where k is independent of the condition of P.

The computation of an approximation P-1A P, in which P-1 is used,

is performed in two steps:

Let

Then

where

B := fi(A,P), C := fi(P-I, B).

E1

:= B - AP, -1 E

2 := C - P B.

= P-1 [ (I+ K1

) (A+ E1P-1) + P

= P-1 (A +F) P ,

F (E + p ) + K (A + E I-I). 1 1 1

(5.2.10)

J p (5.2.11)

(5.2.12)

113

Page 115: An eigenvalue algorithm based on norm-reducing transformations

From (5.1.10) we conclude

IE11,;;; T) IAI !PI ' !E2/ ,;;; T) /P-1

1 /B/. (5.2.14)

These bounds for !E1 r and IE2 r' together with the bound for IIK1 liE

given in (5.2.8) will be used in order to estimate IIFIIE'

From (5.2.14) we derive

lE, P-1

/ ,;;; !E11 !P-1

I ,;;; T) !AI !PI ,P-1

/ •

Hence, since 11 !PI rl'-11 liE ..; nt H IPI IP-

11112 = ntcr(P)'

(5.2.15)

With (5 .2, 7) we derive from (5.2.14)

1 E 1 ,;;; T) c1 + TJ) IP-1

1 1 r + K 1 r AI 1 PI ~ T) r P-1

r 1 AI r P/ • 2 1

Hence

I p E2 p-1, ;; T) IPI jP-11 I Aj I Pj I p-1, •

Thus

(5.2.16)

For theEuclidean norm of the term K1

(A+ E1

P-1) in (5.2.1-3) we

find

HK1 (A + E1 P-1

)liE ,;;; IIK1 liE (IIAIIE + T) !!All 11 I PI [ P-1

1 liE)

~ T) k1 IIAIIE •

Combination of (5.2.15), (5.2.16) and (5.2.17) gives 1 1 .

IIFIIE '-' T) [n2 c (P) (1 + n2 c (P)) + k] IIA!IE' r r 1

We conclude from (5.2.18) that for all matrices A l

o (P) ( 1 + rf o (P)) + k r r 1

(5.2.18)

(5.2.19)

is a stability bound for the numerical computation of P-1 A P.

5.2.2. Forward error analysis. In this case F = 0 and G = E.

We may write

(5.2.20)

114

Page 116: An eigenvalue algorithm based on norm-reducing transformations

-1 K2 .equals the residual matrix P P - I. Let

[K HE k D. (5.2.21) 2 2

--=1 -1 ( If P equals the correctly rounded P , then I!K2

11E.,. l]k c)!

where k is independent of the condition of P. In this subsection

we use the notation introduced in (5.2.9) and (5.2.10). Then we

may write

-1 ( -1) -1 C P A + E P P + E = P A P + G, 1 2

(5.2.22) where

G = P-1E + E + K (P-1A P + P-1 E ) • 1 2 2 1

(5.2.23)

In order to estimate !!Gl!E' we derive upperbounds for liE , -1 -1 -1 . -1

trP '~lE and rnp A P + P E1

jfE relatively to [[P A P!IE.We start

to relate these norms to IIC'~E'

Since I .,.. fJ [P-11 IB[ and B = (P-

1)-1 (c-E ) ~ )we have 2

[E),;; n LP-1

1 [PI [c-E2

[ ~ 11 I I [PI [c[. Hence

~

lfE2

UE.,.. 11 112 c£(P) ilC!!E • (5.2,24)

For [P-1[ [E [ we obtain from (5.2.14)

1

I' P-1 r 1 E

1 r ... 7) r P-

1 r 1 AI !PI. Since A (P-1 )-1 (C-E ) P-1 - E , we derive

2 1

,p-1riE r ... l],p-1 rpr·rcr+K )-111 " +7JIP-111E IIP-111PI 1 . 2 1

~7J[P-1 r-1Pr rei [P-1

[ :P[.

Thus

(5.2.25)

Finally estimating I[P-1 A P + P-1 we find with (5.2.24) :

-1 -1 '( ~p A ? + P E 11'!= = r. I + !I 1 .]:!;

115

Page 117: An eigenvalue algorithm based on norm-reducing transformations

(5.2.24), (5.2.25) and (5.2.26) we find the following upper­

bouEd of !IGIIE : . ],_

frGI!E ~ YJ [n2 c .t(P) (1 + c .t(P)) + ] !!CifE • (5.2.27)

Since C P-1

A P + G, IIP-1 A P!!E ~ I!Cl!E • Hence from (5.2.27) fol­

lows that

(5.2.28)

We conclude from (5.2.28) that

(5.2.29)

i:s' a stability bound for. the numerical computation of P-1 A P •

We would remark that in the estimate (5.2.18), of the hypothetical

backward perturbation F, the right condition number cr(P) appears,

whereas the left condition number c.t(P) occurs in the estimate

(5.2.28) of the hypothetical forward perturbation G.

This difference in the estimates of !fFifE and IIGI!E suggest the type

of error analysis to be chosen in order to obtain a sharp estimate

of the errors generated in the algorithm for the numerical com­

putation of P-1 A P.If c..e(P) << cr(P), then it is advisable to use

backward error analysis; if cr(P) << c /P), then forward error

analysis is preferable. In the next subsection we describe an al­

gorithm for t~e computation of P-1 A P which is of interest if

both c..e(?) and cr(P) are large.

5.2.3. Mixed error ana~sis by means of factorization. In the al­

gorithm for the comp~tation of P-1A P, which will be considered in

this subsection, we suppose that P is the exact product of given

matrices T and U :

P T U.

-1 -1 Uoreover we suppose to have at our disposal T and U , being

the approximate inverses of T and U respectively.

116

Page 118: An eigenvalue algorithm based on norm-reducing transformations

-1 The algorithm for the computation of P A P now consists of four

steps :

A := fe (A,T) 1

A := f-l(A , 3 2

E :=A -AT '

A := f£ (T-1

, A ) 2 1

A := f;;(u-1, A).

4 3 .

E := A - A 1 1 2 2 1

--=-1 (5.3.32) := A -A u E :=A A E ' - u

3 3 2 4 4 3 Let

T-1 -1 (I + K ) :;""1 (I ) -1 = T ' = + u 1

We shall assign the rounding errors of the computation of A to 2

the backward perturbation F, whereas the errors of the second half

of the process will be assigned to the forward perturbation G.

Then

A := P-1 (A +F) P + G, 4

where, similarly to (5.2.13) and (5.2. ),

F = (E + )T-1 + K (A + E T-1 ) 1 1 1

and

G = u-1 + E + K (u-\ U+ u-1 4 2 2

) . From ( 5.1 .1 0) we know

1 E1 r .s; rr 1 AI 1 Tl. , IT-11 lA I

1

[ E31 .s; 1'] IA21 I u I ' I I I A I .• 3

Let further

(5.2. )

The estimate (5.2.18) may be applied here to the backward pertur­

bation F of (5.2.35). Then we obtain • .l.. .l..

!!FilE .s; 11 [ ].12 c (T)(1 + n2 c (T) )+ k] UAIIE • (5.2.38)

r r 1

The estimate (5.2.28) may be applied here to the forward pertur~

117

Page 119: An eigenvalue algorithm based on norm-reducing transformations

bation G of (5.2.36). Then we obtain • _;!,_

!IGIIE,;;;; 1 [ n2 c_e(u)(1 + (u))+ k] 2

(5 .2. 39)

We conclud.e from (5.2. ) and (5.2.39) that _;!,_ -~ _;!,_

ma:x:{n2 c (T)(1 +n2 c (T))+k, n2 c (u)(1 + r r 1 £

(u))+ k } 2.

is a stability bound for the computation of P-1A P with

rithm (5.2.31).

(5.2.40) algo-

If both c;; (P) and cr(P) are very large, then it makes sense to try

to ffnd such factorizations of P = TU that the stability bound

given in (5.2.40) is not

As in subsection 5.1.7 we now apply the result obtained in(5.2.40)

to the particular case that the decomposition P = TU is derived

from the principal factorization P = V /1. U, thus T = V/\.

Then c0(U),;;;; nand c (T) = c (v)...::n.

"' r r Hence, in this case

max (n3/2(1 + n3/2)+ k1' n3/2(1 + n3/2)+ k2)

is a stability bound for the computation of P-1A P with the algo­

rithm (5.2.31 ).

It is appropriate to remark here that the principal factorization

is rather impracticable unless n,;;;; 2. In the next section, however,

we apply the above results to shear similarity transformations.

The error bounds (5.2.38) and (5.2.39) will appear to be of prac­

tical value for these transformations.

5. 3. The general error analysis applied to shear transformation

5.3.0. In this section we apply the results of the preceding sec-

tions to the similarity transformations by unimodular shears :

-1 A 1 :=T AT•m' . ..em "'

118

Page 120: An eigenvalue algorithm based on norm-reducing transformations

We assume to have at our disposal the Jacobi-parameters p, q, r

and s of t~e shear Ttm. In order to simplify the notations we de­

fine P := T • .£m Thus

p = c :J Although generally ps-qr is not exactly equal to one, we take

p:-1:= (s -q) -r P

(ps-qr)P-1 (5.3.3)

as an approximation of P-1 • With this (..e,m)-restriction of T..em-1we

compute an approximation of T.£m-1A T..em in two

C := f.£(T.£m-1, B).

The errors -1

in the computation of this approximation of

T. AT. may be ..vm ,;r,m assigned to a hypothetical backward pe:!:'-

turbation F; we may also assign them

bation G :

to a forward pertur-

5.3.1. We start to examine the errors made in the computation of

the affected elements of A1 , i.e., the elements in the .£-th and

m-th column and row but not

The affected elements of A1

"' . ' belong1ng to A..em·

on the i-th row, where i I .£, i I m, T are obtained from the product of a row vector, say, and the

matrix P :

(a:I_..e a! ) "' (au a. ) (p q) T J.m J.m r. P i l..e,m. J. r s

(5.3.6)

Let T Tp + T T T) P. f.£(r. ,P) = e. (r. + J. J. J.

119

Page 121: An eigenvalue algorithm based on norm-reducing transformations

T T That means, e. and f. are the l l

bation, respectively, caused by

Then, since I ei Tl ..,; TJ I r/1 I Pi ,

the backward perturbation f.T :

forvrard and the backward pertur­

the numeric~l computation of r.TP. l

we find, similarly to (5.1.18),for

l

!I T T I! ..,; 11 c (P)Ifr. 11

2 r l 2

and, similarly

T lie. 'I ..,; r; l 2

to (5.1.22), for the forward perturbation

c ( P) 11 r. T P 1! • £ l 2

(5.3.8)

T

We now consider the affected elements on the i-th column, i I t ,

U m. These elements of A1 are obtained by multiplying the matrix -1 P and a column vector, c. say, Thus

l

(ali) ( s -q) (aJ:i)

a 1 • -r p a . ml m1

P-1 _j ci , i i ,e ,m.

Let

(c. +h.). l. J..

(5.3.11)

That means, g. and h. are the forward and the backward perturba-1 l -

tion, respectively, caused by the numerical computation of P-1 c .• l

'l'hen, since 1 gil .,. n 1 F-11 1 oil, we find directly from c5.1.1s)for

the backward perturbation h. l

[lh.!l ..,;YJ· c0

(P-1) l!c.l! ~YJ c (P) Rc.t[ , · 12 "" 12 r 12

and from (5.1.22) for the forward perturbation

( -1) • ( ) -1 ITg.!l ... Y) 0 p 11 c.\! = YJCn p liP c.rr • 12 r 12 "" 12

As we see from (5.3.8) and (5.3.12), the estimate of the error

which is entirely'assigned to a backward perturbation depends on

cr(P), whereas ct(P) determines the estimate of the error which

is entirely assigned to a forward perturbation,

In this subsection we consider the errors made in the nu-

merical computation of the (t,m)-restriction of At, For the back­

ward(forward) error analysis we investigate and estimate the back-

120

Page 122: An eigenvalue algorithm based on norm-reducing transformations

ward perturbation F n (forward perturbation G0 ) corresponding to hill hill

the errors generated in the numerical computation of the (£,m) -

restriction of A1 • These perturbations Ftm and G}i;m are the (t,m) -

restrictions of the perturbations F and G respective~ which are

introduced in (5.3.5). Consequently

Let -1 -1

liP P - IIIE ..,;;; k1

TJ and liP P - IITE < k2

TJ • j,_

from (5.3.2) and (5.3.3) we have 22 jps-qr-1 I jps-qr-11..,;;; k TJ,

2

<kTJ 1

and

The results of the preceding section now may be used in order to

obtain estimates for lrF1miTE and 'llG1m!fE.

Evidently in this case we have to take n 2.

From (5.2.18) we conclude

I!F.£miTE.;,; 2tcr(P) (1 + 2tc)P)) + k] ITAJ;mlfE (5.3.15)

5.3.3. We shall now take together the results obtained with the

backward and forward ana~sis applied to the affected elements and

to the (t,m)-restriction. For the backward perturbation F holds

where f.T and h. are defined according to (5.3.7) and (5•3.11) re-~ ~

spectively. From (5.3.8), (5.3.12) and (5.3.15) follows 1 1 A

OFIIE2 ;;n2 [22 c (P)(1 + 22 c (P))+ k ] 2 11An llE2 +TJ2 c 2(P)(a + ~), ' r r 1 hm r

wher~, with the notation of (2.2.10),

121

Page 123: An eigenvalue algorithm based on norm-reducing transformations

Co:nsequently A • .1.. 1

IIFIIE,..; 1'J[22 cr(P)(1 + 22 cr( )+ k1

] (IIA.£m!IE2+ a + ~ • (5.3.17)

Similarly we find for the farward perturbation G

I!GIIE 2 = !IG.-em!IE2+ . L: (!lei Tn 2+ I! gill 2)' l;(e,m 2 2

where e. T and g. are defined according to (5.3. 7) and (5.3.11) l l

spectively. From (5.3.9), (5.3.13) and (5.3.16) we may easily

where

ex'+ r' = ._J~ (!a'i..el2+ !a'iml2 + la'..eil2 + [a'mi12 ). lr..v,m

re-

de-

The interpretation of the estimates (5.3.17) and (5.3.18) for the

backward and forward perturbation is very simple. With the back­

ward analysis namely we find that

(5.3.19)

is a stability bound for the computation of T.£m-1A T.£m and with

the forward analysis we find that also :1 1

22 c.£(P) (1 + 22 c.£(P)) + k2

is a stability bound for the computation of T.£m-1A Ttm'

5.3.4. Left and right condition numbers of unimodular shears.

Let P be the (t,m)-restriction of Ttm' where T.£m is a unimodular

shear with pivot-pair (t,m) and Jacobi-parameters p, q, r and s.

In order to compute c.£(T.£m) = ct(P), we have to find the largest

eigenvalue of [P-11 I Pj (j P-11 I PI)*. Thus ct(P) is the square

root of the largest zero or· the polynomial

122

Page 124: An eigenvalue algorithm based on norm-reducing transformations

Hence

o_/(P) = (lpsj-jqrl f+ 2(!prl+jqsj )2 +

+ 2(jprj+jqsj) [(/PI2+1qj2)(irl2+fsl2)

to (2.1.4)

x = I PI 2 +j ql 2 ' y =/ r/2 +j sj2 ' z = pr + qs • ..l.

Consequently jprl + jqsl ~ (xy) 2 • Since we assume that Pis the

(t,m)-restriotion of the unimodular T , tm

xy - I zj2 = I ps - qrl2 = 1 •

Moreover (lpsl-lqrj )2 ~ lps-qrj 2 = 1. With these estimates

we have

C 2 (T 0 ) ~ 1 + 4xy = 5 + 4/ zl 2

• ..e ..vm (5.3.21)

Analogously we find

(5.3.22)

where

;;) (5.3.23)

y'

We would remark that (5.3.21) an upper bound of c (T ) t ..em

which is the same for all matrices row-congruent to T • Since ..em the value of z may be very if T Em (A), the forward ..em ..em error analysis is not appropriate to describe the errors in the

norm-reducing process.

This result contrasts with the conclusion that can be drawn from

the upper bound (5.3.22) for c (T ). In the next section we ~11 r ..em

show that for each A, t and m the class ~..em(A) contains a shear

S 0 for which holds c (sn }<3. This shows that generally it is .-vm r ""m

necessary to assign the -1

putation of Stm A Stm to a

r..e(s..e~l , r..e(A, s..em)) =

backward perturbation F

-1 Stm (A +F) S..em'

123

Page 125: An eigenvalue algorithm based on norm-reducing transformations

5.3.5. We now compute the right condition number of the tri-

representatives and 1£m of mi,m(A) (see section 2.1).

For the upper triangular B£m we have 1

ey~t) :-1) ~ (y:~ c (Yo :t). B £m

2 y

With (5.3.22) we find

c/ (B..em) ,.;; 5 + 41 zj2 jy2. (5.3.24)

For the lower triangular representative 1,em of 7lV (A) tm we have

( :) c_, ") c X~ A

1£m - -t zx ·X zx 1

Hence, with (5.3.22), we find

c/(1£m),.;; 5 + 4jzj 2 /~. (5.3.25)

By direct calculation we can derive that

c (B,e ) r m = ~zj2y-2- + I zj/y (5.3.26)

and

cr(B ,em) f1+T~!2x-2 + jz!/x. (5.3.27)

.!. .!. Since I z! = (:xy-1) 2 < (:xy) 2

.,;;; t(x+y).,;;; max (x,y), we have

min (I z/ x~\ I zjy- 1) < 1.

Consequently

min ( c ( B ) , c ( 1 ) ) < 1 + 2i < 2. 42 • r ,em r .em

The triangular shear in ~ (A) with smallest right condition ,em number will be called the stable representative of ~ (A). This ,em representative will be denoted by S • So we have found that if . . _em X~ y then s B ' if X > y then s n 1 and c (s ) <1 +{2. _em _em Nm _em r _em

124

Page 126: An eigenvalue algorithm based on norm-reducing transformations

5. 4. A numerically stable transformation by the diagonalizing representative of m,tmiA)

For the real norm-reducing algorithm we have found in

section 4.3 that the diagonalizing representative of IJl., (A) ,em may be obtained from the stable representative S by multiplying

tm this shear with an appropriate orthogonal shear V ,em :

- c; V - ~£m ,em •

If x.,.; y then the rotational parameter of V,em can be computed

from formula (4.3.4); if x > y then the angle of rotation can be

obtained from (4.3.6). these matrices S and V

0 , the transformed matrix can be

,em "'m computed with the following algorithm :

A := f£(A,S,e ) A ( -1 A ) := f£ s ' 1 m 2 ,em 1 (5.4.2)

u (v£~1 A := f£(A ,V ) A ' A ) . 3 2 Alm 4 3

A A

Let s := s ,em ' V := V £m and

rr (5·4·3)

We now the rounding errors made in the computation of

to tl':e backward perturbation F, whereas the errors generated in

the second half of the process are assigned to the forward per­

turbation G :

A4

T£~1 (A +F) T£m + G.

Similarly to (5.3.17) and (5.3.18) we find

+a +

and

I' . .,.

E (Vo ) )+ k ] (I! A'\ fE +a'+ p'

.vm 2 "'m

For the triangular shear S£m we easily find that l!ss-1- I!IE.,.;

125

Page 127: An eigenvalue algorithm based on norm-reducing transformations

As· c (s. )<10; 1 + {2, we have : r .vm

Since the transformation with an orthogonal shear Vtm is well

known to be a numerically stable process, algorithm 5.4.2 can be

performed in a stable way if in the factorization T,€m = Stm Vtm

the stable representative Sim of m,im(A) is used.

5.4.2. For the complex norm-reducing algorithm the situation is

a little more complicated. As a matter of fact, in that case the

stable representative stm is derived from the pre-treated matrix

U,e~1A Uim for which holds (u,e;1A Utm)m£ = 0, Hence the diagonal-

i norm-reduction on the pivot-p~ir (t,m) corresponds to the

transformation

A' (Utm 8tm Vim)-1

A (U,em 8tm V.Em),

where V,e is an appropriate unitary shear (described in section m 1

4.4) which "diagonalizes" (U,em S.Em)- A (UJlm Stm).

Since cr(S,em) < 1 + {2 and [S,em s~1 - Ill2

<10;! rJ' A• can be com-

puted in a stable way by the following algorithm

A ( -1 , f ,e(A, u,em)) := f£ u 1 ,em

A ( -1 , f,e(A, S )) (5.4-4) := f,e s 2 ,em 1 ,em

A I= fi(V - 1 , f,e(A , vim)). 3 £m 2

We shall now show that, independent of the condition of S£m' A1

can be computed in a stable way with an algorithm in which only

two numeri.oal similarity transformations are performed.

Let PJlm: UJlm Sim and Q£m = f,e(Uim' S,em). Hence

126

Page 128: An eigenvalue algorithm based on norm-reducing transformations

-1 An approximation of T£m A T£m can be obtained in the following

way

A 1

A := f£(A , V ) , 2 2 £m

A := f£( 2

A ) 1

A := f£(V - 1 , A ) 4 ,em 3

General~ A is only approximately equal to T - 1 A T • We shall 4 ,em ,em

be content, however, if the transformation by V is computed ,em in a stable way.

Let x, y and z be the Euclidean parameters of S E 711., (u - 1 A U \ ,em ,em ,em ,em:' which can be computed with the algorithm described in theorem 2.5.

We now assume x~ y (if x > y, a similar argument can be used).

Then

m (

cos q>

e -ie sincp

p

Since jzjy-1 < 1, we easily find from (5.3.22)

cr2(p,em).;;; 5 +4[zj2Y-2< 9·

Finally,

tiple of

we shall show that !!Q Q ;; ;em ;em

-t+ .!...

11(~ 2.2 2). We have to take

Il' is a moderate mul-12

into account that

(i) the matrix U$m which is actually used for the pre-treatment,

is not exactly unitary;

the determinant of the matrix S..em actually used is not exact­

ly equal to unity.

The Jacobi-parameters of the matrix U£m' actually used for the

computation of ~m' will be denoted by (c,-s,s,c). As we see, c -ie . is an approximation of cos (p, s an approximation of e s~n cp.

The Jacobi-parameters of Sn will be denoted by 1 1 1 .~m

(y -z, zy -z, o, ;i2 ( 1 + e: 1

) ) , where I e: 1

f .;;; i 11 •

With these notations .we find for the (..e,m)- restriction of ~m:

127

Page 129: An eigenvalue algorithm based on norm-reducing transformations

1 1

c c zy -2 c 1 +e: ) - sy2 c 1 +e: ) c 1 +e: ) ) c 1 +e:- ) ) 3 1 4 5

1 1

(szy-2 (1+e: )+ cy2 (1+e: )(1+e: ))(1+e:) , 7 1 8 9

where I e: .1 ..;; 1J t . Hence 1.

det(Q)= cszy-1 [(1+e: )(1+e: )(1+e: )-(1+e: )(1+e: )(1+e: )] + 2 7 9 3 5 6

+ c 1 +e: ) c c2 c 1 +e: ) c 1 +e: ) c 1 +e: ) + I s 12 c 1 +e: ) c 1 +e: H 1 +e: ) J

1 2 8 9 4 56

~ 61 cszY-1Ie:+ (c2 +1 sl 2 )(1+4e:),

where I e: I ,.; t11.

(5.4.7)

-1 Consequently, the upper bound k

111_ of ITQ Q - I!!E will be a ~oder-

ate multiple of 1J if c2 + lsl 2 is very close to unity, so that

the shear u£m is almost unitary.

From the foregoing we conclude that in the complex case algorithm

(5.4.6) enables us to perform in a numerically stable way thenorm­

reducing shear transformation by the diagonalizing shear T£m if,

as in the real case, the stable representative is used in the fac­

torization of T • £m

5. 5. Diqgonal dominance and shear transformations

-1 5.5.0. In this section we shall show that for computing T0

AT "'m £m

in a stable way'· the factorization of the diagonalizing represen-

tative of nL£m(A) is not necessary whenever the off-diagonal ele­

ments of A in the £-th and m-throws and columns are small relati­

vely to a -a .In numerical experiments we have observed that norm-££ mm

reduction with the diagonalizing representative of ~k'~ (~_1 )

(k ~ 0,1, ••••• ) generates a sequence {~}which converges to

a diagonal form. Consequently from some stage in the norm-reducing

process the condition mentioned above was fulfilled, provided the

eigenvalues of A(= A ) were well separated. The above assertion 0

128

Page 130: An eigenvalue algorithm based on norm-reducing transformations

then implies that from that -1 T A_ T can be computffi ,ek'~--k-1 £k,~

in a stable way one shear similarity transformation.

.5.1. First we consider the real optimal norm-reducing trans­

formation by the diagonalizing representative T £m of JJt.,t:m(A). Let

x,,y and z be the real Euclidean parameters of the shears in ~m (A).

We assume that y ~ x (if x > y then the argument is similar). In

this subsection we make use of the notations introduced in (1.2.10)

(1.2.11), (1.2.13) and (1.3.6). As we have seen in section 4 • .3,the

Jacobi-parameters of the diagonalizing representative of ~m(A)are

where

tan2<p yw + 2.:\ vy - 2\z

zy ) c:: -sin<p)

COS(j)

1t 1t --<<p"""-· 4 4

For the estimate of c (T.) we need, as we see from (5.3.22), an r .-vm expression for (pq+rs) 2

• With (5.5.1) and (5.5.2) we find

( \2 [(cos<p + z sin<p)(-sin<p + z cos<p) + y cosm]2 pq+rs; = y sin<p T

[ (y2 + z2- 1) sin2::p + 2zcos2p]2

4l [g - A. - t w(x+y) ]

2

E +vi

We derive from (5 • .3.22) and (5.5.3)

c (T n ) 2 ,.- 5 + 4 [ !l - A - ivv(x+y) )2 r x,m E + w2

Since x+y, i.e. the square of the ~~clidean norm of the (~,m)­

rest.riction of T £m' may be very large, cr(T tm) which is the essen­

tial factor in the given bound for the backward perturbation (see

(5.3.17)), may also be very large. We shall now prove that the

129

Page 131: An eigenvalue algorithm based on norm-reducing transformations

norm-reducing and diagonalizing character of T 0 prevents this dan-. ~

ger on condition that

2 2 ( 2 2 2 2 ) 1 ( )2 ( atm+ am£+ ·..1: ai£+ aim+ a£i+ ami "'"16 a;,[ anun • 5.5.5)

~r"'•m

In terms of the notation of chapter 1 this condition reads

,2 + 112 (.( ... 1 " ,.. +CX+l-' 16 2 \) .

Lemma 5.1. If

then

(1 - 2k2 )v2 "'"E .;;;(1 + 2k

2)

lrl ...:~2 i and

obvious.

2 \) '

As concerns y, since F ~ cxp- y2 ~ 0 (see lemma 1.2.i), we have ~

frr"'" (cx~Y2 .;;;! (ex+~).;;;! k2 v2

Finally D2 ~ (cx!l-f3A.-rv?.;;; (cx2 +~ 2 +/)(r..2 +tl+v2 ).;;; (a:+p) 2 (1+k2)}.

1. .d.. Thus I D I .;;; ( cx+p) ( 1+k2

)'"2 [ v j .;;; ( 1 +k2

) 2

1 v 13 • 0

Lemma 5.2. Let w = -A:x + llY + vz, where x, y and z are the

Euclidean parameters of the class ~£m(A) of unimodular optimal

norm-reducing shears on the pivot-pair (£,m). If

then I wz I .;;; 4 kZiv I • (5.5.10) . 2 2

Proof. If v = o, then A: + /J. + a + ~ = 0 and consequently

m ,em(A) = "";,m • In this case we have z = 0, hence (5.5.10)is satisfied. We now

130

Page 132: An eigenvalue algorithm based on norm-reducing transformations

assume v I 0. Then E ~ (1-2~ )i > 0, as follows from leinma 5.1.

From (1.3.10) and (1.3. 20) follows

I wzl = 1- vD + y(g-E)[ (E-p)2

where p is the negative root of the quartic equation

1.

Since F:;;. 0, certainly \D[/Ip[ < (E-p)(E-2p)""'2, Using this result

in (5.5.11) we find, with lemma 5.1

I wz I < [ vD I I + \r I I < I vI (E-p)(E-2p)2 (E-2p)2

< 4 k2 [vi• 0

Lemma 5.3. Let x, y and z be the Euclidean parameters of ~m(A),

where A is a real matrix. Let T£m be the diagonalizing representa­

tive of ~£m(A) and let T,em = S,em V£m' where S,em is the stable re­

pre"Sentative of m,em(A) and Vtm is an orthogonal shear on the pivot­

pair (,e,m). If <:p is the angle of rotation of V,em and if

A. 2 + ~ 2 + a + p < ~ v2 , 0 < k < !, then

I z Sill(p COBqll < 3k.

Proof, If v = 0 then z = 0 and consequently (5.5.12) holds.

We now assume v I 0. We consider only the case in which y ~ x.Then

the (i,m)-restriction of Ttm is given by (5.5.1) and (5.5.2).Hence

1 z sill(p eo S<:p I < i I z tan2<p I = i I ~ _ + 2 \~/? I· Since y ~ x implies \z[ < y, we find with lemma 5.2

I ~ 4 k2 iv\ + 2 k\vl _ 1 + 2 k lz sin<p cos<:p < 2 ( 1 _ 2k)lv[ - 1 _ 2 k k < 3 k. 0

These lemmas will be used in the proof of

Theorem 5.1. Let o.: + p + A.2 + ~2 < v2 , 0,.;;;; k <

If T,em is the diagonalizing representative of ~£m(A), described in

131

Page 133: An eigenvalue algorithm based on norm-reducing transformations

section 4.3, then the right condition number c (T 0 ) satisfies the r "'m

If v = O, then a+~+ A2 + ~2 = 0; so Tf,m I and(5.5.13)

holds in this case.

We now assume vI 0. Then E ~ (1-2 k2 )v2 > 0, as follows from lem­

ma 5.1.

From (5.5.4), (1.3.10), (1.3.18) and (1.3.19) we derive

c 2 (T ) "' 5 + 4 (u- A - Hx + y)wf r tm E+~

"' 5 + 4 [~-t _ A + D (~-A )D +l(a+§) (E -p) ]2(E + w2 )-1,

p( p - E)

where p is the negative root of the

(p- E) 2 (p2 -F) + D2 (2p -E) = 0.

Thus j:)j/lel "'(E- p )(E- 2p)-i. Consequently

2 (T ) "' 5 + 4 [ I u-.}1 + I u-AIJDI + i(a + s)(E1 - p) J2 im E2 E2 (E - p )(E - 2p)2

~ ~ ~

Since Ill - AI ...;; 22 (A2 + /)2 ...;; 21i'kl vi, we find with lemma 5.1

Finally we estimate the norm of the residual matrix T T - 1 - I. tm £m

This estimate gives a bound for the quantity k(see(5.3.14)),occur-1

ring with c (To ) in the upper bound (5.3.17) for the backward r ,.,m perturbation F (corresponding to the errors generated during the

:-::::; ) computation or T,em A T,em •

If p, q, rand s are the Jaoobi-parameters of Tf,m' then s, -q, ~r

132

Page 134: An eigenvalue algorithm based on norm-reducing transformations

and pare the Jacobi-parameters of T$~ 1 and TtmT~~1 -I= (ps-qr-1)1.

The Jaco bi-parameters of T ~m are computed (by floating-point ari th­

metic) from the Jacobi-parameters of the stable representative

and from those of the orthogonal shear Vtm mentioned

in ( 5. 5 • 1 ) and ( 5 • 5 • 2) • 1 1 1

We assume that the Jacobi-parameters of s. are f~zy-~0 and?(1+e), hill 1

and that those of are c, -s, s and c (being the computed ap-

of and cos~). Then, according to (5.5.1)

(: J where I c: i I ,.; , 1 ,.; i ,.; 9. We find from (5.5.14) that

ps-qr = (1+<::1 ){c 2(1+~)(1 )(1+<::9)+ s2(1+e5)(1+e7)(1+e:8)} +

+ zsc(1 ) { (1 )(1+e::4)(1+e9)-(1+c:6)(1+e:7)(1+e::a)}.

It is now clear that lps-qr-11 will be a moderate multiple of 2-t

on condition that

(i) I z sin~ cos~ I is not

(ii) - 2 - 2 c + s - 1

-t a moderate multiple of 2

(iii) le'S zl is not than. I z siTI(p coscp I • Condition (i) is satisfied if the

longing to the i-th and m-th rows and columns are small relatively

to jaco- a I as we have seen in lemma 5.3. -V-" mm

Condition (ii) is fulfilled if c and s are computed in a stable

mariller ([35], page 276) from the for tan2cp which is de-

scr~bed in formula (5.5.2)

Condition (iii) is satisfied if the

parameter z is calculated with moderate care.

z of the Euclidean

133

Page 135: An eigenvalue algorithm based on norm-reducing transformations

We. conclude from theorem 5.1. and lemma 5.3 that the algorithm

A': ft(T ..e:i, ft(A,T £m)

for the computation of and approximation of T..e:1A T..em' where T..em

is the diagonali representative of ~m(A), can be performed

in a numerically stable way if

71.2 + + a + ~ .,.; If v2

, 0.,.; k.,.; :f. For in that case neither c (T. ) nor the upper bound of

:-::'1 r -vm ~T T"- - Ill /n are large, and consequently the backward pertur-, ..em "'m 2

bation F,estimated in (5.3.17), is small relatively to A.

If c (T. ) or lfTn Tn-1

- Ill /TJ is large (this is only possible r ,vill "'m »ill 2

if the off-diagonal elements ~f A in the ..e-th and m-th rows and

columns are not small relatively to a,e1

- amm)' then the danger of

numerical instability arises. In that case we have to reorganize

the computation in order to guarantee numerical stability. Then,

in fact, the factorization of T..em' described in 5.4.1 has to be

used.

5.5.2. Finally we consider the transformation by the diagonaliz­

ing representative T £m of the class ~m (A) of complex minimizing

shears on the pivot-pair (t,m). T,em is the product of three shears:

T,em = U,em S,em V..em •

The unitar.y shear U,e is used for the pre-treatment (annihilates

( ) , m -1 ) ' , -1

the m,,e -element of U,em A Utm • Let A = Utm A U,em• Stm is

the stable representative of "l'lv" (A') and V n is the uni tarJ shear ,vill -1 ! ,vm

that "diagonalizes" the matrix Stm A S,em'

In this subsection we use the notations introduced in chapter 2 :

(5.5.15)

a :=

y (a. 0 a:. -a". a.) ].,v J.ill ,v]. mJ.

,m

134

Page 136: An eigenvalue algorithm based on norm-reducing transformations

The corresponding functions of A1 = u.-1 A u.

_,vffi hiD (a~ = 0) are indi-

cated by A. 1 )

! I f I ! o,ll,v,a,f3 andy.

with (2.2.13)

' ' I f I ! 2 D := rx ll - y v , E := v I , F :

With these notations we formulate lemma 5.4. The proof of this

lemma is vexy similar to that of lemma 5.1.

Lemma 5.4. If

Let A1 be the pre-treated matrix, thus A.1

0.

(5.5.18)

then 1

( 1-2k2) I v 12 ~ E ~ ( 1 +2k2) I v 12 ' ill r I ~ 22k I vI ' I I 1 2 2 lr ~2k lvl and

1

IDI ~ (1+4k2)2klvl3 •

Analogously to lemma 5.2 we can prove

Lemma 5.5. Let x, y and z be the Euclidean parameters of ~£m(A') I I

and w ll y + v z.

if

then I A 1

2 + ill r 2

+ a + f3 ~ k 21 \) 1

2 ' 0 ~ k ~ i '

Jwl ~ 23/2

k 2 Jvl , jzjy-1 ~~ k, lwzj ~ 4 k 3 jvj and

I w I ( x+y) ~ 8 k 21 vI •

We are now able to formulate the complex analogue of theorem 5.1.

Let p, q, rand s be the Jacobi-parameters of Ttm' where

Tim = Utm Stm Vtm • We take

where (x,y,z) are the Euclidean parameters of ~fm(A').

Since S 0 v. is the "'m .,m

135

Page 137: An eigenvalue algorithm based on norm-reducing transformations

have, according to (4.4.3) and theorem 4.6

A

V X: m

(cos~

I \ ie . \ e SlUrp

-e ) j cos~

where tan2rp = jwj/jv'l , 0 ~ rp ~ n/4 and

w v'/lwv'l if w v' I o, e 0 ~ w v' = o.

Since c: (TX:m) .;;; 5 + 4 lpq + rsl 2 , we first compute lpq + rsl 2 •

From the unitarity of Utm follows that pq + rs only depends on the

Jacobi-parameters of SX:m VX:m' We can easily derive that

I - -, ~ I (coscp + z ei

9sinp)(z cosp - ei

9sincp) + Y i9 . 1

2 pq+rs ~ = y e SlnrpCOSrp

f{;y:2+1 zl2- 1}sin2p + 2Re(z ie) }2 e . co s2p • + 4 Im2(z eiel

4Y2

With the expressions for tan2rp and eie, given above, we find

Re2 {w w-1 ~(x+y) w- ~')} + ~ E + !wl2 y2

Using (5.5.18) and the lemmas 5.4 and 5.5 it is easy to prove

Theorem 5. 2. Let T~m be the diagonalizing representative of

11b" (A). If '"m

I A.j2 + I fll2 + a + ~ .;;; k21 v 12 ' 0 .;;; k ~ t, then c 2 (T ) .;;;. 5 + 1

r ,em

As in the real case we see that, despite the possible largeness of

x+y, the norm-reducing and the diagonalizing character of T.£m pre­

vent. c (Tn) to be very large, provided that the off-diagonal ele-r "'m

ments in the X:-th and m-th columns and rows of A are small rela-

tively to a.££- amm.

136

Page 138: An eigenvalue algorithm based on norm-reducing transformations

, we estimate the norm of the residual matrix T,emT~:- I.

This estimate will give us a bound for the quantity k (s.ee 1

(5.3.14)). With the bounds fork and c (T ) we are able to esti-1 r _em

mate the norm of the backward perturbation corresponding to the

errors made in the computation of an ofT - 1 A Tnm £m ,r,w

(see .3.17)). Let (p, q, r, s) be the Jaoobi-parameters of T_ern, These parameters

will be computed, in floating point from the Jaoobi-

simple, but tedious calculations we find -t be a moderate multiple of 2 if the conditions are

satisfied

(i) I c !2 + Is !2 - 1 and I 12 + I 12

- 1 are both a moderate 1 1 -t

multiple of 2 ;

(ii) ~~~y- 1 not large;

(iii) le s I (~+y) and I s I lzl not 2 2 2

Condition (i) is satisfied if Jacobi-paxameters of U0

and V "'m ,em

are computed in a stable way from (2.2.7b) and (4.4.9) respectively.

If a+~+ lit.l 2 + !11l 2 ..;k2 !vl 2, O..;k..;1z., then the conditions(ii)

and (iii) are fulfilled as follows from the lemmas 5.4 and 5.5.

I I ~~ I w! l·vrl-1• For sin~ cos~ ..;z tan2~

Hence !sin~ cos~! (x+y).,.; (x+y)!v1!-1 < 8 k2 and

!sin~ cos~! !z! .,.; 2{2 k3• if le s I (x+y) and

2 2 I c s I !z! are not essentially than I cos~ si~ I (x+y) and

2 2 ,~, 1"'1 -1 !c2

s2

1 z respectively and moreover z. y is not essentially

larger than !z! y-1, then the conditions (ii) and (iii) are satis­

fied.

We now summarize the results obtained in this subsection.

Let T be the _em representative of ~n (A). If c (Tn ) "'m r "'m

137

Page 139: An eigenvalue algorithm based on norm-reducing transformations

is not and the conditions (i), (ii) and (iii) are satisfied,

then we can -1 '

compute Ttm A Ttm in a numerically stable way with

the algorithm :

Ttm := ft(ft(Utm' Stm)' Vtm)

A1

i= ft('I'£~1 , ft(A, T,em) •

(5.5.19)

(5.5.20)

In (5.5.19) Utm is the unitary shear which A; the

upper triangular representative of~tm(u..e~1 A U,em) and V..em is such

that S,em V,em is the diagonalizing representative of ~m (U£~1 A u1m).

If c (T. ) is large or if the condition )or the condition ) r "'m

is not fulfilled·(this is only possible if the elements

of A in the ..e-th and m-th rows and columns are not small

to a,e,e- amm)' then the factorization of Ttm' mentioned in section

5.4, necessary in order to perform the transfor-

mation in a numerically stable way.

138

Page 140: An eigenvalue algorithm based on norm-reducing transformations

REFERENCES

1. Causey, R.L. 1958, Computing eigenvalues of non-Hermitean ma­

.trices by methods of Jacobi type, J.S.I.A.M, ~,172-181.

2. Dimsdale, B. 1958, The non-convergence of a characteristic

root method, J.S.I.A.M., ~' 23-25.

3. Durand, E. 1960, Solutions n'.lllleriques des equations alge­

briques,Tome II, Masson & Cie, VIII + 447 p.

4. Eberlein, P.J, 1962, A Jacobi-like Method for the Automatic

Computation of Eigenvalues and Eigenvectors of an Ar­

bitrary Matrix, J.S.I.A.M., lQ, 74-88.

5· Eberlein, P.J., 1965, On Measures of non-normality of Ma­

trices, Arn,Wath. Monthly, 72, 995-996.

6, Eberlein, P.J. and Boothroyd, John, 1968, Solution to the

eigenproblem by a norm-reducing Jacobi type method,

Nurn, Math., 11, 1-12.

7, Eberlein, P.J ., 1968, Solution to the Complex Eigenproblem

by a Norm-Reducing Jacobi type Method, unpublished,

8, Forsyth, G.E. and Henrici, P., 1960, The cyclic Jacobi method

for computing the principal values of a complex matrix,

Trans. Am. M.s., 2!• 1-23.

9. Goldstine, H.H., Murray, F,J, and von Neurnan.11., J., 1959, The Jacobi method for real symmetric matrices,J.Ass.

Comp.Mach., ~' 59-96.

10, Goldstine, H. H. and Horwi tz, L.P., 1959, A procedure for the

diagonalization of normal matrices, J.Ass.Comp. Mach.,

§_, 176-195.

11. Greenstadt, J., 1955, A method for finding roots of arbi­

t~:cy matrices, lVfath. Tab. and other aids Comput., 2_,

47-52.

1 39

Page 141: An eigenvalue algorithm based on norm-reducing transformations

12. Greenstadt, J., 1960, ~he determination of the characteris­

tic roots of a matrix by the Ja.cobi II'Tethod, l\IJa.themati-

ca.l Methods for Digital Computers

Wilf), Wiley, New York, 84-91.

ed. Ra.lston and

13. Gregory, R.T., 1953, Computing eigenvalues and eigenvectors

of a. symmetric matrix on the I.L.L.I.A.C ., Math. Tab.

and other aids comput., l• 215-220.

14. Henrici, P., 1962, Bounds for iterates, inverses, spectral

variation and fields of values of non-normal matrices,

Num. ::VIath., 4_, 24-40.

15. Hoff'man, A.J. and Wielandt, H.W., 1953, The variation of the

spectrum of a normal Duke Math.J., 20, 37-39·

16. Jacobi, C.G.J., 1846, Doer ein leichtes Verfahren die in der

Theorie der Secularstorungen vorkommenden Gleichungen

numerisch aufzulosen,Crelle's J., 30, 51-94.

17. Kempen, E.P.M.van, 1966, Cn the convergence of the classical

Jacobi method for real symmetric matrices with non­

distinct eigenva.lues, Num. Math., 2J 11-18.

18. Kempen, II.:?.M.van, 1966, On the quadratic convergence of the

serial Jacobi method, Num.Math., 2, 19-22.

19. Lotkin, M., 1956, Characteristic values of arbitrary matrice~

Quart. Appl. llfa.th., 14_, 267-275.

20. },'[arcus, M. and N'Jinc, H., 1964, A survey of Matrix Theory and

Matrix Inequalities, Allyn and Bacon, Boston, XVIII +

180 p.

21. IvJarcus, M. and Mine, H., 1965, Introduction to Linear Algebra,

The Macmillan Compa~, New York, X + 261 p.

22. Mirsky, L., 1958, On the minimization of Matrix Norms,A.m.Math.

Monthly, ~ 1 06-1 07.

140

Page 142: An eigenvalue algorithm based on norm-reducing transformations

23. Osborne, E.E., On preconditioning of matrices, J. Ass.Comp.

Mach., l• 338-345·

24. Pope, D.A. and Tompkins, c., 1957, Maximizing functions of ro­

tation-experiments concerning speed of diagonalisation

of symmetric matrices using Jacobi 1s method, J.Ass.Comp.

Mach., !, 459-466.

25. Ruhe, A., 1967, On the quadratic convergence of the Jacobi

method for normal matrices, B.I.T., l' 305-313.

26, Ruhe, A., 1968, On the quadratic convergence of a generalisa­

tion of the Jacobi method to arbitrary matrices, B.I.T.,

§_, 21 0-232.

27. Ruhe, A., 1969, The norm of a matrix after a similarity'Trans­

fomation, B.I.T., 2_, 53-58.

28. Rutishauser, H., 1964, Une methode pour le calcul des valeurs

propres des matrices non symetriques, C.R. Acad.Sc.,

t.259, 2758.

29. Rutishauser, H., 1966, The Jacobi-Method for Real Symmetric

Matrices, Num. Math., 2_, 1-10.

30. Smith, R.A., 1967, The condition number of the matrix eigen­

value problem, Num.Math., 1.Q_, 232-240.

31. Voyevodin, V .V., The solution of the complete problem of eigen­

values by a generalised method of rotations, Comp.Meth.

and Progr.III (Comp.Centre of Moscow Univ.) 89-105.

(Russian, transl. by A.Korlaar)

32. Voyevodin, V.V., An extension of the method of Jacobi, Comp.

Meth.and Progr. VIII(Comp.Centre of Moscow Univ.)216-228.

(Russian, transl. by A.Korlaar)

33 •. Wilkinson, J .H., 1962, Note on the quadratic convergence of

the cyclic Jacobi process, Num.Math., i• 296-300.

141

Page 143: An eigenvalue algorithm based on norm-reducing transformations

34.• Wilkinson, J .H., 1963, Rounding Errors in Algebraic Processes,

H.M.s.o., London, VI + 161 p.

35. Wilkinson, J.II., 1965, The Algebraic Eigenvalue Problem,

Clarendon Press, Oxford, XVIII + 662 p.

142

Page 144: An eigenvalue algorithm based on norm-reducing transformations

SAMENVATTING

In dit proefschrift wordt een algorithme beschreven voor de bere­

kening van de eigenwaarden van een matrix. De gebruikte methode is

van het Jacobi-type: in de algorithme wordt iteratief een rij van

matrices geconstrueerd en elke iteratie-stap is een gelijkvormig­

heidstransformatie met een shear (d.i. een matrix die slechts in

een 2 x 2 submatrix verschilt van de eenheidsmatrix).

In de eerste fase van de algorithme wordt de oorspronkelijke ma­

trix met niet-unitaire norm-reducerende shears getransformeerd in

een "bijna normale" matrix.

Als gevolg van de invariantie van de Euclidische norm van een ma­

trix onder unitaire transformatie brengt elke shear in een klasse

van z.g. rij-congruente shears dezelfde norm-reductie teweeg. Zo'n

klasse wordt bepaa.ld door ha.ar Euclidische parameters. De Eucli­

dische norm van de met een shear getransformeerde matrix is een

eenvoudige uitdrukking van deze parameters.Voor unimodulaire shear

transformaties is de functie, die het kwadraat van de Euclidische

norm van de getransformeerde matrix beschrijft in termen van de

Euclidische parameters van de transformatie shear, kwadratisch en

het defini tiegebied er van is een ble"d. van een hyperboloide.

In de hoofdstukken 1 en 2 wordt een methode beschreven ter bereke­

ning van de Euclidische parameters van de klasse van rij-congruen­

te unimodulaire optimaal norm-reducerende shears.Daartoe wordt be­

rekend waar zich op de hyperboloide het infimum bevindt van de bo­

ven vermelde kwadratische functie.

In hoofdstuk 3 wordt het effect van successief toegepaste norm­

reducerende shear transformaties onderzocht. Bij een welgekozen

pivot-strategie zal de rij van matrices, die ontstaat door bij

elke iteratie de norm optimaal te reduceren, convergeren naar nor­

maliteit, d.w.z. convergeren naar de klasse van normale matrices

met hetzelfde ~pectrum als de matrices uit de rij.

143

Page 145: An eigenvalue algorithm based on norm-reducing transformations

In de tweede fase van de algorith~e wordt de door norm-reductie

"bijna normale" matrix A met unitaire shears getransfor-t

meerd in een "bijna diagonaal11 matrix.

Allereerst wordt daartoe, met Jacobi de bijna normale

matrix A zodanig getransformeerd dat het Hermitische gedeelte van

de getransformeerde matrix A bijna 1

is. De matrix A 1

blijkt een bijna blok diagonale structuur te hebben. Van elk dia-

gonaal blok is het Hermitische gedeelte bijna een veelvoud van de

eenheidsmatrix. Daarom kan aangetoond warden dat bij de vervolgens

uit te voeren diagonalisatie van de scheef-Hermitische gedeelten

van deze blokken (met Jacobi rotaties), het bijna diagonaal karak­

ter van het Hermitische gedeelte niet verloren gaat.

De uiteindelijk verkregen matrix

deelte waarvan de norm van boven

heeft een niet diagonaal ge­

wordt door een continue

functie die nul is als A normaal is, het Hermitische gedeelte van

A een diagonaal matrix is en de scheef-Hermitische gedeelten van 1

de blokken van A ook diagonaal 2

Voor reele matrices is de matrix A bijna een kanonieke vorm van 1 .

l\furnaghan mi ts de matrix, indien er een paar complex geconjugeerde

eigenwaarden ~ + i v bestaat, geen andere eigenwaarden heeft met

reeel gedeelte ~·

In hoofdstuk '5 wordt de numerieke stabiliteit van het norm-reduce­

rende proces onderzocht. De resultaten dragen bij tot de

verklaring van de nauwkeurigheid van berekende eigenwaarden in nu­

merieke experimenten met procedures die gebaseerd waren op de in

dit proefschrift beschreven algorithms.

144

Page 146: An eigenvalue algorithm based on norm-reducing transformations

CURRICULUM VITAE

De schrijver van dit proefschrift werd geboren te Zoeterwoude op

16 september 193-1 •

In 1950 behaalde hij het diploma H.B.S.-b aan het Titus Brandsma

College te Oss. Vervolgens studeerde bij wis- en natuurkunde aan

de Rijksuniversitei t te Leiden, waar hij in 1961 het doctoraal .ex­

amen wiskunde met natuurkunde en mechanica behaalde.Tot september

1961 was leraar aan achtereenvolgens het St.Antonius College

te Gouda, de Gemeentelijke M.M.S en H.B.S. voor Meisjes te Leiden

en het R.K. Llfceum St. Bonaventura te Leiden.

Vanaf September 1961 is hij als wetenschappelijk medewerker ver­

bonden aan de Onderafdeling der Wiskunde van de Technische Hoge­

school te Eindhoven.

Page 147: An eigenvalue algorithm based on norm-reducing transformations

STELLINGEN

I Zij D, E en F de volgens (1.2.13) van dit proefschrift ge­

definieerde functies van een reele matrix A en de pivots (t,m). Dan

geldt voor iedere reele .unimodulaire shear T ,em met pivots (,e,m):

D(A; t,m) = D(T - 1 tm A Ttm; t,m) , E(A; t,m) = E(T£~1

A T,em; ,e,m)en

F(A; t,m) = F(T - 1 tm A Ttni; £,m).

II Zij bij een reele matrix A en een pivot-paar (t,m) a• ~· Y•

~. ~' v, e en a gedefinieerd als in hoofdstuk 1 van dit proefschrift.

Zij (x,y,z) de Euclidische parameters van een shear op het pivot­

paar (,e,m).

Zij <p := o:x + ~y + 2yz en w := -A.x + ~y + vz.

Dan is

en

inf S2 (T,e-1

AT,e)= inf (<p+W+a), T£mE t£m m m (x,y,z)Eq:

waarbij ttm' S en t!( zijn gedefinieerd volgens de defini ties 0. 7 ,0.8

en .1.3 van dit proefschrift.

III Als ~. + iv ., a. en h". (j = 1 ,2, ••• , n) de eigenwaarden J J J J

zijn van respectievelijk de matrices A, i(A +A*) en i(A - A*) ~n

tJ. (A) de afwijking van de normali tei t van A is, dan bestaan er per­

mutaties p en q z6 dat

n 2:: (~ • - a ( . ) f .;; tJ.

2 (A) j=1 J p J

IV Zij T een niet singuliere matrix.

Als voor iedere matrix A geldt IIT-1 A TilE ijA[[Edan is Teen veel­

voud van een unitaire matrix.

V Zij f : R ..,. R , f E C2 • n 1

Als grad f Min en slechts een nulpunt X E R heeft dan geldt o n f(x ) extreem impliceert f(x ) globaal extreem.

0 0

Page 148: An eigenvalue algorithm based on norm-reducing transformations

VI Zij voor p > 0. 00

n(log n) 1 + p

1 --:r:tp ) • n

S(p) := Z n=2

Dan bestaat lim S(p) en deze limiet is positief. plO

VII Rutisha.user, Schwarz en Stiefel gaan in hun beschouw:i.ng over

de onnauwkeurigheid in de numerieke oplossing van het stelsel verge­

lijkingen Ax = b ten onrechte uit van wat zij noemen de onnauwkeurig­

heid van de berekening van Ax - b voor veotoren x in de omgeving van

de oplossingsvector xt.

H.Rutisbauser, E,Stiefel, H.R.Schwarz, NUmerik symmetrischer

Matrizen, Stuttgart, 1968.

VIII De door Dekker in de formule van Newton gebruikte benadering

van de multipliciteit van een nulpunt bij eindige arithmetiek,

niet geschikt om de bepaling van een meervoudig nulpunt te versnel­

len. ~.J.Dekker, Newton-Laguerre iteration, Report MR 82,

Mathematisch Centrum, Amsterdam, 1966.

IX In het door Bonset geschetste projektonderwijs wordt onvol­

doende aandacht geschonken aan kennisoverdraoht en vaardigheidstrai­

ning. '· H.Bonset,"Nooit met je rug naar de klas!"

Amsterdam, 1969.

X Het verdient aanbeveling Westerlaken 1s advies inzake het be-

rijden van verkeerspleinen niet op te volgen.

H.M.W.Westerlaken, "Eisen voor rijvaardigheid"

Uitgave ANWB- KNAC - KNMV, Se dxuk.

XI Het is wenselijk de procedure ter beoordeling van weten-

schappelijke medewerkers aan Universiteiten en Hogescholen ex­

pliciet te formuleren.

Eindhoven, 2 december 1969 M.H.C. Paardekooper