A QR-decomposition of block tridiagonal matrices generated by the block Lanczos process

Similar documents
Updating the QR decomposition of block tridiagonal and block Hessenberg matrices generated by block Krylov space methods

Updating the QR decomposition of block tridiagonal and block Hessenberg matrices

Block Krylov Space Solvers: a Survey

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Block Bidiagonal Decomposition and Least Squares Problems

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

arxiv: v1 [hep-lat] 2 May 2012

6.4 Krylov Subspaces and Conjugate Gradients

Arnoldi Methods in SLEPc

Numerical Methods in Matrix Computations

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

The Lanczos and conjugate gradient algorithms

FEM and sparse linear system solving

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

AMS526: Numerical Analysis I (Numerical Linear Algebra)

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Matrices, Moments and Quadrature, cont d

Index. for generalized eigenvalue problem, butterfly form, 211

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Math 504 (Fall 2011) 1. (*) Consider the matrices

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

Krylov Subspace Methods that Are Based on the Minimization of the Residual

On the influence of eigenvalues on Bi-CG residual norms

APPLIED NUMERICAL LINEAR ALGEBRA

BLOCK KRYLOV SPACE METHODS FOR LINEAR SYSTEMS WITH MULTIPLE RIGHT-HAND SIDES: AN INTRODUCTION

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

Solving large scale eigenvalue problems

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES

ANONSINGULAR tridiagonal linear system of the form

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

Iterative methods for Linear System

Krylov Space Solvers

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Generalized MINRES or Generalized LSQR?

Scientific Computing: An Introductory Survey

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Lecture 9 Least Square Problems

Eigenvalues and eigenvectors

Sparse matrix methods in quantum chemistry Post-doctorale cursus quantumchemie Han-sur-Lesse, Belgium

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Applied Numerical Linear Algebra. Lecture 8

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Large-scale eigenvalue problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

11.3 Eigenvalues and Eigenvectors of a Tridiagonal Matrix

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Course Notes: Week 1

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

LARGE SPARSE EIGENVALUE PROBLEMS

Krylov Subspaces. Lab 1. The Arnoldi Iteration

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Iterative Methods for Linear Systems of Equations

Notes on Eigenvalues, Singular Values and QR

Eigenvalue and Eigenvector Problems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Orthonormal Transformations

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Solving Ax = b, an overview. Program

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Lecture 10: Eigenvectors and eigenvalues (Numerical Recipes, Chapter 11)

Orthonormal Transformations and Least Squares

Block Lanczos Tridiagonalization of Complex Symmetric Matrices

THE QR ALGORITHM REVISITED

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Eigenvalue Problems and Singular Value Decomposition

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

S.F. Xu (Department of Mathematics, Peking University, Beijing)

A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

Computing tomographic resolution matrices using Arnoldi s iterative inversion algorithm

Orthogonal Transformations

AN ITERATIVE METHOD WITH ERROR ESTIMATORS

Lecture 5 Singular value decomposition

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.

CG VERSUS MINRES: AN EMPIRICAL COMPARISON

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Orthogonal iteration revisited

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 0

Householder reflectors are matrices of the form. P = I 2ww T, where w is a unit vector (a vector of 2-norm unity)

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

This can be accomplished by left matrix multiplication as follows: I

Bindel, Spring 2016 Numerical Analysis (CS 4220) Notes for

Algorithms that use the Arnoldi Basis

Transcription:

1 A QR-decomposition of block tridiagonal matrices generated by the block Lanczos process Thomas Schmelzer Martin H Gutknecht Oxford University Computing Laboratory Seminar for Applied Mathematics, ETH Zurich Oxford University ETH Zurich Wolfson Building, Parks Road ETH-Zentrum, HG G522 Oxford, OX1 3UQ United Kingdom CH-892 Zurich, Switzerland Phone: +44-1865-273855, Fax: +44-1865-273839 Phone: +41-1-632-3464, Fax: +41-1-632-114 E-mail: thoms@comlaboxacuk E-mail: mhg@sammathethzch Abstract - For MinRes and SymmLQ it is essential to compute a QR decomposition of a tridiagonal coefficient matrix gained in the Lanczos process This QR decomposition is constructed by an update scheme applying in every step a single Givens rotation Using complex Householder reflections we generalize this idea to block tridiagonal matrices that occur in generalizations of MinRes and SymmLQ to block methods for systems with multiple right-hand sides Some implementation details are given, and we compare the method with an algorithm based on Givens rotations used in block QMR Our approach generalizes to the QR decomposition of upper block Hessenberg matrices resulting from the block Arnoldi process and is applicable in block GMRes Keywords block Lanczos process, block Krylov space methods I The symmetric Lanczos algorithm In 1975 Christopher Paige and Michael Saunders [PAI 75] proposed two iterative Krylov subspace methods called MinRes and SymmLQ for solving sparse Hermitian indefinite linear systems Ax = b 1 with A = A H C N N, x C N and b C N An orthonormal basis for the Krylov subspace K n A,r := span { r,ar,,a n 1 r } = span {y,y 1,,y n 1 } Krylov subspaces K m A,y m = 1,2,, νy,a compute, for n = 1,2,, the following: 1 Apply A to y n 1 K n 1 A,y : ỹ n := Ay n 1 2 2 Subtract the projections of ỹ n on the last two basis vectors: ỹ n := ỹ n y n 2 β n 2 if n > 1, 3 α n 1 := y n 1,ỹ n, 4 ỹ n := ỹ n y n 1 α n 1 5 3 Normalize the vector ỹ n K n A,y if ỹ n > : β n 1 := ỹ n, 6 y n := ỹ n /β n 1 7 The method stops when ỹ n = Then n = ν A,y For a basis of K n A,r we start with y = r / r 2 The coefficients of the recurrence relation in the Lanczos process are the entries of a tridiagonal matrix Theorem 1 The vectors { } y,y 1,,y νa,y 1 constructed by this algorithm are orthonormal For every n = 1,2, ν A,y the first n vectors are a basis for K n A,y Moreover, for n < ν A,y where AY n = Y n+1 T n, 8 Y n = y y 1 y n 1, is constructed by the Lanczos process The smallest index n with n = dim K n A,r = dim K n+1 A,r is called the grade of A with respect to r and denoted by ν A,r Algorithm 1 Sym Lanczos algorithm Let a Hermitian matrix A and a vector y with y 2 = 1 be given For constructing a nested set of orthonormal bases {y,y 1,,y m 1 } for the nested T n = α β β α 1 βn 2 β n 2 α n 1 β n 1 R n+1 n 9

and β i > for all i = 1,2,n 1 In MinRes and SymmLQ a QR decomposition of T n is used to construct approximations for the solution of the linear system 1 However, it is enough to apply an update scheme exploiting the tridiagonal structure of T n In every iteration only one Givens rotation has to be constructed II A block Lanczos process The Lanczos process can be generalized for multiple initial vectors, both in the Hermitian [CUL 74], [UND 75], [GOL 77], [LEW 77], [O L 87] and the non- Hermitian case [O L 87], [ALI ], [BAI 99], [FRE 97b] This idea is of interest when a linear system with s right-hand sides has to be solved or when eigenvalues of geometric multiplicity at most s have to be computed an application not treated here Approximating all systems in the same large space may accelerate convergence a lot Usually that space is a direct sum of the corresponding Krylov spaces: B n A,r := s K n i=1 A,r i = block span { r,ar,,a n 1 r } The block Lanczos process creates in exact arithmetic the block vectors which is just a fancy word for a high and skinny matrix to emphasize its character as a list of a few columns y,y 1,,y n 1 Their orthonormal columns are a basis for B n A,v B n A,r = block span {y,y 1,,y n 1 } Deflation is crucial in the underlying Lanczos process but not investigated in this paper Deflation means to delete those columns of y n which are already contained in B n A,r The block vector y i has s i columns, where s s i s i+1, i = 1,2, We denote by t n A,v the dimension of B n A,v, which implies t n A,v = n 1 i= s i Definition 2 The smallest index n with t n A,v = t n+1 A,v is called the block grade of A with respect to v and denoted by ν A,v Algorithm 2 Hermitian Block Lanczos algorithm Let a Hermitian matrix A and an orthonormal block vector y be given For constructing a nested set of orthonormal block bases {y,y 1,,y m 1 } for the nested Krylov subspaces B m A,y m = 1,2, ν A,y compute, for n = 1,2,, the following: 1 Apply A to y n 1 B n 1 A,y : ỹ n := Ay n 1 1 2 Subtract the projections of ỹ n on the last two basis block vectors: ỹ n := ỹ n y n 2 β H n 2 if n > 1, 11 α n 1 := y H n 1ỹn, 12 ỹ n := ỹ n y n 1 α n 1 13 3 QR factorization of ỹ n B n A,y with rankỹ n = s n s n 1 : where: π n y n y n ρ n ρ n ỹ n =: =: yn y n yn y n ρ n ρ n ρ n β n 1 β n 1 π T n, 14 is an s n 1 s n 1 permutation matrix is an N s n block vector with full numerical column rank, which goes into the basis is an N s n 1 s n matrix that will be deflated, is an s n s n upper triangular, nonsingular matrix is an s n s n 1 s n matrix ρ n is an upper triangular s n 1 s n s n 1 s n matrix with ρ n F = Oσ sn+1, where σ sn+1 is the largest singular value of ỹ n smaller or equal to tol The permutations are encapsulated in the block coefficients β i Let P n : block diag π 1,,π n be the permutation matrix that describes all these permutations Note that P T n = P 1 n If tol = we speak

of exact deflation The algorithm uses block permutations as well as deflation It is therefore not immediately apparent that the fundamental Lanczos relationships still hold Theorem 3 With exact deflation the columns of the block vectors y,y 1, constructed by Algorithm 2 are orthonormal For every n = 1,2,, ν A,y the columns of the first n block vectors are an orthonormal basis for B n A,y Moreover, for n < ν A,y where T n = AY n = Y n+1 T n, 15 Y n = y y 1 y n 1, α β H β α 1 C tn+1 tn β H n 2 β n 2 α n 1 β n 1 and α i = α H i for all i = 1,2,n 1 16 III An update scheme for the QR decomposition of block tridiagonal matrices The subdiagonal blocks of T n P n are upper trapezoidal This is used to construct an update scheme for this block tridiagonal matrix Let T n P n = Q n+1 R n be the QR decomposition of T n P n so that Q n+1 is a unitary matrix and R n is an upper triangular matrix with full column rank The block tridiagonal matrix T n P n is α π 1 β H π 2 β π 1 α 1 π 2 T n P n = β 1 π 2 β H n 2 π n αn 1 π n The t n+1 t n matrix R n has the form where: α i β i γ i R n = α β γ α 1 β1 γn 3 sn t n βn 2 α n 1 is an s i s i upper triangular block with full rank, is an s i s i+1 block and is an s i s i+2 lower trapezoidal block 18 We determine the unitary matrix Q n+1 in its factored form With Q 1 = I s 19 we introduce the recurrence relation where U n := Q n+1 := Qn tn s n sn t n I sn U n 2 I tn 1 tn 1 s n 1+s n 21 sn 1+s n t n 1 Û n Here Ûn is a unitary s n 1 + s n s n 1 + s n matrix Given a sequence of unitary transformations Û 1,Ûn it is possible to compute Q n+1 with a simple scheme If the matrix Q n has the form Q n = q q 1 q n 2 q n 1 where q i C tn si and where: Û n = Ûn,u Û n,d Ûn,u,l Û = n,u,r Û n,d,l Û n,d,r 22 where: α i π i+1 β i π i+1 is an s i s i block and β n 1 π n is an s i+1 s i upper trapezoidal block, β i π i+1 = ρ i+1 ρ i+1 17 Û n,u Û n,d Û n,u,l Û n,u,r Û n,d,l Û n,d,r is an s i 1 s i 1 + s i matrix, is an s i s i 1 + s i matrix, is an s i 1 s i 1 matrix, is an s i 1 s i matrix, is an s i s i 1 matrix, is an s i s i matrix,

Apply the upper part of Ûi: 5 Q n+1 1 : t i,t i 1 + 1 : t i+1 1 15 = Q n+1 1 : t i,t i 1 + 1 : t i Ûi,u Apply the lower part of Ûi: Q n+1 t i + 1 : t i+1,t i 1 + 1 : t i+1 = Ûi,d 24 2 5 1 15 2 a The block tridiagonal structure of a matrix T n P n The matrix is not symmetric, but all subdiagonal blocks are upper trapezoidal The multiplication of T n P n by Q H n annihilates all subdiagonal elements except those below the diagonal of the last s n 1 columns, or if we regard T n P n as a matrix with n block columns, it does not annihilate the subdiagonal elements in the last block column, ie Q H n I sn T n P n 5 1 15 = α β γ α 1 β1 γn 3 α n 2 βn 2 µ n 25 2 β n 1 π n then 5 1 15 2 b The corresponding block structure of the matrix R n qn 2 q Q n+1 = n 1 Û n,u,l q n 1 Û n,u,r Û n,d,l Û n,d,r 23 In particular the matrix Q n+1 has a trapezoidal structure The update scheme 23 yields the following algorithm, which we describe using Matlab notation: Algorithm 3 Construction of Q n+1 Let Û1,Ûn be the constructed sequence of unitary matrices where Ûi is of order s i 1 +s i and let Q n+1 = I tn+1 For the recursive definition of Q n+1, for i = 1,,n: For the entries in the last block column, respectively the entries in the last s n 1 columns we have γ n 3 β n 2 µ n ÛH n 2 = Isn 3 Û H n 1 I sn 1 sn 3 s n 1 β H n 2π n α n 1 π n 26 For annihilating all subdiagonal elements we have to construct a unitary matrix U n such that µ n β n 1 π n = Ûn αn 1 sn s n 1, 27 where α n 1 is an upper triangular nonsingular block vector

Algorithm 4 An update scheme for the QR decomposition By applying a sequence of unitary matrices Û 1,,Ûm on the matrix T m P m of 17 the upper triangular matrix R m of 18 is recursively constructed as follows For n = 1,,m: 1 Let α n 1 := α n 1 π n, and β n 2 := β H n 2π n if n > 1 If n > 2, apply U H n 2 to the new last column of T n : γn 3 sn 3 s := β ÛH n 1 n 2 ; n 2 β n 2 if n > 1, apply U H n 1 to the last column of U H n 2T n : βn 2 α n 1 := ÛH n 1 βn 2 α n 1 2 Let µ n := α n 1 and compute ÛH n according to 27 3 Compute Q n+1 according to Algorithm 3 There are various possible ways to construct such a matrix Ûn Freund and Malhotra [FRE 97a] apply Givens rotations since a problem with a single righthand side should be a special case of the general block problem In the case of one right-hand side it is enough to apply a single Givens rotation So the rationale behind this approach was to generalize this special case [FRE 4] For our goals the most efficient way is to use a product of complex Householder reflections IV Complex Householder reflections Let y = T y 1 y n be a complex vector The goal is to construct a unitary matrix U C n n such that Uy = αe 1 where α C As U is unitary we note that α = y 2 for any nonzero v C n the matrix H v = I n 2 v vh v,v = I n + β v v H, 28 where β = 2/ v,v R, is called a Householder reflection The matrix H v describes a reflection at the complimentary subspace orthogonal to v We note that H v is Hermitian and unitary, ie H H v = H v and H v H H v = I n The vector y is mapped to H v y = y + βv v,y 29 Keeping in mind our goal we demand This yields H v y = αe 1 y αe 1 = β v,y v 3 In particular v span {y αe 1 } As H v = H λv for all λ C {} we can choose v = y αe 1 without loss of generality By 3 this choice implies v,y = β 1 R Let y 1 = y 1 e iθ and α = y 2 e iθα Then v,y = y αe 1,y = y 2 2 y 2 e iθα e iθ y 1 So either α = + y 2 e iθ or α = y 2 e iθ The second choice is better, otherwise cancellation effects may occur in the first component of v Finally we note that β 1 = v,y = y 2 y 2 + y 1 Algorithm 5 Implicit construction of H v Let y = y 1 y n T be a complex vector A Householder reflection H v of order n with H v y = αe 1 is constructed as follows Let y 1 = y 1 e iθ Compute α and β: α = y 2 e iθ, β = Compute the vector v: 1 y 2 y 2 + y 1 31 v = y αe 1 32 It is not necessary to compute the actual matrix H v It is much more economical and accurate to store the vector v and the coefficient β and apply the identity 29 Parlett [PAR 71] presents a thorough discussion on the choice of the sign of α when computing Householder reflectors Lehoucq [LEH 96] compares different variants for the choice of the vector v and the corresponding coefficient β He specifically compares the different computations of an elementary unitary matrix in EISPACK, LINPACK, NAG and LAPACK The above scheme is due to Wilkinson [WIL 65, pp 49-5] V QR decomposition of a lower banded matrix The idea is to use the trapezoidal structure of the matrix ν n := β n 1 π n in 27 Recall that µ n is s n s n,

while ν n is s n s n 1 For example, if s n 1 = 5 and s n = 4, µn ν n = We determine s n 1 Householder reflections H 1,n,,H sn 1,n such that αn 1 sn s n 1 = H sn 1,n H 1,n µn ν n, 33 where α n 1 is an upper triangular matrix In particular Û n = H 1,n H sn 1,n 34 Assume that reflections H 1,n,H 2,n have been computed such that H 2,n H 1,n µn ν n = The highlighted section of the third column vector determines the next Householder reflection In step i this vector has the size l i,n = s n 1 i + 1 }{{} + min i,s n }{{}, size of upper part size of lower part and the last entry is in row In this example we have e i,n = l i,n + i 1 l 3,n = 5 3 + 1 + 3 = 6, e i,n = 6 + 3 1 = 8 Hence the Householder reflection is given by H i,n = diag I i 1,Ĥi,n,I sn mini,s n, where Ĥi,n is a Householder reflection in the sense of our original definition 28: a reflection at a hyperplane but in a space of dimension l i,n only When we apply this reflection we only compute those entries which are not invariant In this example the first two and the last row would be not influenced at all All we have to do is to apply the reflection Ĥi,n on the submatrix whose left column is exactly given by the vector generating Ĥi,n Here the submatrix is highlighted: α This submatrix is updated and we proceed with the construction of the next reflection Algorithm 6 Implicit construction of Ûn Let µ n a s n 1 s n 1 and ν n an upper trapezoidal s n s n 1 block In a implicit way we construct s n 1 Householder reflections such that 33 holds Let For i = 1,,s n 1 : Compute l i,n and e i,n : M = µn ν n l i,n = s n 1 i+1+min i,s n, e i,n = l i,n +i 1 35 Create implicitly by Algorithm 5 the Householder reflection Ĥi,n, that is compute β and the vector v, using the vector y i,n = Mi : e i,n,i 36 Apply Ĥi,n to the corresponding submatrix of M: Mi : e i,n,i + 1 : s n 1 = Mi : e i,n,i + 1 : s n 1 +βv v H Mi : e i,n,i + 1 : s n 1 37 VI Householder reflections vs Givens rotations Given an upper trapezoidal matrix M with s 1 = s 2 = w, Householder reflections, applied as described in Section IV, are an efficient way to construct the QR decomposition of M An alternative are Givens rotations: instead of the Householder reflection of Algorithm 5 a set of w 1 Givens rotations is applied in

norm norm x Frobenius norm of QR M 1 15 4 2 2 4 6 8 1 3 x 1 15 Frobenius norm of QH Q I 2 1 2 4 6 8 1 Fig 1 Experiment 1: Accuracy of the QR decomposition of 1 random 1 5 upper trapezoidal matrices M The solid line represents results gained by using Householder reflections The dashed line corresponds to Givens rotations cpu time in seconds for 1 testmatrices 8 7 6 5 4 3 2 1 5 1 15 2 25 3 width w Fig 2 Experiment 2: Computation time for the QR decomposition of 1 random 2w w upper trapezoidal matrices M The solid line represents results gained by using Householder reflections The dashed line corresponds to Givens rotations In both cases we have updated the remaining columns simultaneously an explicit loop Therefore w 1 roots have to be computed In Algorithm 6 the block-wise update in 37 by a single Householder reflection applied to a whole submatrix of M must be replaced by applying the w 1 Givens rotations to suitable rows of this submatrix of M In a first experiment we compare the accuracy of both approaches Experiment 1 We apply both approaches for the QR decomposition Householder reflections and Givens rotations to a set of 1 random 1 5 upper trapezoidal matrices M The results are shown in Fig 1 The accuracy of both methods turns out to be on the same level: except in a few cases the Frobenius norms of M QR and of Q H Q I are for both methods of the same magnitude The norm of Q H Q I is typically slightly smaller if Householder reflections are used For the explicit computation of Q we apply the unitary transformations to I 1 1 as in 34 Next we compare the computing time of the two approaches The time-critical second step involves the application of the implicitly constructed unitary matrix to the remaining columns The update in 37, where a Householder reflection is applied to all these columns at once, can be done by BLAS2 operations, whereas with Givens rotations we have to apply the corresponding BLAS1 operations in a loop from the bottom row to the top row of the remaining submatrix of M Experiment 2 In order to compare the speed of the two approaches we measure the cpu time for constructing the QR decomposition of 1 random matrices M of size s 1 = s 2 = w, using Householder reflections or Givens rotations, respectively See Fig 2 For a ma- trix M of width 3 the Givens rotations turn out to be about 7% slower, but for small matrices the efficiency difference is small The block Lanczos process introduced in this paper allows a block-wise construction of T n as opposed to the column-wise construction of Ruhe [RUH 79] or Freund and Malhotra [FRE 97a] The block-wise construction of T n allows us to update its QR decomposition block-wise Hence we can profit from the possibility to update several columns at once in 37 Recall that in this application M is a submatrix of T n In contrast, in the implementation of block QMR by Freund and Malhotra there is only a routine for rotating a single vector by a set of Givens rotations, since in block QMR the matrix T n is generated column-wise, and its QR decomposition is updated whenever a new column of T n becomes available In particular, in the first iteration a vector of length s + 1 is obtained as first column of T n A set of s Givens rotations is constructed to annihilate all except the first entry In the next iteration a second vector of length s + 2 appears as second column, etc At every stage we have to apply the previously constructed rotations or reflections Hence, instead of applying a reflection at once on a matrix of width k < s it is necessary to apply a reflection or a set of s rotations k times on a column vector This is equivalent in theory However, in practice the latter is far slower In a third experiment we compare the block-wise update of the QR decomposition using Householder reflections with the column-wise update of the QR decomposition using Givens rotations Experiment 3 We measure the cpu time for constructing the QR decomposition of 1 random matrices M of size s 1 = s 2 = w using Householder re-

cpu time in seconds for 1 testmatrices 25 2 15 1 5 2 4 6 8 1 12 14 16 width w Fig 3 Experiment 3: Computation time for the QR decomposition of 1 random 2w w upper trapezoidal matrices M The solid line represents results gained by using Householder reflections and block-wise updates The dashed line corresponds to using Givens rotations and column-wise updates flections with simultaneous update of the remaining columns as in 37 on the one hand, and using Givens rotations with column-wise update as described above on the other hand See Fig 3 VII Conclusions and generalizations The symmetric block Lanczos process produces a growing banded matrix T n, which consists of a symmetric block tridiagonal matrix T n, extended by some additional rows, and a permutation matrix P n In a forthcoming paper we will discuss the need for those permutations and deflations in the symmetric block Lanczos process and give details and results for the block MinRes and block SymmLQ algorithms, which require to compute the full QR decomposition T n P n = Q n R n The standard approach for this computation has been an update algorithm based on Givens rotations, a generalization of the well known update algorithm for tridiagonal matrices It was recommended to compute both T n and R n column by column We promote instead a block-wise construction of T n and a block-wise update algorithm based on Householder reflections for the QR decomposition It turns out that our QR decomposition is equally accurate as the one based on Givens rotations and that even on a serial computer it is much faster than column-wise updates with Givens rotations At least formally our approach can be generalized quickly from the symmetric block Lanczos to the unsymmetric block Lanczos and the block Arnoldi processes, and from the QR decomposition of banded symmetric block tridiagonal matrices to the one of banded unsymmetric block tridiagonal matrices or block Hessenberg matrices as they come up in block QMR and block GMRes, respectively REFERENCES [ALI ] Aliaga J I, Boley D L, Freund R W, Hernandez V, A Lanczos-type method for multiple starting vectors, Math Comp, vol 69, p 1577 161, 2 [BAI 99] Bai Z, Day D, Ye Q, ABLE: an adaptive block Lanczos method for non-hermitian eigenvalue problems, SIAM J Matrix Anal Appl, vol 2, p 16 182 electronic, 1999, Sparse and structured matrices and their applications Coeur d Alene, ID, 1996 [CUL 74] Cullum J, Donath W E, A block generalization of the symmetric s-step Lanczos algorithm, Rapport no 4845, IBM TJ Watson Research Center, may 1974 [FRE 97a] Freund R W, Malhotra M, A block QMR algorithm for non-hermitian linear systems with multiple right-hand sides, Linear Algebra and Its Applications, vol 254, p 119 157, 1997 [FRE 97b] Freund R W, Malhotra M, A block QMR algorithm for non-hermitian linear systems with multiple right-hand sides, Linear Algebra Appl, vol 254, p 119 157, 1997 [FRE 4] Freund R W, QR Zerlegung im Lanczos Prozess, private note, 24 [GOL 77] Golub G H, Underwood R, The block Lanczos method for computing eigenvalues, Mathematical Software, III Proc Sympos, Math Res Center, Univ Wisconsin, Madison, Wis, 1977, p 361 377 Publ Math Res Center, no 39, Academic Press, New York, 1977 [LEH 96] Lehoucq R B, The computations of elementary unitary matrices, ACM Trans Math Software, vol 22, p 393-4, 1996 [LEW 77] Lewis J, Algorithms for Sparse Matrix Computations, PhD thesis, Stanford University, Standford, CA, 1977 [O L 87] O Leary D P, Parallel implementation of the block conjugate gradient algorithm, Parallel Computing, vol 5, p 127-139, 1987 [PAI 75] Paige C C, Saunders M A, Solution of Sparse Indefinite Systems of Linear Equations, SIAM J Numer Anal, vol 12, p 617 629, 1975 [PAR 71] Parlett B N, Analysis of algorithms for reflectors in bisectors, SIAM Review, vol 13, p 197 28, 1971 [RUH 79] Ruhe A, Implementation aspects of band Lanczos algorithms for computation of eigenvalues of large sparse symmetric matrices, Math Comp, vol 33, p 68 687, 1979 [UND 75] Underwood R, An iterative block Lanczos method for the solution of large sparse symmetric eigenproblems, PhD thesis, Stanford University, Stanford, CA, 1975 [WIL 65] Wilkinson J H, The Algebraic Eigenvalue Problem, Oxford University Press, 1965