ABSTRACT OF DISSERTATION. Wei Zhang. The Graduate School. University of Kentucky

Size: px
Start display at page:

Download "ABSTRACT OF DISSERTATION. Wei Zhang. The Graduate School. University of Kentucky"

Transcription

1 ABSTRACT OF DISSERTATION Wei Zhang The Graduate School University of Kentucky 2007

2 GMRES ON A TRIDIAGONAL TOEPLITZ LINEAR SYSTEM ABSTRACT OF DISSERTATION A dissertation submitted in partial fulfillment of the requirements of the degree of Doctor of Philosophy in the College of Arts and Sciences at the University of Kentucky By Wei Zhang Lexington, Kentucky Director: Dr. Ren-Cang Li, Dr. Qiang Ye, Department of Mathematics Lexington, Kentucky 2007 Copyright c Wei Zhang 2007

3 ABSTRACT OF DISSERTATION GMRES ON A TRIDIAGONAL TOEPLITZ LINEAR SYSTEM The Generalized Minimal Residual method (GMRES) is often used to solve a nonsymmetric linear system Ax = b. But its convergence analysis is a rather difficult task in general. A commonly used approach is to diagonalize A = XΛX and then separate the study of GMRES convergence behavior into optimizing the condition number of X and a polynomial minimization problem over A s spectrum. This artificial separation could greatly overestimate GMRES residuals and likely yields error bounds that are too far from the actual ones. On the other hand, considering the effects of both A s spectrum and the conditioning of X at the same time poses a difficult challenge, perhaps impossible to deal with in general but only possible for certain particular linear systems. This thesis will do so for a (nonsymmetric) tridiagonal Toeplitz system. Sharp error bounds on and sometimes exact expressions for residuals are obtained. These expressions and/or bounds are in terms of the three parameters that define A and Chebyshev polynomials of the first kind or the second kind. KEYWORDS: GMRES, rate of convergence, tridiagonal Toeplitz matrix, linear system, Chebyshev Polynomials Wei Zhang 23 July 2007

4 GMRES ON A TRIDIAGONAL TOEPLITZ LINEAR SYSTEM By Wei Zhang Dr. Ren-Cang Li, Dr. Qiang Ye Director of Dissertation Dr. Serge Ochanine Director of Graduate Studies 23 July 2007

5 RULES FOR THE USE OF DISSERTATIONS Unpublished dissertations submitted for the Master s and Doctor s degrees and deposited in the University of Kentucky Library are as a rule open for inspection, but are to be used only with due regard to the rights of the authors. Bibliographical references may be noted, but quotations or summaries of parts may be published only with the permission of the author, and with the usual scholarly acknowledgments. Extensive copying or publication of the dissertation in whole or in part requires also the consent of the Dean of the Graduate School of the University of Kentucky.

6 DISSERTATION Wei Zhang The Graduate School University of Kentucky 2007

7 GMRES ON A TRIDIAGONAL TOEPLITZ LINEAR SYSTEM DISSERTATION A dissertation submitted in partial fulfillment of the requirements of the degree of Doctor of Philosophy in the College of Arts and Sciences at the University of Kentucky By Wei Zhang Lexington, Kentucky Director: Dr. Ren-Cang Li, Dr. Qiang Ye, Department of Mathematics Lexington, Kentucky 2007 Copyright c Wei Zhang 2007

8 ACKNOWLEDGMENTS I would like to express my gratitude to all those who gave me the possibility to complete this thesis. I would like to thank Dr. Zhongwei Shen, Dr. David Johnson, Dr. Robert Molzon, Dr. Russell Brown, Dr. Peter Hislop and Dr. Serge Ochanine from Department of Mathematics of University of Kentucky, and Dr. Yuming Zhang from Department of Electrical Engineering of University of Kentucky, for their continuous support and encouragement since I joined the mathematics program. I am deeply indebted to my advisor Dr. Ren-Cang Li and Dr. Qiang Ye from Department of Mathematics of University of Kentucky, who helped and encouraged me in all the time of research for and writing of this thesis and I would like to give my special thanks to my parents whose support enabled me to complete this work. iii

9 Contents Acknowledgments List of Tables List of Figures iii vi vii Chapter Introduction. Introduction Objective Notation Chapter 2 Preliminary 6 2. Projection Method Basic Idea Properties Krylov Subspace GMRES method Arnoldi Algorithm GMRES Algorithm Practical Implementation of GMRES Chebyshev Polynomials Chebyshev Polynomials of the First Kind Chebyshev Polynomials of the Second Kind Chapter 3 Convergence Analysis Using Chebyshev Polynomials of the First Kind Residual Formulation for a Diagonalizable Linear System iv

10 3.2 Residual Reformulation Using Chebyshev Polynomials of the First Kind Estimation of Residual in General Case Residual with General Right-hand Sides The Second Part of the Proof Numerical Examples Special Right-hand Sides: b = e, b = e N Right-hand Sides b = e Numerical Examples for b = e Right-hand Sides b = e N Right-hand Sides b () e + b (N) e N Worst Convergence Speed Chapter 4 Convergence Analysis Using Chebyshev Polynomials of the Second Kind Residual Reformulation Using Chebyshev Polynomials of the Second Kind Estimation of Residual in General Case Residual with General Right-hand Sides Numerical Examples Exact Residual for Special Right-hand Sides: b = e, b = e N The Structure of Ml Chapter 5 Conclusions 84 Bibliography 89 Vita 94 v

11 List of Tables 3.3. Parameters Setting vi

12 List of Figures 2.3. AV m = V m H m + h m+,m v m+ e T m AV j = V j H j The structure of M l with N = 8 for L =, 2, 3, The structure of M l with N = 8 for L = 5, 6, 7, Computation of M (i,j) GMRES residuals for random b with τ = 0.8, and the upper bounds by Theorem GMRES residuals for random b with τ =.2, and the upper bounds by Theorem GMRES residuals for random b with τ =, and the upper bounds by Theorem GMRES residuals for random b with τ = 0.8, and the upper bounds by Theorem GMRES residuals for random b with τ =.2, and the upper bounds by Theorem GMRES residuals for random b with τ =, and the upper bounds by Theorem GMRES residuals for b = e, τ = 0.8, and the bounds by Theorem GMRES residuals for b = e, τ =.2, and the bounds by Theorem GMRES residuals for b = e, τ =, and the bounds by Theorem GMRES residuals for b = e + e N, τ = 0.8, and the bounds by Theorem GMRES residuals for b = e + e N, τ =.2, and the bounds by Theorem GMRES residuals for b = e + e N, τ =, and the bounds by Theorem vii

13 4.2. GMRES residuals for random b with τ = 0.8, and the upper bounds by Theorem 3.3. and Theorem GMRES residuals for random b with τ =.2, and the upper bounds by Theorem 3.3. and Theorem GMRES residuals for random b with τ =, and the upper bounds by Theorem 3.3. and Theorem Upper bound ratios between the bounds obtained from Theorem 4.2. and those from Theorem viii

14 Chapter Introduction. Introduction Iteration methods are often used to solve large sparse systems of linear equations. The Generalized Minimal Residual (GMRES) method [32, 4] is such an algorithm and is often used for solving a non-symmetric linear system Ax = b, (..) where A is an N N nonsingular matrix, and b is a vector with dimension N. The basic idea is to seek approximate solutions, which minimize the residual norm, within the Krylov subspaces. Specifically, the kth approximation, x k, is sought so that the kth residual, r k = b Ax k, satisfies [32] (without loss of generality, we take initially x 0 = 0 and thus r 0 = b) r k 2 = min y K k b Ay 2, where the kth Krylov subspace K k K k (A, b) of A on b is defined as K k K k (A, b) def = span{b, Ab,..., A k b}, (..2) and generic norm 2 is the usual l 2 norm of a vector or the spectral norm of a matrix. According to R. W. Freund, N. M. Nachtigal [2] and M. Embree [0], the residual norms by other algorithms, such as QMR [2, 3] and BiCGSSTAB [39, 20], are somehow

15 related to the GMRES residual norm. Hence, understanding convergence for GMRES helps us to study convergence of other algorithms. Roughly speaking, there are three kinds of convergence bounds for GMRES:. The bound based on eigenvalues with the eigenvector condition number [8, 32]; 2. The bound based on the field of values [6, 7]; 3. The bound based on pseudospectra [37, 36]. In this thesis we are most interested in the first kind. It starts by diagonalizing A = XΛX and then separating the study of GMRES convergence behavior into optimizing the condition number of X and a polynomial minimization problem over A s spectrum. This artificial separation could greatly overestimate GMRES residuals and likely yields error bounds that are too far from the actual ones. On the other hand, considering the effects of both A s spectrum and the conditioning of X at the same time poses a difficult challenge, perhaps impossible to deal with in general but possible for certain particular linear systems..2 Objective This thesis is concerned with the convergence analysis of GMRES on a linear system Ax = b whose coefficient matrix A is a (nonsymmetric) tridiagonal Toeplitz coefficient matrix λ µ ν A = µ, ν λ where λ, µ, ν are assumed nonzero and possibly complex. Linear systems as such have been studied quite extensively in the past. For the nonsymmetric case, i.e., µ ν as we 2

16 are interested in here, most up-to-date and detailed studies are due to Liesen and Strakoš [26] and Ernst []. Motivated to better understand the convergence behavior of GMRES on a convectiondiffusion model problem [27], Liesen and Strakoš and Ernst established various bounds on residual ratios. Most results in Liesen and Strakoš [26] are of a qualitative nature, intended to explain GMRES convergence behaviors for such linear systems. In particular, Liesen and Strakoš showed that GMRES for tiny µ behaves much like GMRES after setting µ to 0. Instead of eigenvalue information, Ernst used the field of values to assess the convergence rate. Our object here is to analyze the kth residual for the GMRES on tridiagonal Toeplitz A directly and arrive at simple quantitative results. The remainder of this thesis is organized as follows. Chapter 2 reviews necessary material about GMRES method and Chebyshev polynomials. Section 2. reviews projection methods and some of their properties. Section 2.2 introduces Krylov subspaces. Based on projection methods and Krylov subspaces, Section 2.3 explains the Arnoldi process and GMRES algorithm. Section 2.4 introduces Chebyshev polynomials of the first kind and the second kind. Chapter 3 covers the rate of convergence analysis of GMRES Method on a tridiagonal Toeplitz linear system by applying the Chebyshev polynomial of the first kind. Section 3. gives the calculation of the kth residual of GMRES method. Section 3.2 analyzes the norm of the residual based on rectangular Vandermonde matrices and the Chebyshev polynomial of the first kind. Section 3.3 follows the calculation of the residual in Section 3.2, and shows an estimation of the upper bound of the kth residual in a general case, which is given by Theorem Section 3.3 finishes the proof of Theorem 3.3. by analyzing the decomposition of the residual. Some numerical examples are also given 3

17 in Section 3.3. Section 3.4 lists the estimation result with some special right-hand sides: e, e N, or b () e + b (N) e N. The estimation comes from Theorem 3.3. and gives a tight upper bound. Section 3.5 estimates what the worst convergence speed could be. Chapter 4 applies Chebyshev Polynomials of the second kind to estimate the kth residual, r k, of GMRES method on a tridiagonal Teoplitz linear system. Section 4. calculates the residual by applying rectangular Vandermonde matrices and the Chebyshev polynomial of the second kind. The computation is similar to those in Chapter 3, and similar bounds are obtained. Section 4.2 follows the computation in the previous section, and gives an estimation of residuals of GMRES with a general right-hand side. The residuals with special right-hand side: b = e or b = e N are calculated exactly in Section 4.3. Some of the complicated computations, needed in Section 4.2, are presented in Section 4.4. Chapter 5 presents our concluding remarks..3 Notation Throughout this thesis, K n m is the set of all n m matrices with entries in K, where K is C (the set of complex numbers) or R (the set of real numbers), K n = K n, and K = K. I n (or simply I if its dimension is clear from the context) is the n n identity matrix, and e j is its jth column. The superscript takes conjugate transpose while T takes transpose only. The smallest singular value of X is denoted by σ min (X). We shall also adopt MATLAB-like convention to access the entries of vectors and matrices. The set of integers from i to j inclusive is i : j. For vector u and matrix X, u (j) is u s jth entry, X (i,j) is X s (i, j)th entry, diag(u) is the diagonal matrix with (diag(u)) (j,j) = u (j) ; X s submatrices X (k:l,i:j), X (k:l,:), and X (:,i:j) consists of intersections of row k to row l and column i to column j, row k to row l and all columns, and all rows 4

18 and column i to column j, respectively. Finally p ( p ) is the l p norm of a vector or the l p operator norm of a matrix, defined as ( ) /p u p = u (j) p j, X p = max Xu p. u p= α be the largest integer that is no bigger than α, and α the smallest integer that is no less than α. Throughout this thesis, exact arithmetic is assumed, A is N-by-N, and k is GMRES iteration index. Since in exact arithmetic GMRES computes the exact solution in at most N steps, r N = 0. For this reason, we restrict k < N at all times. This restriction is needed to interpret our later results concerning (worst) asymptotic speed in terms of certain limits of r k /k as k. Copyright c Wei Zhang

19 Chapter 2 Preliminary In this chapter, we briefly present the projection method, the Krylov subspace method, the GMRES method, and Chebyshev polynomials. The projection methods and some of their properties are presented in Section 2.. In Section 2.2, the Krylov subspace method is introduced as a special case of the Projection method. Section 2.3 introduces the GMRES method, including Arnoldi process and the general GMRES algorithm. Finally Section 2.4 presents Chebyshev polynomials of the first kind and the second kind. 2. Projection Method 2.. Basic Idea A projection method [2, 5] is to find an approximation to the solution of a linear system from a subspace, and is widely used for solving large linear systems. Without loss of generality, we take initially x 0 = 0. A general projection method for solving the linear system Ax = b is a method which seeks an approximate solution x m from a subspace J m of dimension m by imposing the Petrov-Galekin condition: b Ax m L m, 6

20 i.e., b Ax m, w = 0, w L m, (2..) where L m is another subspace of dimension m, and, is the inner product. The subspace J m is the search subspace, which contains the approximate solution x m. L m is called the subspace of constraints, i.e. the constraints in the Petrov-Galekin condition. J m and L m can be the same subspace, which results in an orthogonal projection method. If they are different, it is called an oblique projection method. Let J m have a basis {v, v 2,..., v m }, and V = [v, v 2,..., v m ] be an n m matrix. Let L m have a basis {w, w 2,..., w m }, and W = [w, w 2,..., w m ] be an n m matrix. Then let x m = V y, where y is a vector with dimension m. According to (2..), we have W T AV y = W T b. (2..2) If W T AV is nonsingular, then x m can be written as x m = V y = V (W T AV ) W T b. (2..3) The projection method can be presented by Algorithm 2.. [29]. Algorithm 2.. Projection Method : repeat 2: Select a pair of subspace J m and L m ; 3: Choose bases V = [v, v 2,..., v m ] and W = [w, w 2,..., w m ] for J m and L m ; 4: r := b Ax; 5: y := (W T AV ) W T r; 6: x := x + V y; 7: until convergence 2..2 Properties Obviously, Algorithm 2.. works only if W T AV is nonsingular. A special case, where W T AV is nonsingular, is presented by [29, Proposition 5.]. 7

21 Proposition 2... If A is nonsingular and L m = AJ m, then the matrix W T AV is nonsingular for any basis matrices V and W of J m and L m, respectively. Proof: Since L m = AJ m, then we have W = AV G, where G is a nonsingular m m matrix. Hence W T AV = G T (AV ) T AV. Since A is a nonsingular N N matrix, and V is a basis of J m with dimension N m, then N m matrix AV is of full column rank and (AV ) T AV is nonsingular.. [29, Proposition 5.] also presents another particular case as follows, Proposition If A is positive definite and L m = J m, then the matrix W T AV is nonsingular for any basis matrices V and W of J m and L m, respectively. Proof: Since L m = J m, then W = V G, where G is a nonsingular m m matrix. Hence W T AV = G T V T AV. Since A is positive definite, V T AV is also positive definite. Hence W T AV = G T V T AV is nonsingular.. According to the above propositions, L m is chosen as J m or AJ m very often. Let L m = AJ m. The projection method minimizes 2-norm of the residual r = b Ax [29, Proposition 5.3]. 8

22 Proposition Let A be an arbitrary square matrix, L m = AJ m. Given an initial solution x 0 = 0, a vector x is the result of an projection method onto J m, orthogonally to L m if and only if it minimizes the 2-norm of the residual vector b A x over x J m, i.e. b A x 2 = min x K m b Ax 2. Proof: If x is the minimizer of b Ax 2, it is necessary and sufficient that b A x is orthogonal to all vectors of the form v = Ay, where y J m. Since L m = AJ m, v = Ay L m. Hence x is the minimizer of b Ax 2, if and only if, b A x, v = 0, v L m, i.e. the Petrov-Galerkin condition is satisfied, and x is the approximated solution. Proposition 2..3 is applied in GMRES method. For L m = J m, there is a similar property, which is used in this thesis and omitted here. 2.2 Krylov Subspace A Krylov subspace method [29, 3] is a projection method for which the subspace J m is the Krylov subspace J m = K m (A, b) = span{b, Ab, A 2 b,..., A m b}. In this thesis, K m (A, b) will be denoted by K m. The approximations obtained from a Krylov subspace method are of the form A b x m = q m (A)b, where q m is a certain polynomial of degree (m ). 9

23 The dimension of the subspace increases by one at each step of the iterative process. The subspace K m is the subspace of all vectors in R N which can be written as x = p(a)b, where p is a polynomial of degree not exceeding m. 2.3 GMRES method The Generalized Minimum Residual Method(GMRES) [32, 4] is a projection method based on taking J m = K m and L m = AK m, in which K m is the m-th Krylov subspace with v = r 0 / r 0 2. The GMRES method minimizes the residual norm over all vectors in the Krylov subspace K m Arnoldi Algorithm Arnoldi s process [2, 30] is applied to build an orthogonal basis of the Krylov subspace K m. In exact arithmetic, the Arnoldi algorithm is described as Algorithm Algorithm 2.3. Arnoldi algorithm : Choose a vector v of norm ; 2: for j =,2,...,m do 3: Compute h ij = Av j, v i for i =,2,...,j; 4: Compute w j = Av j j i= h ijv i ; 5: h j+,j = w j 2 ; 6: if h j+,j = 0 then stop; 7: v j+ = w j /h j+,j ; 8: end for According to Algorithm 2.3., {v, v 2,...v m } forms an orthogonal basis of the Krylov subspace K m. At each step, the algorithm multiplies the previous vector v j by A and then orthonormalizes the resulting vector w j against all previous v i s by a standard Gram- Schmidt procedure. If w j vanishes, then the process stops. If w j does not vanish, then another vector v j+ is obtained, and the algorithm produces a bigger Krylov subspace K j+. If A m b K m, then the dimension of K m+ is equal to the dimension of K m. 0

24 0 A V m = V m H m +h m+,m v m+ e T m Figure 2.3.: AV m = V m H m + h m+,m v m+ e T m. According to Algorithm 2.3., the following property is obtained [29]. Proposition Let V m = [v, v 2,..., v m ], where {v,..., v m } are obtained from Algorithm 2.3., and H m be the (m + ) m Hessenberg matrix whose nonzero entries h ij are defined by Algorithm 2.3., and H m obtained from H m by deleting the last row. Then AV m = V m+ Hm (2.3.4) = V m H m + h m+,m v m+ e T m, (2.3.5) and hence V T m AV m = H m. (2.3.6) The proof just follows from Algorithm 2.3., and is omitted here. Obviously, the relation (2.3.4) and (2.3.5) are equivalent, the relation (2.3.6) is obtained by multiplying both sides of (2.3.5) by Vm T. The relation (2.3.5) is represented in Figure In practice, the Modified Gram-Schmidt [34] or the Householder Arnoldi [40] is used instead of the standard Gram-Schmidt algorithm. The Modified Gram-Schmidt results in Algorithm [34].

25 Algorithm Arnoldi-Modified Gram-Schmidt : Choose a vector v of norm ; 2: for j =,2,...,m do 3: Compute w j = Av j ; 4: for i =,2,...,j do 5: h ij = w j, v i ; 6: w j = w j h ij v i ; 7: end for 8: h j+,j = w j 2. If h j+,j = 0 then stop; 9: v j+ = w j /h j+,j ; 0: end for The Arnoldi-Modified Gram-Schmidt and Algorithm 2.3. are mathematically equivalent in exact arithmetic. If we consider the round-off in practice, Algorithm is more reliable. If even Algorithm is inadequate, then double orthogonalization or the Housholder Arnoldi can be used. The Arnoldi-Modified Gram-schmidt algorithm is applied in the GMRES algorithm in the next subsection GMRES Algorithm By using the Arnoldi-Modified Gram-schmidt algorithm [34] to build an orthogonal basis of the Krylov subspace K m, the GMRES method seek an approximation minimizing the residual norm over all vectors in the Krylov subspace K m. Any vector x in the Krylov subspace K m can be written as: x = V m y, where y is an m-vector. Then b Ax 2 = b AV m y 2. 2

26 Furthermore, we have b Ax = b AV m y = βv V m+ Hm y = V m+ (βe H m y), where β = b 2. Since the column-vectors of V m+ are othonormal, then b Ax 2 = βe H m y 2. (2.3.7) The GMRES approximation gives the unique vector in K m, which minimizes b Ax 2. According to (2.3.7), this approximation can be obtained by seeking an optimal y m that minimizes βe H m y 2, hence the approximation is x m = V m y m. The y m is easier to compute since it is the solution of an (m + ) m least-squares problem where m is not big generally. The algorithm is described as Algorithm [32]. Algorithm GMRES : Compute r 0 = b, β = r 0 2, and v = r 0 /β; 2: Define the (m + ) m matrix H m = {h ij } i m+, j m ; 3: for j =,2,...,m do 4: Compute w j = Av j ; 5: for i =,2,...,j do 6: h ij = w j, v i ; 7: w j = w j h ij v i ; 8: end for 9: h j+,j = w j 2. If h j+,j = 0 set m = j and go to 2; 0: v j+ = w j /h j+,j ; : end for 2: Compute y m, the minimizer of βe H m y 2, and x m = V m y m. Algorithm calculates the GMRES approximation as: x m = V m y m, (2.3.8) where y m = argmin y βe H m y 2. (2.3.9) 3

27 Similar to Algorithm 2.3., Algorithm will stop if w j vanishes, i.e., when h j+,j = 0 at some step j. If the algorithm stops in this way, that means the residual vector is zero, hence the approximation of GMRES will be the exact solution. The following proposition is Proposition 2 in [32]. Proposition The solution x j produced by GMRES at step j is exact if and only if the following four equivalent conditions hold:. The algorithm breaks down at step j. 2. v j+ = h j+,j = The degree of the minimal polynomial of A on the initial residual vector r 0 is equal to j, where the minimal polynomial of A on r 0 is the monic polynomial of least degree such that p(a)r 0 = 0. The breakdown results in the exact solution. Because the degree of the minimal polynomial of v can not exceed N for an N-dimensional problem, GMRES terminates in at most N steps [32]. We also can illustrate the above property by Figure 2.3.2, which is obtained by letting h j+,j = 0 in Figure Practical Implementation of GMRES In order to solve the least-squares problem min βe H m y 2 in the last step of Algorithm 2.3.3, the Hessenberg matrix is transformed into upper triangular form by using 4

28 0 A V j = V j H j Figure 2.3.2: AV j = V j H j + 0. Givens rotations [6, 32]. Define the rotation matrices as Q i =... c i s i s i c i..., (2.3.0) where c 2 i + s 2 i =. Given H m as an (m + ) m matrix, Q i s are (m + ) (m + ) matrices for i m. The idea can be best explained for m = 4. We have h h 2 h 3 h 4 β h 2 h 22 h 23 h 24 H 4 = h 32 h 33 h 34 h 43 h 44, γ = (2.3.) h 54 0 Multiply H 4 and γ by where c s s c Q =, c = s = h, h 2 + h 2 2 h 2. h 2 + h 2 2 5

29 Then we have H () 4 = h () h () 2 h () 3 h () 4 h () 22 h () 23 h () 24 h 32 h 33 h 34 h 43 h 44 h 54 c β, s β γ() = 0 0. (2.3.2) 0 Now h 2 has been eliminated, and we can multiply Q 2 to eliminate h 32, and continue the elimination process until the upper triangular H (4) 4 is obtained, h (4) h (4) 2 h (4) 3 h (4) 4 γ c β h (4) 22 h (4) 23 h (4) 24 (4) R 4 = H 4 = h (4) 33 h (4) γ 2, γ (4) = 34 γ 3 h (4) γ 4 = c 2 s β c 3 s 2 s β c 4 s 3 s 2 s β, (2.3.3) 44 0 γ 5 s 4 s 3 s 2 s β where c i and s i are defined as c i = s i = h (i ) ii (h (i ) ii h i+,i (h (i ) ii ) 2 + h 2 i+,i ) 2 + h 2 i+,i, (2.3.4). (2.3.5) Let Q = Q 4 Q 3 Q 2 Q, then R 4 = Q H 4, (2.3.6) γ (4) = Q γ = Qβe, (2.3.7) and γ (4) R 4 y = Qβe Q H 4 y (2.3.8) = Q(βe H 4 y). (2.3.9) Since Q is unitary, min γ (4) R 4 y 2 = min βe H 4 y 2. (2.3.20) 6

30 The solution to min γ (4) R 4 y 2 can be calculated by solving the upper triangular system, Moreover, min γ (4) R 4 y 2 = γ 5. h (4) h (4) 2 h (4) 3 h (4) 4 h (4) 22 h (4) 23 h (4) 24 h (4) 33 h (4) 34 h (4) 44 y = γ γ 2 γ 3 γ 4. (2.3.2) 2.4 Chebyshev Polynomials Chebyshev polynomials [28, ], named after Pafnuty Chebyshev, are a sequence of orthogonal polynomials which are related to de Moivre s formula and which are easily defined recursively. In this thesis, we calculate Chebyshev polynomials of the first kind which are denoted by T n (x) and Chebyshev polynomials of the second kind which are denoted by U n (x) Chebyshev Polynomials of the First Kind The Chebyshev polynomials of the first kind [28] are a set of orthogonal polynomials defined as the solutions to the Chebyshev differential equation. They are used as an approximation to a least squares fit. They are also intimately connected with trigonometric multiple-angle formulas. They are normalized such that T n () =. The Chebyshev polynomials of the first kind can be obtained from the generating function g(t, x) = xt 2xt + t = T 2 n (x)t n, n=0 for x < and t <. 7

31 The Chebyshev polynomials of the first kind are: T 0 (x) =, T (x) = x, T 2 (x) = 2x 2,..., T n+2 (x) = 2T n+ x T n (x). The Chebyshev polynomials of the first kind can also be defined through the identity T n (cos θ) = cos(nθ). Let x = cos(θ). Then T n (x) = cos(nθ) = cos(n arccos(x)). The sequence of Chebyshev polynomials T n (x) composes a sequence of orthogonal polynomial. Two different polynomials T n (x), T m (x) in the sequence are orthogonal to each other in the following sense dx 0, if m n, T n (x) T m (x) = π, if m = n = 0, x 2 π/2, if m = n 0. (2.4.22) Chebyshev Polynomials of the Second Kind The Chebyshev polynomials of the second kind [28] are denoted U n (x) and are defined by a different generating function g(t, x) = 2xt + t = U 2 n (x)t n, n=0 for x < and t <. 8

32 The Chebyshev polynomials of the second kind are: U 0 (x) =, U (x) = 2x, U 2 (x) = 4x 2,..., U n+2 (x) = 2U n+ x U n (x). Letting x = cos(θ) allows the Chebyshev polynomials of the second kind to be written as U n (x) = sin[(n + )θ)]. sin(θ) The sequence of U n (x) also composes a orthogonal polynomial sequence. Two different polynomials U n (x), U m (x) in the sequence are orthogonal to each other in the following sense U n (x) U m (x) x 2 dx = { 0, if m n, π/2, if m = n. (2.4.23) Copyright c Wei Zhang

33 Chapter 3 Convergence Analysis Using Chebyshev Polynomials of the First Kind In this and next chapter, we will present our main result of the convergence analysis of GMRES, i.e. estimation of the residuals of a tridiagonal Toeplitz matrix. Section 3. introduces how the residuals are calculated. Section 3.2 applies Chebyshev polynomial to calculate the residuals. In Section 3.3, the upper bounds of estimated residuals in general case are given, and numerical examples are provided to show how sharp our error bounds are. More error bounds of special cases are obtained in Section 3.4 and 3.5. In this chapter, a general case means a linear system with a general right hand side; special cases are ones with special right hand sides, such as b = e, b = e N. 3. Residual Formulation for a Diagonalizable Linear System The result in this section applies to any Ax = b with diagonalizable A. Suppose A = XΛX, (3..) Λ = diag(λ,..., λ N ), (3..2) and X is an N N nonsingular matrix. Recall we assumed, without loss of generality, the initial approximation x 0 = 0 and 20

34 thus the initial residual r 0 = b Ax 0 = b. The kth GMRES residual is r k 2 = min p k (A)b 2 p k (0)= = min Xp k (Λ)X b 2 p k (0)= = min u () = Y V T k+,nu 2, (3..3) where Y = X diag(x b), (3..4) V k+,n is the (k + ) N rectangular Vandermonde matrix def λ λ 2 λ N V k+,n =......, (3..5) λ k λ k 2 λ k N having nodes {λ j } N j=[4, Lemma 2.] and the coefficients of p k (A) forms a vector u with u () =. 3.2 Residual Reformulation Using Chebyshev Polynomials of the First Kind An N N tridiagonal Toeplitz A takes this form λ µ ν A = µ CN N. (3.2.) ν λ Given parameters ν, λ, and µ are nonzero, A is diagonalizable. In fact [33, pp.3-5] (see also [26]), A = XΛX, (3.2.2) 2

35 where Λ = diag(λ,..., λ N ), (3.2.3) λ j = λ 2 µν t j, (3.2.4) t j = cos θ j, θ j = jπ N +, (3.2.5) X = ΩZ, (3.2.6) µν Ω = diag(ξ 0, ξ,..., ξ N+ ), ξ = ν, (3.2.7) 2 Z (:,j) = N + (sin jθ,..., sin jθ N ) T. (3.2.8) It can be verified that Z T Z = I N. So A is normal if ξ =, i.e., µ = ν > 0. By (3.2.4), we have λ j = ω(t j τ), j N, (3.2.9) where ω = 2 µν, (3.2.0) τ = λ 2 µν. (3.2.) Any branch of µν, once picked and fixed, is a valid choice in this thesis. According to (3..3), we have r k 2 = min u () = Y V T k+,nu 2 Y 2 min u () = V T k+,nu 2 X 2 max (X b) (i) min V i u () k+,nu T 2 = X 2 X b 2 min u () = V T k+,nu 2 X 2 X 2 b 2 min u () = V T k+,nu 2 κ(x) b 2 min max p k (0)= i 22 p k (λ i ), (3.2.2)

36 where p k is a polynomial of degree no higher than k. Thus, together with r 0 = b, they imply r k 2 / r 0 2 κ(x) min max p k (λ i ). (3.2.3) p k (0)= i Inequality (3.2.3) is often the starting point in existing quantitative analysis on GMRES convergence [8, Page 54], as it seems that there is no easy way to do otherwise. It simplifies the analysis by separating the study of GMRES convergence behavior into optimizing the condition number of X and a polynomial minimization problem over A s spectrum, but it could potentially overestimate GMRES residuals. This is partly because, as observed by Liesen and Strakoš [26], possible cancelations of huge components in X and/or X were artificially ignored for the sake of the convergence analysis. However, for tridiagonal Toeplitz matrix A the rich structure (the Chebyshev polynomial) allows us to estimate differently, namely starting with (3..3) directly. We define the mth Translated Chebyshev Polynomial of the first kind in z of degree m as T m (z; ω, τ) def = T m (z/ω + τ) (3.2.4) = a mm z m + a m m z m + + a m z + a 0m, (3.2.5) where the coefficients a jm a jm (ω, τ) are functions of ω and τ, and forms an upper triangular R m+ in terms of ω and τ as R m+ R m+ (ω, τ) def = a 00 a 0 a 02 a 0 m a a 2 a m a 22 a 2 m, (3.2.6).... a m m 23

37 i.e., the (m + )th column consists of the coefficients of T m (λ; ω, τ). Set T 0 (t ) T 0 (t 2 ) T 0 (t N ) def T (t ) T (t 2 ) T (t N ) T N =..., (3.2.7) T N (t ) T N (t 2 ) T N (t N ) and V N = V N,N for short. Then V T N R N = T T N. (3.2.8) Equation (3.2.8) yields V T N = T T NR N. Extracting the first k + columns from both sides of V T N = T T NR N yields V T k+,n = (T T N) (:,:k+) R k+. (3.2.9) Now notice Y = X diag(x b) and X = ΩZ with Z in (3.2.6) being real and orthogonal to get Y V T k+,n = ΩZ diag(z T Ω b) (T T N) (:,:k+) R k+ = ΩM (:,:k+) R k+ (3.2.20) = ΩM (:,:k+) Ω k+ Ω k+r k+. (3.2.2) where Ω k+ = Ω (:k+,:k+), the (k + )th leading submatrix of Ω, M = Z diag(z T Ω b) T T N. (3.2.22) Now we can estimate GMRES residual Then r k 2 = min u () = Y V T k+,nu 2 = min u () = ΩM (:,:k+)ω k+ Ω k+r k+ u 2. (3.2.23) σ min (ΩM (:,:k+) Ω k+ ) r k 2 min u() = Ω k+ R k+ u 2 ΩM (:,:k+) Ω k+ 2. (3.2.24) Hence, the upper bound and lower bound of the residual r k 2 could be estimated. 24

38 3.3 Estimation of Residual in General Case Set { ζ = min ξ, }, (3.3.) ξ { τ ρ = max + τ 2, τ } τ 2. (3.3.2) Note ρ always because (τ + τ 2 )(τ τ 2 ) =. In particular if λ R, µ < 0 and ν > 0, then ρ = τ + τ 2 +. Recall Chebyshev polynomials of the first kind Define T m (t) = cos(m arccos t), for t, (3.3.3) = 2 ( t + ) m ( t 2 + t m t 2 2 ), for t. (3.3.4) Φ (+) k+ (τ, ξ) def = Φ ( ) k+ (τ, ξ) def = Φ k+ (τ, ξ) def = k ξ 2j T j (τ) 2, (3.3.5) j=0 k ξ 2j T j (τ) 2, (3.3.6) j=0 k j=0 { } ζ 2j T j (τ) 2 min Φ (+) k+ (τ, ξ), Φ( ) k+ (τ, ξ), (3.3.7) where j means the first term is halved. Obviously, if ξ, then Φ k+ (τ, ξ) = Φ (+) k+ (τ, ξ); otherwise, Φ k+ (τ, ξ) = Φ ( ) k+ (τ, ξ) Residual with General Right-hand Sides Given a tridiagonal Toeplitz with a general right hand side, we have 25

39 Theorem For Ax = b, where A is tridiagonal Toeplitz as in (3.2.) with nonzero (real or complex) parameters ν, λ, and µ. Then the kth GMRES residual r k satisfies for k < N r k 2 [ ] /2 k + r Φ k+(τ, ξ). (3.3.8) The proof of Theorem 3.3. involves a complicated computation. We will prove the theorem for the case ξ first. Recall the computation in the previous section, the proof follows the inequality (3.2.24) σ min (ΩM (:,:k+) Ω k+ ) r k 2 min u() = Ω k+ R k+ u 2 Rewrite the second inequality in (3.2.24): ΩM (:,:k+) Ω k+ 2. r k 2 min u () = Ω k+r k+ u 2 ΩM (:,:k+) Ω k+ 2. (3.3.9) This is our foundation to prove Theorem There are two quantities to deal with min Ω k+r u () = k+ u 2 and ΩM (:,:k+) Ω k+ 2. (3.3.0) For the first part, the following lemma was proven in [23, 24], and also implied by the proof of [25, Theorem 2.]. See also [22]. Lemma If W has full column rank, then min W u 2 = [ ] e T (W W ) /2 e. (3.3.) u () = In particular if W is nonsingular, min u() = W u 2 = W e 2. Proof: Set v = W u. Since W has full column rank N, its Moore-Penrose pseudoinverse is W = (W W ) W [35], and thus u = W v. This gives a one-one and onto mapping between u C m and the column space v span(w ), where span(w ) = span{w (: 26

40 , ), W (:, 2),..., W (:, N)}. Then W u 2 min W u 2 = min u () = u u () where the last min is achieved at = min v span(w ) v 2 e T W v min v v 2 e T W v = et W 2, (3.3.2) v opt = (e T W ) = W (W W ) e span(w ), which implies the in (3.3.2) is an equality, and u opt = W v opt e T W v opt. Finally e T W 2 = e T W (W ) e = e T (W W ) e. The above proof is taken from [23]. By this lemma, we have (note a 00 = ) [ ] /2 min Ω k+r u () = k+ u 2 = Ω k+ R k+e 2 = 2 + Φ(+) k+ (τ, ξ). (3.3.3) This gives the first quantity in (3.3.0). Rewrite Theorem3.3. as [ ] /2 r k 2 r 0 2 k Φ k+(τ, ξ) [ ] /2 b 2 k Φ k+(τ, ξ) [ ] /2 2 + Φ k+(τ, ξ) b 2 k +. If we can show ΩM (:,:k+) Ω k+ 2 b 2 k +, (3.3.4) then according to inequality (3.3.9) and (3.3.3), Theorem 3.3. is proved. We prove the inequality (3.3.4) in the next subsection and finish the proof of Theorem

41 3.3.2 The Second Part of the Proof Now let s investigate ΩM (:,:k+) Ω k+ and prove the inequality (3.3.4). It can be seen that ΩM (:,:k+) Ω k+ = (ΩMΩ ) (:,:k+) since Ω is diagonal. To compute ΩMΩ, we shall investigate M in (3.2.22) first. M = Z diag(z T Ω b) T T N N = Z diag(zω b (l) e l ) T T N = = = l= N b (l) ξ l Z diag(ze l ) T T N l= N b (l) ξ l Z diag(z (:,l) ) T T N l= N b (l) ξ l M l, (3.3.5) l= where M l = Z diag(z (:,l) ) T T N and b (l) is the lth element of the right hand side b. It is not easy to obtain M directly, but M l can be calculated in Lemma In Lemma and in the proof of Lemma below, without causing notational conflict, we will temporarily use k as a running index, as opposed to the rest of the paper where k is reserved for GMRES step index. Lemma For θ j = N cos lθ k = k= j π and integer l, N+ N, if l = 2m(N + ) for some integer m,, if l is even, but l 2m(N + ) for any integer m, 0, if l is odd. (3.3.6) Proof: If l = 2m(N + ) for some integer m, then lθ k = 2mkπ and thus cos lθ k =. Assume that l 2m(N + ) for any integer m. Set Ω = lπ/(n + ). We have [7, p.30] N cos lθ k = k= N k= cos kω = cos N + Ω 2 NΩ sin 2. sin Ω 2 28

42 Now notice cos N+ 2 Ω = cos l 2 π = 0 for odd l and ( )l/2 for even l, and sin NΩ 2 = sin( l 2 π Ω) = ( )l/2 sin Ω for even l to conclude the proof. Lemma is Lemma 3. in [23]. Lemma Let M l def = Z diag(z (:,l) )T T N for l N. Then the entries of M l are zeros, except at those positions (i, j), graphically forming four straight lines: (a) i + j = l +, (b) i j = l, (c) j i = l +, (d) i + j = 2(N + ) l +. (3.3.7) (M l ) (i,j) = /2 for (a) and (b), except at their intersection (l, ) for which (M l ) (l,) =. (M l ) (i,j) = /2 for (c) and (d). Notice no valid entries for (c) if l N and no valid entries for (d) if l 2. Proof: For i, j N, 2(N + ) (M l ) (i,j) = 4 = 4 = 2 = N sin kθ i sin lθ k cos(j )θ k k= N sin iθ k sin lθ k cos(j )θ k k= N [cos(i l)θ k cos(i + l)θ k ] cos(j )θ k k= N [cos(i + j l )θ k + cos(i j l + )θ k k= cos(i + j + l )θ k cos(i j + l + )θ k ]. Since all i = i + j l, i 2 = i j l +, i 3 = i + j + l, i 4 = i j + l +, 29

43 are either even or odd at the same time, Lemma implies (M l ) (i,j) = 0 unless one of them takes the form 2m(N + ) for some integer m. We now investigate all possible situations as such, keeping in mind that i, j, l N.. i = i + j l = 2m(N + ). This happens if and only if m = 0, and thus i + j = l +. Then i 2 = 2j + 2, i 3 = 2l, i 4 = 2j + 2l + 2. They are all even. i 3 and i 4 do not take the form 2m(N + ) for some integers m. This is obvious for i 3, while i 4 = 2m(N + ) implies m = 0 and j = l +, and thus i = 0 which cannot happen. However if i 2 = 2m(N + ), then m = 0 and j =, and thus i = l. So Lemma implies (M l ) (i,j) = /2 for i+j = l+ and i l, while (M l ) (l,) =. 2. i 2 = i j l + = 2m(N + ). This happens if and only if m = 0, and thus i j = l. Then i = 2j 2, i 3 = 2j + 2l 2, i 4 = 2l. They are all even. i 3 and i 4 do not take the form 2m(N + ) for some integers m. This is obvious for i 4, while i 3 = 2m(N + ) implies m = and j = N + 2 l, and thus i = N + which cannot happen. However if i = 2m(N + ), then m = 0 and thus j = and i = l which has already been considered in Item. So Lemma implies (M l ) (i,j) = /2 for i j = l and i l, while (M l ) (l,) =. 3. i 3 = i + j + l = 2m(N + ). This happens if and only if m =, and thus i + j = 2(N + ) l +. Then i = 2(N + ) 2l, i 2 = 2(N + ) 2j 2l + 2, i 4 = 2(N + ) 2j

44 They are all even. i and i 2 do not take the form 2m(N + ) for some integers m. This is obvious for i, while i 2 = 2m(N + ) implies m = 0 and j = N + 2 l, and thus i = N + which cannot happen. However if i 4 = 2m(N + ), then m = and thus j = and i = 2(N + ) l which is bigger than N + 2 and not possible. So Lemma implies (M l ) (i,j) = /2 for i + j = 2(N + ) l i 4 = i j + l + = 2m(N + ). This happens if and only if m = 0, and thus j i = l +. Then i = 2j 2l 2, i 2 = 2l, i 3 = 2j 2. They are all even, and do not take the form 2m(N + ) for some integers m. This is obvious for i 2. i = 2m(N + ) implies m = 0 and j = l +, and thus i = 0 which cannot happen. i 3 = 2m(N + ) implies m = 0 and thus j = and i = l which cannot happen either. So Lemma implies (M l ) (i,j) = /2 for j i = l +. This completes the proof. Figure 3.3. gives the example for M l with N = 8 for L =, 2, 3, 4. The structure of M l, i.e. the nonzero entries, is showed in each graph. The entries on the red lines have the value 2 except on the first column with M l(l, ) = ; the entries on the black lines have the value. The lines are labeled with a, b, c, d according to Lemma Figure gives the example for M l with N = 8 for L = 5, 6, 7, 8 similarly. Now we know M l. We still need to find out ΩMΩ. Let us examine it for N = 5 in 3

45 0 M(): with N=8 0 a M(2): with N=8 2 3 c 2 3 c b b nz = nz = 3 0 M(3): with N=8 0 M(4): with N=8 2 a c 2 a c b d b d nz = nz = 3 Figure 3.3.: The structure of M l with N = 8 for L =, 2, 3, 4. 32

46 0 M(5): with N=8 0 M(6): with N=8 2 3 a c 2 3 a c b d 7 8 b d nz = nz = 3 0 M(7): with N=8 0 M(8): with N= a 3 4 a d 6 7 d 8 b nz = nz = 4 Figure 3.3.2: The structure of M l with N = 8 for L = 5, 6, 7, 8. 33

47 order to get some idea about what it may look like. ΩMΩ for N = 5 is b () 2 ξ2 b (2) 2 ξ2 b () + 2 ξ4 b (3) 2 ξ4 b (2) + 2 ξ6 b (4) 2 ξ6 b (3) + 2 ξ8 b (5) b (2) b 2 () + b 2 (3) ξ 2 b 2 (4) ξ 4 2 ξ2 b () + 2 ξ6 b (5) 2 ξ4 b (2) b (3) b 2 (2) + 2 ξ2 b (4) b 2 () + b 2 (5) ξ ξ2 b () 2 ξ6 b (5) b (4) 2 b (3) + b 2 (5) ξ 2 b 2 (2) b 2 () b 2 (5) ξ 4 b. 2 (4) ξ 4 b (5) b 2 (4) b 2 (3) b 2 (5) ξ 2 b 2 (2) 2 ξ2 b (4) b 2 () b 2 (3) ξ 2 We observe that for N = 5, the entries of ΩMΩ are polynomials in ξ with at most two terms. This turns out to be true for all N. Lemma The following statements hold.. The first column of ΩMΩ is b. Entries in every other columns taking one of the three forms: (b (n )ξ m + b (n2 )ξ m 2 )/2 with n n 2, b (n )ξ m /2, and 0, where n, n 2 N and m i 0 are nonnegative integer. 2. In each given column of ΩMΩ, any particular entry of b appears at most twice. As the consequence, we have ΩM (:,:k+) Ω k+ 2 k + b 2, if ξ. Proof: Notice M = N l= b (l)ξ l M l and consider M s (i, j)th entry which comes from the contributions from all M l. But not all of M l contribute as most of them are zero at the position. Precisely, with the help of Lemma 3.3.3, those M l that contribute nontrivially to the (i, j)th position are the following ones subject to satisfying the given inequalities. (a) if i + j N or equivalently i + j N +, M i+j gives a /2. (b) if i j + N or equivalently i j, M i j+ gives a /2. (c) if j i N or equivalently j i + 2, M j i gives a /2. 34

48 i + j N + (a) (d) i + j N + 3 i j 2 (c) (b) i j 0 (a) (c) (a). (c) (b) (d) (b) (d) Figure 3.3.3: Computation of M (i,j). (d) if 2(N + ) (i + j) + N or equivalently i + j N + 3, M 2(N+) (i+j)+ gives a /2. These inequalities, effectively 4 of them, are described in Figure The left graph shows the regions of entries as divided by inequalities in (a) and (d); the middle shows Regions of entries as divided by inequalities in (b) and (c). These four regions divided entries of M into nine possible regions showed as the right graph of Figure We shall examine each region one by one. Recall (ΩMΩ ) (i,j) = ξ i+ M (i,j) ξ j = ξ j i M (i,j), and let γ a = γ b = 2 b (i+j )ξ 2j 2, 2 b (i j+), γ c = 2 b (j i )ξ 2(j i ), γ d = 2 b (2(N+) (i+j)+)ξ 2(N+) (i+j). Each entry in the 9 possible regions in the rightmost plot of Figure is as follows.. (a) and (b): (ΩMΩ ) (i,j) = γ a + γ b. 2. (a) and (c): (ΩMΩ ) (i,j) = γ a + γ c. 35

49 3. (b) and (d): (ΩMΩ ) (i,j) = γ b + γ d. 4. (c) and (d): (ΩMΩ ) (i,j) = γ c + γ d. 5. (a) and i j = : (ΩMΩ ) (i,j) = γ a. 6. (b) and i + j = N + 2: (ΩMΩ ) (i,j) = γ b. 7. (c) and i + j = N + 2: (ΩMΩ ) (i,j) = γ c. 8. (d) and i j = : (ΩMΩ ) (i,j) = γ d. 9. i j = and i + j = N + 2: (ΩMΩ ) (i,j) = 0. In this case, i = (N + )/2 and j = (N + 3)/2. So there is only one such entry when N is odd, and none when N is even. With this profile on the entries of ΩMΩ, we have Item of the lemma immediately. Item 2 is the consequence of M = N l= b (l)ξ l M l and Lemma which implies that there are at most two nonzero entries in each column of M l. As the consequence of Item and Item 2, each column of ΩMΩ can be expressed as the sum of two vectors w and v such that w 2, v 2 b 2 /2 when ξ, and thus (ΩMΩ ) (:,j) 2 b 2 for all j N. Therefore as expected. ΩM (:,:k+) Ω k+ 2 k+ (ΩMΩ ) (:,j) 2 2 k + b 2, j= We can prove Theorem 3.3. now. Proof: First we can prove [ ] /2 r k 2 b 2 k Φ(+) k+ (τ, ξ), for ξ. (3.3.8) 36

50 Assume ξ. Inequality (3.3.8) is the consequence of estimation of residuals (3.2.24) σ min (ΩM (:,:k+) Ω k+ ) r k 2 min u() = Ω k+ R k+ u 2 the equation(3.3.3) and Lemma ΩM (:,:k+) Ω k+ 2, [ ] /2 min Ω k+r u () = k+ u 2 = Ω k+ R k+e 2 = 2 + Φ(+) k+ (τ, ξ), ΩM (:,:k+) Ω k+ 2 k + b 2, if ξ. For the other case, ξ >, the proof can be turned into this case as follows. Let Π = (e N,..., e 2, e ) R N N be the permutation matrix. Notice Π T AΠ = A T and thus Ax = b is equivalent to A T Π T x = (Π T AΠ)(Π T x) = Π T b. (3.3.9) Note K k (A T, Π T b) = K k (Π T AΠ, Π T b) = Π T K k (A, b), and r k 2 = min b Ay 2 = min Π T (b AΠ Π T y) 2 y K k (A,b) Π T y Π T K k (A,b) Since (3.3.8) is proven true, then for ξ > we have = min Π T b A T w 2. w K k (A T,Π T b) [ ] /2 r k 2 Π T b 2 k Φ( ) k+ (τ, ξ), because the ξ for A T is the reciprocal of the one for A. Remark The leftmost inequality in (3.2.24) gives a lower bound on r k 2 in terms of σ min (ΩM (:,:k+) Ω k+ ) which, however, is hard to bound from below because it can be as small as zero, unless we know more about b such as b = e or e N as in Theorems 3.4. and

51 The upper bounds can also be calculated according to the inequality (3.2.3). Theorem For Ax = b, where A is tridiagonal Toeplitz as in (3.2.) with nonzero (real or complex) parameters ν, λ, and µ. Then the kth GMRES residual r k satisfies for k < N Proof: According to (3.2.3), [ k ] /2 r k 2 κ(x) T j (τ) 2. (3.3.20) r 0 2 j=0 r k 2 / r 0 2 κ(x) min max p k (λ i ). (3.3.2) p k (0)= i Now we calculate min pk (0)= max i p k (λ i ) according to (3.2.8) Apply Lemma 3.3., we have min max p k (λ i ) = min V p k (0)= i u () k+,nu T 2 = = min u () = (T T N) (:,:k+) R k+ u 2. (3.3.22) min max p k (λ i ) = min (T T p k (0)= i u () N) (:,:k+) R = k+ u 2 = [ e T (((T T N) (:,:k+) R k+ ) (T T N) (:,:k+) R k+ ) e ] /2 = [ e T (R k+ R k+ ) e ] /2 = [ ] e T (R k+ Rk+)e /2 [ k ] /2 = T j (τ) 2. (3.3.23) j=0 Now (3.2.3) can be written as (3.3.20) Numerical Examples In all numerical examples in this paper, we always take N = 50 and λ =, and choose different τ and ξ, then calculate µ and ν as following: µ = ξ, µ = ± µ, and ν = ν = 2 τ 2 τξ. 38

52 0 0, upper bounds ( τ =0.8, ξ =0.7) N = for µ=0.4375, ν= for µ= , ν= k+ 0 0, upper bounds ( τ =0.8, ξ =.2) N = for µ=0.75, ν= for µ= 0.75, ν= k+ Figure 3.3.4: GMRES residuals for random b with τ = 0.8, and the upper bounds by Theorem

53 0 0, upper bounds ( τ =.2, ξ =0.7) N = for µ=0.2967, ν= for µ= , ν= k+ 0 0, upper bounds ( τ =.2, ξ =.2) N = for µ=0.5, ν= for µ= 0.5, ν= k+ Figure 3.3.5: GMRES residuals for random b with τ =.2, and the upper bounds by Theorem

54 0 0, upper bounds ( τ =, ξ =0.7) N = for µ=0.35, ν= for µ= 0.35, ν= k+ 0 0, upper bounds ( τ =, ξ =.2) N = for µ=0.6, ν= for µ= 0.6, ν= k+ Figure 3.3.6: GMRES residuals for random b with τ =, and the upper bounds by Theorem

55 Thus µ, ν R, and in fact ν > 0 always. When µ > 0, ξ = µ/ν < 0 and τ = /(2 µν) > 0, but when µ < 0, both ξ = ι µ/ν and τ = ι/(2 µν ) are imaginary, where ι = is the imaginary unit. Figures 3.3.4, 3.3.5, plot residual histories for examples of GMRES with each of b s entries being uniformly random in [, ], and their upper bounds (dashed lines) by Theorem The parameters, τ and ξ are set in table Table 3.3.: Parameters Setting τ ξ Figure Figure Figure Each figure has two graphs with different ξ, 0.7 or.2. Each graph has two cases for a real τ and a pure imaginary τ. In each graph, the plot for µ > 0, i.e. τ is real, is above that for µ < 0, i.e. τ is purely imaginary. In other words, this indicates that GMRES converges much faster for µ < 0 than for µ > 0 in each of the plots. There is a simple explanation for this: the eigenvalues of A (see (3.2.9) below) are further away from the origin for a pure imaginary τ than for a real τ for any fixed τ. All plots indicate that our upper bounds are tight, except for the last few steps. Note that the upper bounds for the case µ > 0 are visually indistinguishable from the horizontal line 0 0 when τ, suggesting slow convergence. Figures 3.3.7, 3.3.8, plot residual histories for examples of GMRES with each of b s entries being uniformly random in [, ], and their upper bounds (dashed lines) by Theorem In each figure, the upper bounds are much bigger than the residuals. 42

The Rate of Convergence of GMRES on a Tridiagonal Toeplitz Linear System

The Rate of Convergence of GMRES on a Tridiagonal Toeplitz Linear System Numerische Mathematik manuscript No. (will be inserted by the editor) The Rate of Convergence of GMRES on a Tridiagonal Toeplitz Linear System Ren-Cang Li 1, Wei Zhang 1 Department of Mathematics, University

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. (009) 11:67 93 DOI 10.1007/s0011-008-006- Numerische Mathematik The rate of convergence of GMRES on a tridiagonal Toeplitz linear system Ren-Cang Li Wei Zhang Received: 1 November 007 / Revised:

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

ABSTRACT OF DISSERTATION. Ping Zhang

ABSTRACT OF DISSERTATION. Ping Zhang ABSTRACT OF DISSERTATION Ping Zhang The Graduate School University of Kentucky 2009 Iterative Methods for Computing Eigenvalues and Exponentials of Large Matrices ABSTRACT OF DISSERTATION A dissertation

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

c 2005 Society for Industrial and Applied Mathematics

c 2005 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 26, No. 4, pp. 1001 1021 c 2005 Society for Industrial and Applied Mathematics BREAKDOWN-FREE GMRES FOR SINGULAR SYSTEMS LOTHAR REICHEL AND QIANG YE Abstract. GMRES is a

More information

Linear Analysis Lecture 16

Linear Analysis Lecture 16 Linear Analysis Lecture 16 The QR Factorization Recall the Gram-Schmidt orthogonalization process. Let V be an inner product space, and suppose a 1,..., a n V are linearly independent. Define q 1,...,

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.

More information

Lecture 3: QR-Factorization

Lecture 3: QR-Factorization Lecture 3: QR-Factorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QR-factorization of matrices It also outlines some applications of this factorization

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Antti Koskela KTH Royal Institute of Technology, Lindstedtvägen 25, 10044 Stockholm,

More information

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition AM 205: lecture 8 Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition QR Factorization A matrix A R m n, m n, can be factorized

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems

GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University December 4, 2011 T.-M. Huang (NTNU) GMRES

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

ELE/MCE 503 Linear Algebra Facts Fall 2018

ELE/MCE 503 Linear Algebra Facts Fall 2018 ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 4: Iterative Methods PD

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Applied Linear Algebra

Applied Linear Algebra Applied Linear Algebra Gábor P. Nagy and Viktor Vígh University of Szeged Bolyai Institute Winter 2014 1 / 262 Table of contents I 1 Introduction, review Complex numbers Vectors and matrices Determinants

More information

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education MTH 3 Linear Algebra Study Guide Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education June 3, ii Contents Table of Contents iii Matrix Algebra. Real Life

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1

More information

On the solution of large Sylvester-observer equations

On the solution of large Sylvester-observer equations NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 200; 8: 6 [Version: 2000/03/22 v.0] On the solution of large Sylvester-observer equations D. Calvetti, B. Lewis 2, and L. Reichel

More information

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

Intrinsic products and factorizations of matrices

Intrinsic products and factorizations of matrices Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 5 3 www.elsevier.com/locate/laa Intrinsic products and factorizations of matrices Miroslav Fiedler Academy of Sciences

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

Lecture 6. Numerical methods. Approximation of functions

Lecture 6. Numerical methods. Approximation of functions Lecture 6 Numerical methods Approximation of functions Lecture 6 OUTLINE 1. Approximation and interpolation 2. Least-square method basis functions design matrix residual weighted least squares normal equation

More information

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{96{2 TR{3587 Iterative methods for solving Ax = b GMRES/FOM versus QMR/BiCG Jane K. Cullum

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm References: Trefethen & Bau textbook Eigenvalue problem: given a matrix A, find

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Applied Numerical Linear Algebra. Lecture 8

Applied Numerical Linear Algebra. Lecture 8 Applied Numerical Linear Algebra. Lecture 8 1/ 45 Perturbation Theory for the Least Squares Problem When A is not square, we define its condition number with respect to the 2-norm to be k 2 (A) σ max (A)/σ

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to

More information

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated. Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces

More information

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9 1. qr and complete orthogonal factorization poor man s svd can solve many problems on the svd list using either of these factorizations but they

More information

MAA507, Power method, QR-method and sparse matrix representation.

MAA507, Power method, QR-method and sparse matrix representation. ,, and representation. February 11, 2014 Lecture 7: Overview, Today we will look at:.. If time: A look at representation and fill in. Why do we need numerical s? I think everyone have seen how time consuming

More information

5.6. PSEUDOINVERSES 101. A H w.

5.6. PSEUDOINVERSES 101. A H w. 5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

IDR(s) as a projection method

IDR(s) as a projection method Delft University of Technology Faculty of Electrical Engineering, Mathematics and Computer Science Delft Institute of Applied Mathematics IDR(s) as a projection method A thesis submitted to the Delft Institute

More information

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i ) Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I) CS 106 Spring 2004 Homework 1 Elena Davidson 8 April 2004 Problem 1.1 Let B be a 4 4 matrix to which we apply the following operations: 1. double column 1, 2. halve row 3, 3. add row 3 to row 1, 4. interchange

More information

THESIS. Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University

THESIS. Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University The Hasse-Minkowski Theorem in Two and Three Variables THESIS Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University By

More information

Class notes: Approximation

Class notes: Approximation Class notes: Approximation Introduction Vector spaces, linear independence, subspace The goal of Numerical Analysis is to compute approximations We want to approximate eg numbers in R or C vectors in R

More information

Tikhonov Regularization of Large Symmetric Problems

Tikhonov Regularization of Large Symmetric Problems NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2 MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

The Solution of Linear Systems AX = B

The Solution of Linear Systems AX = B Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has

More information

On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell

On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell DAMTP 2014/NA02 On fast trust region methods for quadratic models with linear constraints M.J.D. Powell Abstract: Quadratic models Q k (x), x R n, of the objective function F (x), x R n, are used by many

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

Linear Algebra. Min Yan

Linear Algebra. Min Yan Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Review of similarity transformation and Singular Value Decomposition

Review of similarity transformation and Singular Value Decomposition Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm

More information

Householder reflectors are matrices of the form. P = I 2ww T, where w is a unit vector (a vector of 2-norm unity)

Householder reflectors are matrices of the form. P = I 2ww T, where w is a unit vector (a vector of 2-norm unity) Householder QR Householder reflectors are matrices of the form P = I 2ww T, where w is a unit vector (a vector of 2-norm unity) w Px x Geometrically, P x represents a mirror image of x with respect to

More information

Matrices, Moments and Quadrature, cont d

Matrices, Moments and Quadrature, cont d Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that

More information

On Solving Large Algebraic. Riccati Matrix Equations

On Solving Large Algebraic. Riccati Matrix Equations International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

Block Lanczos Tridiagonalization of Complex Symmetric Matrices

Block Lanczos Tridiagonalization of Complex Symmetric Matrices Block Lanczos Tridiagonalization of Complex Symmetric Matrices Sanzheng Qiao, Guohong Liu, Wei Xu Department of Computing and Software, McMaster University, Hamilton, Ontario L8S 4L7 ABSTRACT The classic

More information

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1... Chapter Matrices We review the basic matrix operations What is a Matrix? An array of numbers a a n A = a m a mn with m rows and n columns is a m n matrix Element a ij in located in position (i, j The elements

More information

(v, w) = arccos( < v, w >

(v, w) = arccos( < v, w > MA322 F all206 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: Commutativity:

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

MTH 2032 SemesterII

MTH 2032 SemesterII MTH 202 SemesterII 2010-11 Linear Algebra Worked Examples Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education December 28, 2011 ii Contents Table of Contents

More information