A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem

Size: px
Start display at page:

Download "A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem"

Transcription

1 A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van Barel Department of Computer Science Katholieke Universiteit Leuven Belgium International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

2 Outline Introduction Cholesky LR algorithm Laguerre s shift Implementation Numerical examples Conclusions International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

3 Introduction semiseparable: every submatrix from the lower or upper triangular part has rank at most 1. diagonal-plus-semiseparable (DPSS): the sum D + S of a diagonal D and a semiseparable S. Givens-vector representation of a DPSS matrix is based on a vector f = [f 1,..., f n ] T, Givens [ ] ci s rotations G i = i, i = 1,..., n 1, and a diagonal d = [d 1,..., d n ] T as s i c i c 1 f 1 + d 1 c 2 s 1 f 1 c n 1 s n 2:1 f 1 s n 1:1 f 1 c 2 s 1 f 1 c 2 f 2 + d 2 c n 1 s n 2:1 f 2 s n 1:2 f 2 D + S =....., c n 1 s n 2:1 f 1 c n 1 s n 2:2 f 2 c n 1 f n 1 + d n 1 s n 1 f n 1 s n 1:1 f 1 s n 1:2 f 2 s n 1 f n 1 f n + d n where s a:b = s a s a 1 s b. When they appear, we assume that c n = 1 and s n = 0. We denote D + S = diag(d) + Giv(c, s, f). Our goal is to compute the smallest (or all) eigenvalues of a s.p.d. DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

4 Motivation Let A be an n n symmetric matrix. One can find an orthogonal matrix Q such that B = Q T AQ is tridiagonal in O(n 3 ) flops. Next, the eigendecomposition of B is computed in O(n 2 ) flops. Instead of the tridiagonal structure one can use DPSS matrices. Vandebril, Van Camp, Van Barel, Mastronardi (2004): For an arbitrary diagonal matrix D there exists an orthogonal matrix Q such that Q T AQ = D+S is a DPSS matrix. The reduction is O(n 2 ) more expensive than in the tridiagonal case. The reduction algorithm has a Lanczos-Ritz convergence behaviour and performs a kind of nested subspace iteration at each step. A good choice of the diagonal can compensate slower reduction. Algorithms for the eigendecomposition of a symmetric DPSS matrix: Chandrasekaran, Gu (2004), in Numer. Math.: divide and conquer, Mastronardi, Van Camp, Van Barel (2003), tech. report: divide and conquer, Bini, Gemignani, Pan (2003), tech. report: QR algorithm, Fasino (2004), tech. report: QR algorithm, Van Camp, Delvaux, Van Barel, Vandebril, Mastronardi (2005), tech. report: implicit QR algorithm, Mastronardi, Van Camp, Van Barel, Vandebril (2004), tech. report: computation of eigenvectors International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

5 Cholesky LR algorithm Let A be a s.p.d. matrix. A 0 = A k = 0, 1, 2,... choose shift σ k A k σ k I = V k V T k A k+1 = V T k V k + σ k I (Cholesky decomposition, V k is lower-triangular) Two steps of the zero shift Cholesky LR are equivalent to one step of the zero shift QR. The shift σ k should be such that A k σ k I is positive definite. When applied to a s.p.d. DPSS matrix D + S, the shift can be included into the diagonal part. Grad, Zakrajšek (1972): Cholesky LR with Laguerre s shifts for the symmetric tridiagonal matrices. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

6 DPSS is invariant to Cholesky LR Theorem: Let A = Giv(c, s, f) + diag(d) be a s.p.d DPSS matrix. 1. If A = V V T is the Cholesky decomposition of A, then V = tril(giv(c, s, f)) + diag( d), 2. If B = V T V, then B is a s.p.d. DPSS matrix B = Giv(ĉ, ŝ, f) + diag(d). A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3. V 4 = B 4 = s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4 c 1 f1 + d 1 c 2 s 1 f1 c 2 f2 + d 2 c 3 s 2 s 1 f1 c 3 s 2 f2 c 3 f3 + d 3 s 3 s 2 s 1 f1 s 3 s 2 f2 s 3 f3 f4 + d 4 ĉ 1 f1 + d 1 ĉ 2 ŝ 1 f1 ĉ 2 f2 + d 2 ĉ 3 ŝ 2 ŝ 1 f1 ĉ 3 ŝ 2 f2 ĉ 3 f3 + d 3.. ŝ 3 ŝ 2 ŝ 1 f1 ŝ 3 ŝ 2 f2 ŝ 3 f3 f4 + d 4 International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

7 Cholesky decomposition A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4, V 4 = c 1 f1 + d 1 c 2 s 1 f1 c 2 f2 + d 2 c 3 s 2 s 1 f1 c 3 s 2 f2 c 3 f3 + d 3 s 3 s 2 s 1 f1 s 3 s 2 f2 s 3 f3 f4 + d 4. We compare the diagonal and the main subdiagonal of A and V V T : c k f k + d k = c k+1 s k f k = k 1 (c k s k 1 s j fj ) 2 + (c k fk + d k ) 2 = c 2 k q k + (c k fk + d k ) 2, j=1 k 1 c k c k+1 s k (s k 1 s j fj ) 2 c k+1 s k fk (c k fk + d k ) = c k c k+1 s k q k + c k+1 s k fk (c k fk + d k ), j=1 where The solution is q k := k 1 (s k 1 s k 2 s j fj ) 2. j=1 f k = f k c k q k dk + c k (f k c k q k ), dk = d k, k = 1,..., n. dk + c k (f k c k q k ) International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

8 Algorithm for the Cholesky decomposition If A = Giv(c, s, f) + diag(d) is s.p.d., then A = V V T for V = tril(giv(c, s, f)) + diag( d). function [ f, d] = Cholesky(c, s, f, d) c n = 1 q 1 = 0 for k = 1,..., n : z k = f k c k q k y k = d k + c k z k f k = z k /y k d k = d k /y k q k+1 = s 2 k (q k + f 2 k ) Flops: 11n + O(1). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

9 Algorithm for the Cholesky decomposition If A = Giv(c, s, f) + diag(d) is s.p.d., then A = V V T for V = tril(giv(c, s, f)) + diag( d). function [ f, d] = Cholesky(c, s, f, d) c n = 1 q 1 = 0 for k = 1,..., n : z k = f k c k q k y k = d k + c k z k f k = z k /y k d k = d k /y k q k+1 = s 2 k (q k + f 2 k ) Flops: 11n + O(1). It follows from c k f k + d k = c 2 k q k + (c k fk + d k ) 2 that y k is in fact the diagonal element of V : c k fk + d k = d k + c k (f k c k q k ) = d k + c k z k = y k. A negative or zero value under the square root appears if A is not positive definite. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

10 Equations for V T V The product B = V T V is a s.p.d. DPSS matrix. The lower triangular elements of B are where k = 1,..., n and j > k. b kk = (c k fk + d k ) 2 + (s k fk ) 2, b jk = s k s k+1 s j 1 fk ( f j + c j dj ), B = Giv(ĉ, ŝ, f) + diag(d). For k = 1,..., n 1 we have ŝ 2 f n 2 k k = b 2 jk = f n 2 k p k,, where p k = (s k s k+1 s j 1 ) 2 ( f j + c j dj ) 2. j=k+1 For p k we can apply the recursion p n = 0 and p k = s 2 k j=k+1 ( p k+1 + ( f k+1 + c k+1 dk+1 ) 2). From ĉ k fk + d k = (c k fk + d k ) 2 + (s k fk ) 2 and the Cholesky decomposition we get ĉ k fk = c k z k + (s k fk ) 2. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

11 Equations for V T V The product B = V T V is a s.p.d. DPSS matrix. The lower triangular elements of B are where k = 1,..., n and j > k. b kk = (c k fk + d k ) 2 + (s k fk ) 2, b jk = s k s k+1 s j 1 fk ( f j + c j dj ), B = Giv(ĉ, ŝ, f) + diag(d). For k = 1,..., n 1 we have ŝ 2 f n 2 k k = b 2 jk = f n 2 k p k,, where p k = (s k s k+1 s j 1 ) 2 ( f j + c j dj ) 2. j=k+1 For p k we can apply the recursion p n = 0 and p k = s 2 k j=k+1 ( p k+1 + ( f k+1 + c k+1 dk+1 ) 2). From ĉ k fk + d k = (c k fk + d k ) 2 + (s k fk ) 2 and the Cholesky decomposition we get ĉ k fk = c k z k + (s k fk ) 2. Now, ĉ k, ŝ k, and f k can be computed from ĉ k fk = c k z k + (s k fk ) 2, ŝ k fk = f k pk. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

12 Algorithm for the V T V product Let A = Giv(c, s, f) + diag(d) be s.p.d. = A = V V T, where V = tril(giv(c, s, f)) + diag( d), = B = V T V = Giv(ĉ, ŝ, f) + diag(d). function [ĉ, ŝ, f] = VTV(c, s, f, z) c n = 1 f n = ( f n + d n ) 2 d n p n = 0 for k = n 1,..., 2, 1 p k = s 2 k (p k+1 + ( f k+1 + c k+1 dk+1 ) 2) [ĉ k, ŝ k, f k ] = Givens(c k z k + s 2 k f 2 k, f k pk ) [c, s, f] = Givens(x, y) returns the Givens transformation such that [ c s s c ] [ ] x y = [ ] f. 0 Flops: 16n + O(1) = one step of the zero shift Cholesky LR algorithm: 27n + O(1) flops. As the eigenvalues are invariant to the sign of ŝ k, k = 1,..., n 1, we do not care about it. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

13 Laguerre s shift Let A be a s.p.d. n n matrix with eigenvalues 0 < λ n λ n 1... λ 1. We apply Laguerre s method on the characteristic polynomial f(λ) = det(a λi). If x is an approximation for an eigenvalue of A and S 1 (x) = S 2 (x) = n i=1 n i=1 1 λ i x = f (x) f(x) 1 = f 2 (x) f(x)f (x) (λ i x) 2 f 2 (x) then the next approximation x by Laguerre s method is given by the equation n x = x +. S 1 (x) + (n 1)(nS 2 (x) S 21 (x)) Two important properties of Laguerre s method: if λ n is a simple eigenvalue and if x < λ n then x < x < λ n, the convergence towards λ n is cubic (linear for a multiple eigenvalue). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

14 Laguerre s shift Let A be a s.p.d. n n matrix with eigenvalues 0 < λ n λ n 1... λ 1. We apply Laguerre s method on the characteristic polynomial f(λ) = det(a λi). If x is an approximation for an eigenvalue of A and S 1 (x) = S 2 (x) = n i=1 n i=1 1 λ i x = f (x) f(x) 1 = f 2 (x) f(x)f (x) (λ i x) 2 f 2 (x) = trace((a xi) 1 ), = trace((a xi) 2 ), then the next approximation x by Laguerre s method is given by the equation n x = x +. S 1 (x) + (n 1)(nS 2 (x) S 21 (x)) Two important properties of Laguerre s method: if λ n is a simple eigenvalue and if x < λ n then x < x < λ n, the convergence towards λ n is cubic (linear for a multiple eigenvalue). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

15 Computation of Laguerre s shift We need S 1 (σ) = trace((a σi) 1 ) and S 2 (σ) = trace((a σi) 2 ). If A σi = V V T is the Cholesky decomposition and W = V 1, then S 1 (σ) = trace(w T W ) = W 2 F, S 2 (σ) = trace(w T W W T W ) = trace(w W T W W T ) = W W T 2 F. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

16 Computation of Laguerre s shift We need S 1 (σ) = trace((a σi) 1 ) and S 2 (σ) = trace((a σi) 2 ). If A σi = V V T is the Cholesky decomposition and W = V 1, then S 1 (σ) = trace(w T W ) = W 2 F, S 2 (σ) = trace(w T W W T W ) = trace(w W T W W T ) = W W T 2 F. Delvaux, Van Barel (2004): Let V = tril(giv(c, s, f)) + diag( d) be nonsingular and d i 0 for i = 1,..., n. Then W = V 1 = tril(giv( c, s, f)) + diag( d), where d 1 i = d i for i = 1,..., n. Let us assume that W = tril(giv( c, s, f)) + diag( d). The final algorithm will also be correct when W is not DPSS, which happens when d i = 0 for some i = 2,..., n 1. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

17 Computation of Laguerre s shift A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4. Lemma: If A = Giv(c, s, f) + diag(d) is a symmetric n n DPSS matrix then A 2 F = n (c k f k + d k ) k=1 n 1 k=1 s 2 k f 2 k. Lemma: If W = tril(giv( c, s, f)) + diag( d) and c k 0 for k = 2,..., n 1, then W W T 2 F = W 2 F = ( ) n n 1 (W W T ) 2 kk + 2 (W W T 2 ) k+1,k, k=1 n (W W T ) kk. k=1 k=1 c k+1 International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

18 Algorithm for the Laguerre s shift The following algorithm computes S 1 = W 2 F and S 2 = W W T 2 F. function [S 1, S 2 ] = invtrace(c, s, f, d, y) c n = c n = 1 for k = n 1,..., 2, 1 : r 1 = 0 [ c k, s k ] = Givens(c k c k+1 y k+1, c k+1 s k dk ) for k = 1,..., n 1 β k = c k+1 s k fk /( c k+1 y k y k+1 ) = w k+1,k / c k+1 ω k = c 2 k r k + y 2 k = (W W T ) kk ξ k = c k s k r k + β k /y k = (W W T ) k+1,k / c k+1 r k+1 = s 2 k r k + β 2 k ω n = r n + y 2 n S 1 = n k=1 ω k S 2 = n k=1 ω2 k + 2 n 1 k=1 ξ2 k Flops: 31n + O(1) flops. The algorithm is also correct if d k = 0 for some k = 2,..., n 1. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

19 A 4 = Implementation c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4. If s k is small enough for some k = 1,..., n 1, then we decouple the problem into two smaller problems with matrices A(1 : k, 1 : k) and A(k + 1 : n, k + 1 : n). If s n 1 is small enough, we take f n + d n as an eigenvalue of A and continue with vectors c(1 : n 2), s(1 : n 2), f(1 : n 1), and d(1 : n 1). New initial shift is f n + d n. Even if σ k < λ n, the Cholesky factorization can fail if the difference is too small. This is a problem since Laguerre s shifts converge faster to the smallest eigenvalue than (A k ) nn. One strategy is to relax the shift with a factor τ close to 1, for instance, τ = The computation of Laguerre s shift requires more than half of the operations for one step of the Cholesky LR. We can save work by fixing the shift once the shift improvement is small enough. The algorithm for Laguerre s shift fails if c k = 0 for some k = 2,..., n 1. A simple solution is to perturb c k into some small δ (for instance, ) whenever c k = 0. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

20 Numerical examples Numerical results were obtained with Matlab 7.0 running on a Pentium4 2.6 GHz Windows XP. We compared a Matlab implementation of the Cholesky LR algorithm, a Matlab implementation of the implicit QR algorithm for DPSS matrices [Van Camp, Delvaux, Van Barel, Vandebril, Mastronardi (2005)], the Matlab function eig. Exact eigenvalues were computed in Mathematica 5 using variable precision. The cutoff criterion for both Cholesky LR and implicit QR is With the maximum relative error we denote max 1 i n λ i λ i λ i. When the initial matrix is not DPSS, the reduction into a similar DPSS matrix is done by the algorithm of Vandebril, Van Camp, Van Barel, and Mastronardi (2004). There is a connection between the Lanczos method and the reduction into a similar DPSS matrix which causes that the largest eigenvalues of A are approximated by the lower right diagonal elements of the DPSS matrix. As this is not good for the Cholesky LR method, we reverse the direction of the columns and rows of the DPSS matrix. This can be done in linear time, see, e.g., Vandebril (2004). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

21 Numerical example 1 We take random s.p.d. DPSS matrices of the form A = diag(1,..., n) + triu(uv T, 1) + triu(uv T, 1) T + αi, where u and v are vectors of uniformly distributed random entries on [0, 1], obtained by the Matlab function rand, and the shift α is such that the smallest eigenvalue of A is 1. The condition numbers of these matrices are approximately n. Cholesky LR Implicit QR eig n t steps error t steps error t error International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

22 Numerical example 2 We take symmetric positive definite matrices A = Q diag(1 : n)q T, where Q is a random orthogonal matrix. Now we have to transform the matrix into a similar DPSS matrix before we can apply Cholesky LR or implicit QR. Cholesky LR Implicit QR eig n t steps error t steps error t error : times for Cholesky LR and Implicit QR do not include the reduction to a DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

23 Numerical example 3 We take symmetric positive definite matrices A = Q diag(λ 1,..., λ n )Q T, where Q is a random orthogonal matrix and 8+7 k 1 λ k = 10 n 1. Cholesky LR Implicit QR eig n t steps error t steps error t error : times for Cholesky LR and Implicit QR do not include the reduction to a DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

24 Conclusions Cholesky LR algorithm exploits the structure of s.p.d. DPSS matrices. The method can be combined with Laguerre s shifts. It seems natural to compare the method to the implicit QR for DPSS matrices. In Cholesky LR the eigenvalues are computed from the smallest to the largest eigenvalue, therefore the method is very appropriate for applications where one is interested in few of the smallest eigenvalues. If the complete spectrum is computed, Cholesky LR is more expensive than implicit QR, but, as it tends to be slightly more accurate, it presents an alternative. The proposed method combined with the reduction to DPSS matrices can also be applied to a general s.p.d. matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Orthogonal similarity transformation of a symmetric matrix into a diagonal-plus-semiseparable one with free choice of the diagonal Raf Vandebril, Ellen Van Camp, Marc Van Barel, Nicola Mastronardi Report

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science The Lanczos-Ritz values appearing in an orthogonal similarity reduction of a matrix into semiseparable form M Van Barel R Vandebril N Mastronardi Report TW 36, May 23 Katholieke Universiteit Leuven Department

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Using semiseparable matrices to compute the SVD of a general matrix product/quotient

Using semiseparable matrices to compute the SVD of a general matrix product/quotient Using semiseparable matrices to compute the SVD of a general matrix product/quotient Marc Van Barel Yvette Vanberghen Paul Van Dooren Report TW 508, November 007 n Katholieke Universiteit Leuven Department

More information

Structured Matrix Methods for Polynomial Root-Finding

Structured Matrix Methods for Polynomial Root-Finding Structured Matrix Methods for Polynomial Root-Finding Luca Gemignani Department of Mathematics University of Pisa Largo B. Pontecorvo 5, 5617 Pisa, Italy gemignan@dm.unipi.it ABSTRACT In this paper we

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Direct methods for symmetric eigenvalue problems

Direct methods for symmetric eigenvalue problems Direct methods for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 4, 2008 1 Theoretical background Posing the question Perturbation theory

More information

Eigen-solving via reduction to DPR1 matrices

Eigen-solving via reduction to DPR1 matrices Computers and Mathematics with Applications 56 (2008) 166 171 www.elsevier.com/locate/camwa Eigen-solving via reduction to DPR1 matrices V.Y. Pan a,, B. Murphy a, R.E. Rosholt a, Y. Tang b, X. Wang c,

More information

COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME

COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME GARETH I HARGREAVES Abstract For an n n tridiagonal matrix we exploit the structure of its QR factorization

More information

A multiple shift QR-step for structured rank matrices

A multiple shift QR-step for structured rank matrices A multiple shift QR-step for structured rank matrices Raf Vandebril, Marc Van Barel and Nicola Mastronardi Raf Vandebril Department of Computer Science KU Leuven, Belgium Raf.Vandebril@cs.kuleuven.be Marc

More information

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm References: Trefethen & Bau textbook Eigenvalue problem: given a matrix A, find

More information

Matrix Computations and Semiseparable Matrices

Matrix Computations and Semiseparable Matrices Matrix Computations and Semiseparable Matrices Volume I: Linear Systems Raf Vandebril Department of Computer Science Catholic University of Louvain Marc Van Barel Department of Computer Science Catholic

More information

ECS130 Scientific Computing Handout E February 13, 2017

ECS130 Scientific Computing Handout E February 13, 2017 ECS130 Scientific Computing Handout E February 13, 2017 1. The Power Method (a) Pseudocode: Power Iteration Given an initial vector u 0, t i+1 = Au i u i+1 = t i+1 / t i+1 2 (approximate eigenvector) θ

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science A QR method for computing the singular values via semiseparable matrices. Raf Vandebril Marc Van Barel Nicola Mastronardi Report TW 366, August 2003 Katholieke Universiteit Leuven Department of Computer

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric

More information

Numerical Experiments for Finding Roots of the Polynomials in Chebyshev Basis

Numerical Experiments for Finding Roots of the Polynomials in Chebyshev Basis Available at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 12, Issue 2 (December 2017), pp. 988 1001 Applications and Applied Mathematics: An International Journal (AAM) Numerical Experiments

More information

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue

More information

Fast computation of eigenvalues of companion, comrade, and related matrices

Fast computation of eigenvalues of companion, comrade, and related matrices BIT Numer Math (204) 54:7 30 DOI 0007/s0543-03-0449-x Fast computation of eigenvalues of companion, comrade, and related matrices Jared L Aurentz Raf Vandebril David S Watkins Received: 6 December 202

More information

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n. Lecture # 11 The Power Method for Eigenvalues Part II The power method find the largest (in magnitude) eigenvalue of It makes two assumptions. 1. A is diagonalizable. That is, A R n n. A = XΛX 1 for some

More information

A Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix

A Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix A Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix Nicola Mastronardi Marc Van Barel Raf Vandebril Report TW 461, May 26 n Katholieke Universiteit

More information

Matrices, Moments and Quadrature, cont d

Matrices, Moments and Quadrature, cont d Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Rank structures preserved by the QR-algorithm: the singular case Steven Delvaux Marc Van Barel Report TW 4, August 24 Katholieke Universiteit Leuven Department of Computer Science Celestijnenlaan 2A B-31

More information

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete

More information

Accurate eigenvalue decomposition of arrowhead matrices and applications

Accurate eigenvalue decomposition of arrowhead matrices and applications Accurate eigenvalue decomposition of arrowhead matrices and applications Nevena Jakovčević Stor FESB, University of Split joint work with Ivan Slapničar and Jesse Barlow IWASEP9 June 4th, 2012. 1/31 Introduction

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Logistics Notes for 2016-08-26 1. Our enrollment is at 50, and there are still a few students who want to get in. We only have 50 seats in the room, and I cannot increase the cap further. So if you are

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

Matrix decompositions

Matrix decompositions Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Diagonal Representation of Certain Matrices

Diagonal Representation of Certain Matrices Diagonal Representation of Certain Matrices Mark Tygert Research Report YALEU/DCS/RR-33 December 2, 2004 Abstract An explicit expression is provided for the characteristic polynomial of a matrix M of the

More information

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2 The QR algorithm The most common method for solving small (dense) eigenvalue problems. The basic algorithm: QR without shifts 1. Until Convergence Do: 2. Compute the QR factorization A = QR 3. Set A :=

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Linear algebra & Numerical Analysis

Linear algebra & Numerical Analysis Linear algebra & Numerical Analysis Eigenvalues and Eigenvectors Marta Jarošová http://homel.vsb.cz/~dom033/ Outline Methods computing all eigenvalues Characteristic polynomial Jacobi method for symmetric

More information

Transformations to rank structures by unitary similarity

Transformations to rank structures by unitary similarity Linear Algebra and its Applications 402 (2005) 126 134 www.elsevier.com/locate/laa Transformations to rank structures by unitary similarity Roberto Bevilacqua, Enrico Bozzo, Gianna M. Del Corso Dipartimento

More information

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated. Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Numerical Methods I: Eigenvalues and eigenvectors

Numerical Methods I: Eigenvalues and eigenvectors 1/25 Numerical Methods I: Eigenvalues and eigenvectors Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu November 2, 2017 Overview 2/25 Conditioning Eigenvalues and eigenvectors How hard are they

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Krylov subspace projection methods

Krylov subspace projection methods I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks

More information

Solving linear equations with Gaussian Elimination (I)

Solving linear equations with Gaussian Elimination (I) Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

arxiv: v1 [math.na] 24 Jan 2019

arxiv: v1 [math.na] 24 Jan 2019 ON COMPUTING EFFICIENT DATA-SPARSE REPRESENTATIONS OF UNITARY PLUS LOW-RANK MATRICES R BEVILACQUA, GM DEL CORSO, AND L GEMIGNANI arxiv:190108411v1 [mathna 24 Jan 2019 Abstract Efficient eigenvalue solvers

More information

Solutions Problem Set 8 Math 240, Fall

Solutions Problem Set 8 Math 240, Fall Solutions Problem Set 8 Math 240, Fall 2012 5.6 T/F.2. True. If A is upper or lower diagonal, to make det(a λi) 0, we need product of the main diagonal elements of A λi to be 0, which means λ is one of

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization

More information

Orthogonal iteration to QR

Orthogonal iteration to QR Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Structures preserved by matrix inversion Steven Delvaux Marc Van Barel Report TW 414, December 24 Katholieke Universiteit Leuven Department of Computer Science Celestijnenlaan 2A B-31 Heverlee (Belgium)

More information

Math 504 (Fall 2011) 1. (*) Consider the matrices

Math 504 (Fall 2011) 1. (*) Consider the matrices Math 504 (Fall 2011) Instructor: Emre Mengi Study Guide for Weeks 11-14 This homework concerns the following topics. Basic definitions and facts about eigenvalues and eigenvectors (Trefethen&Bau, Lecture

More information

Chapter 1. Matrix Algebra

Chapter 1. Matrix Algebra ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

EIGENVALUE PROBLEMS (EVP)

EIGENVALUE PROBLEMS (EVP) EIGENVALUE PROBLEMS (EVP) (Golub & Van Loan: Chaps 7-8; Watkins: Chaps 5-7) X.-W Chang and C. C. Paige PART I. EVP THEORY EIGENVALUES AND EIGENVECTORS Let A C n n. Suppose Ax = λx with x 0, then x is a

More information

Compression of unitary rank structured matrices to CMV-like shape with an application to polynomial rootfinding arxiv: v1 [math.

Compression of unitary rank structured matrices to CMV-like shape with an application to polynomial rootfinding arxiv: v1 [math. Compression of unitary rank structured matrices to CMV-like shape with an application to polynomial rootfinding arxiv:1307.186v1 [math.na] 8 Jul 013 Roberto Bevilacqua, Gianna M. Del Corso and Luca Gemignani

More information

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems Mario Arioli m.arioli@rl.ac.uk CCLRC-Rutherford Appleton Laboratory with Daniel Ruiz (E.N.S.E.E.I.H.T)

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Eigenvalue and Eigenvector Problems

Eigenvalue and Eigenvector Problems Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 9, 2008 Today

More information

22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes Numerical Determination of Eigenvalues and Eigenvectors 22.4 Introduction In Section 22. it was shown how to obtain eigenvalues and eigenvectors for low order matrices, 2 2 and. This involved firstly solving

More information

Math Matrix Algebra

Math Matrix Algebra Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:

More information

Fundamentals of Numerical Linear Algebra

Fundamentals of Numerical Linear Algebra Fundamentals of Numerical Linear Algebra Seongjai Kim Department of Mathematics and Statistics Mississippi State University Mississippi State, MS 39762 USA Email: skim@math.msstate.edu Updated: November

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview

More information

632 CHAP. 11 EIGENVALUES AND EIGENVECTORS. QR Method

632 CHAP. 11 EIGENVALUES AND EIGENVECTORS. QR Method 632 CHAP 11 EIGENVALUES AND EIGENVECTORS QR Method Suppose that A is a real symmetric matrix In the preceding section we saw how Householder s method is used to construct a similar tridiagonal matrix The

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial

More information

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 19, 2010 Today

More information

Matrix Theory, Math6304 Lecture Notes from October 25, 2012

Matrix Theory, Math6304 Lecture Notes from October 25, 2012 Matrix Theory, Math6304 Lecture Notes from October 25, 2012 taken by John Haas Last Time (10/23/12) Example of Low Rank Perturbation Relationship Between Eigenvalues and Principal Submatrices: We started

More information

A factorization of the inverse of the shifted companion matrix

A factorization of the inverse of the shifted companion matrix Electronic Journal of Linear Algebra Volume 26 Volume 26 (203) Article 8 203 A factorization of the inverse of the shifted companion matrix Jared L Aurentz jaurentz@mathwsuedu Follow this and additional

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 208 LECTURE 3 need for pivoting we saw that under proper circumstances, we can write A LU where 0 0 0 u u 2 u n l 2 0 0 0 u 22 u 2n L l 3 l 32, U 0 0 0 l n l

More information

A CHOLESKY LR ALGORITHM FOR THE POSITIVE DEFINITE SYMMETRIC DIAGONAL-PLUS-SEMISEPARABLE EIGENPROBLEM

A CHOLESKY LR ALGORITHM FOR THE POSITIVE DEFINITE SYMMETRIC DIAGONAL-PLUS-SEMISEPARABLE EIGENPROBLEM A CHOLESKY LR ALGORITHM FOR THE POSITIVE DEFINITE SYMMETRIC DIAGONAL-PLUS-SEMISEPARABLE EIGENPROBLEM BOR PLESTENJAK, ELLEN VAN CAMP, AND MARC VAN BAREL Abstract. We preset a Cholesky LR algorithm with

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Improved initialization of the accelerated and robust QR-like polynomial root-finding

Improved initialization of the accelerated and robust QR-like polynomial root-finding Improved initialization of the accelerated and robust QR-like polynomial root-finding Dario A. Bini, Luca Gemignani, and Victor Y. Pan Abstract. We approximate polynomial roots numerically as the eigenvalues

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Gaussian Elimination without/with Pivoting and Cholesky Decomposition

Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination WITHOUT pivoting Notation: For a matrix A R n n we define for k {,,n} the leading principal submatrix a a k A

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

Matrices and Linear Algebra

Matrices and Linear Algebra Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2

More information

arxiv: v2 [math.na] 21 Sep 2015

arxiv: v2 [math.na] 21 Sep 2015 Forward stable eigenvalue decomposition of rank-one modifications of diagonal matrices arxiv:1405.7537v2 [math.na] 21 Sep 2015 N. Jakovčević Stor a,1,, I. Slapničar a,1, J. L. Barlow b,2 a Faculty of Electrical

More information

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 8, 2009 Today

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

R-Linear Convergence of Limited Memory Steepest Descent

R-Linear Convergence of Limited Memory Steepest Descent R-Linear Convergence of Limited Memory Steepest Descent Frank E. Curtis, Lehigh University joint work with Wei Guo, Lehigh University OP17 Vancouver, British Columbia, Canada 24 May 2017 R-Linear Convergence

More information

Math 577 Assignment 7

Math 577 Assignment 7 Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October Finding normalized and modularity cuts by spectral clustering Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics marib@math.bme.hu Ljubjana 2010, October Outline Find

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

EE364a Review Session 7

EE364a Review Session 7 EE364a Review Session 7 EE364a Review session outline: derivatives and chain rule (Appendix A.4) numerical linear algebra (Appendix C) factor and solve method exploiting structure and sparsity 1 Derivative

More information

Solving large scale eigenvalue problems

Solving large scale eigenvalue problems arge scale eigenvalue problems, Lecture 4, March 14, 2018 1/41 Lecture 4, March 14, 2018: The QR algorithm http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich E-mail:

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information