A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem
|
|
- Dennis Beasley
- 6 years ago
- Views:
Transcription
1 A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van Barel Department of Computer Science Katholieke Universiteit Leuven Belgium International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
2 Outline Introduction Cholesky LR algorithm Laguerre s shift Implementation Numerical examples Conclusions International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
3 Introduction semiseparable: every submatrix from the lower or upper triangular part has rank at most 1. diagonal-plus-semiseparable (DPSS): the sum D + S of a diagonal D and a semiseparable S. Givens-vector representation of a DPSS matrix is based on a vector f = [f 1,..., f n ] T, Givens [ ] ci s rotations G i = i, i = 1,..., n 1, and a diagonal d = [d 1,..., d n ] T as s i c i c 1 f 1 + d 1 c 2 s 1 f 1 c n 1 s n 2:1 f 1 s n 1:1 f 1 c 2 s 1 f 1 c 2 f 2 + d 2 c n 1 s n 2:1 f 2 s n 1:2 f 2 D + S =....., c n 1 s n 2:1 f 1 c n 1 s n 2:2 f 2 c n 1 f n 1 + d n 1 s n 1 f n 1 s n 1:1 f 1 s n 1:2 f 2 s n 1 f n 1 f n + d n where s a:b = s a s a 1 s b. When they appear, we assume that c n = 1 and s n = 0. We denote D + S = diag(d) + Giv(c, s, f). Our goal is to compute the smallest (or all) eigenvalues of a s.p.d. DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
4 Motivation Let A be an n n symmetric matrix. One can find an orthogonal matrix Q such that B = Q T AQ is tridiagonal in O(n 3 ) flops. Next, the eigendecomposition of B is computed in O(n 2 ) flops. Instead of the tridiagonal structure one can use DPSS matrices. Vandebril, Van Camp, Van Barel, Mastronardi (2004): For an arbitrary diagonal matrix D there exists an orthogonal matrix Q such that Q T AQ = D+S is a DPSS matrix. The reduction is O(n 2 ) more expensive than in the tridiagonal case. The reduction algorithm has a Lanczos-Ritz convergence behaviour and performs a kind of nested subspace iteration at each step. A good choice of the diagonal can compensate slower reduction. Algorithms for the eigendecomposition of a symmetric DPSS matrix: Chandrasekaran, Gu (2004), in Numer. Math.: divide and conquer, Mastronardi, Van Camp, Van Barel (2003), tech. report: divide and conquer, Bini, Gemignani, Pan (2003), tech. report: QR algorithm, Fasino (2004), tech. report: QR algorithm, Van Camp, Delvaux, Van Barel, Vandebril, Mastronardi (2005), tech. report: implicit QR algorithm, Mastronardi, Van Camp, Van Barel, Vandebril (2004), tech. report: computation of eigenvectors International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
5 Cholesky LR algorithm Let A be a s.p.d. matrix. A 0 = A k = 0, 1, 2,... choose shift σ k A k σ k I = V k V T k A k+1 = V T k V k + σ k I (Cholesky decomposition, V k is lower-triangular) Two steps of the zero shift Cholesky LR are equivalent to one step of the zero shift QR. The shift σ k should be such that A k σ k I is positive definite. When applied to a s.p.d. DPSS matrix D + S, the shift can be included into the diagonal part. Grad, Zakrajšek (1972): Cholesky LR with Laguerre s shifts for the symmetric tridiagonal matrices. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
6 DPSS is invariant to Cholesky LR Theorem: Let A = Giv(c, s, f) + diag(d) be a s.p.d DPSS matrix. 1. If A = V V T is the Cholesky decomposition of A, then V = tril(giv(c, s, f)) + diag( d), 2. If B = V T V, then B is a s.p.d. DPSS matrix B = Giv(ĉ, ŝ, f) + diag(d). A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3. V 4 = B 4 = s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4 c 1 f1 + d 1 c 2 s 1 f1 c 2 f2 + d 2 c 3 s 2 s 1 f1 c 3 s 2 f2 c 3 f3 + d 3 s 3 s 2 s 1 f1 s 3 s 2 f2 s 3 f3 f4 + d 4 ĉ 1 f1 + d 1 ĉ 2 ŝ 1 f1 ĉ 2 f2 + d 2 ĉ 3 ŝ 2 ŝ 1 f1 ĉ 3 ŝ 2 f2 ĉ 3 f3 + d 3.. ŝ 3 ŝ 2 ŝ 1 f1 ŝ 3 ŝ 2 f2 ŝ 3 f3 f4 + d 4 International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
7 Cholesky decomposition A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4, V 4 = c 1 f1 + d 1 c 2 s 1 f1 c 2 f2 + d 2 c 3 s 2 s 1 f1 c 3 s 2 f2 c 3 f3 + d 3 s 3 s 2 s 1 f1 s 3 s 2 f2 s 3 f3 f4 + d 4. We compare the diagonal and the main subdiagonal of A and V V T : c k f k + d k = c k+1 s k f k = k 1 (c k s k 1 s j fj ) 2 + (c k fk + d k ) 2 = c 2 k q k + (c k fk + d k ) 2, j=1 k 1 c k c k+1 s k (s k 1 s j fj ) 2 c k+1 s k fk (c k fk + d k ) = c k c k+1 s k q k + c k+1 s k fk (c k fk + d k ), j=1 where The solution is q k := k 1 (s k 1 s k 2 s j fj ) 2. j=1 f k = f k c k q k dk + c k (f k c k q k ), dk = d k, k = 1,..., n. dk + c k (f k c k q k ) International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
8 Algorithm for the Cholesky decomposition If A = Giv(c, s, f) + diag(d) is s.p.d., then A = V V T for V = tril(giv(c, s, f)) + diag( d). function [ f, d] = Cholesky(c, s, f, d) c n = 1 q 1 = 0 for k = 1,..., n : z k = f k c k q k y k = d k + c k z k f k = z k /y k d k = d k /y k q k+1 = s 2 k (q k + f 2 k ) Flops: 11n + O(1). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
9 Algorithm for the Cholesky decomposition If A = Giv(c, s, f) + diag(d) is s.p.d., then A = V V T for V = tril(giv(c, s, f)) + diag( d). function [ f, d] = Cholesky(c, s, f, d) c n = 1 q 1 = 0 for k = 1,..., n : z k = f k c k q k y k = d k + c k z k f k = z k /y k d k = d k /y k q k+1 = s 2 k (q k + f 2 k ) Flops: 11n + O(1). It follows from c k f k + d k = c 2 k q k + (c k fk + d k ) 2 that y k is in fact the diagonal element of V : c k fk + d k = d k + c k (f k c k q k ) = d k + c k z k = y k. A negative or zero value under the square root appears if A is not positive definite. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
10 Equations for V T V The product B = V T V is a s.p.d. DPSS matrix. The lower triangular elements of B are where k = 1,..., n and j > k. b kk = (c k fk + d k ) 2 + (s k fk ) 2, b jk = s k s k+1 s j 1 fk ( f j + c j dj ), B = Giv(ĉ, ŝ, f) + diag(d). For k = 1,..., n 1 we have ŝ 2 f n 2 k k = b 2 jk = f n 2 k p k,, where p k = (s k s k+1 s j 1 ) 2 ( f j + c j dj ) 2. j=k+1 For p k we can apply the recursion p n = 0 and p k = s 2 k j=k+1 ( p k+1 + ( f k+1 + c k+1 dk+1 ) 2). From ĉ k fk + d k = (c k fk + d k ) 2 + (s k fk ) 2 and the Cholesky decomposition we get ĉ k fk = c k z k + (s k fk ) 2. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
11 Equations for V T V The product B = V T V is a s.p.d. DPSS matrix. The lower triangular elements of B are where k = 1,..., n and j > k. b kk = (c k fk + d k ) 2 + (s k fk ) 2, b jk = s k s k+1 s j 1 fk ( f j + c j dj ), B = Giv(ĉ, ŝ, f) + diag(d). For k = 1,..., n 1 we have ŝ 2 f n 2 k k = b 2 jk = f n 2 k p k,, where p k = (s k s k+1 s j 1 ) 2 ( f j + c j dj ) 2. j=k+1 For p k we can apply the recursion p n = 0 and p k = s 2 k j=k+1 ( p k+1 + ( f k+1 + c k+1 dk+1 ) 2). From ĉ k fk + d k = (c k fk + d k ) 2 + (s k fk ) 2 and the Cholesky decomposition we get ĉ k fk = c k z k + (s k fk ) 2. Now, ĉ k, ŝ k, and f k can be computed from ĉ k fk = c k z k + (s k fk ) 2, ŝ k fk = f k pk. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
12 Algorithm for the V T V product Let A = Giv(c, s, f) + diag(d) be s.p.d. = A = V V T, where V = tril(giv(c, s, f)) + diag( d), = B = V T V = Giv(ĉ, ŝ, f) + diag(d). function [ĉ, ŝ, f] = VTV(c, s, f, z) c n = 1 f n = ( f n + d n ) 2 d n p n = 0 for k = n 1,..., 2, 1 p k = s 2 k (p k+1 + ( f k+1 + c k+1 dk+1 ) 2) [ĉ k, ŝ k, f k ] = Givens(c k z k + s 2 k f 2 k, f k pk ) [c, s, f] = Givens(x, y) returns the Givens transformation such that [ c s s c ] [ ] x y = [ ] f. 0 Flops: 16n + O(1) = one step of the zero shift Cholesky LR algorithm: 27n + O(1) flops. As the eigenvalues are invariant to the sign of ŝ k, k = 1,..., n 1, we do not care about it. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
13 Laguerre s shift Let A be a s.p.d. n n matrix with eigenvalues 0 < λ n λ n 1... λ 1. We apply Laguerre s method on the characteristic polynomial f(λ) = det(a λi). If x is an approximation for an eigenvalue of A and S 1 (x) = S 2 (x) = n i=1 n i=1 1 λ i x = f (x) f(x) 1 = f 2 (x) f(x)f (x) (λ i x) 2 f 2 (x) then the next approximation x by Laguerre s method is given by the equation n x = x +. S 1 (x) + (n 1)(nS 2 (x) S 21 (x)) Two important properties of Laguerre s method: if λ n is a simple eigenvalue and if x < λ n then x < x < λ n, the convergence towards λ n is cubic (linear for a multiple eigenvalue). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
14 Laguerre s shift Let A be a s.p.d. n n matrix with eigenvalues 0 < λ n λ n 1... λ 1. We apply Laguerre s method on the characteristic polynomial f(λ) = det(a λi). If x is an approximation for an eigenvalue of A and S 1 (x) = S 2 (x) = n i=1 n i=1 1 λ i x = f (x) f(x) 1 = f 2 (x) f(x)f (x) (λ i x) 2 f 2 (x) = trace((a xi) 1 ), = trace((a xi) 2 ), then the next approximation x by Laguerre s method is given by the equation n x = x +. S 1 (x) + (n 1)(nS 2 (x) S 21 (x)) Two important properties of Laguerre s method: if λ n is a simple eigenvalue and if x < λ n then x < x < λ n, the convergence towards λ n is cubic (linear for a multiple eigenvalue). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
15 Computation of Laguerre s shift We need S 1 (σ) = trace((a σi) 1 ) and S 2 (σ) = trace((a σi) 2 ). If A σi = V V T is the Cholesky decomposition and W = V 1, then S 1 (σ) = trace(w T W ) = W 2 F, S 2 (σ) = trace(w T W W T W ) = trace(w W T W W T ) = W W T 2 F. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
16 Computation of Laguerre s shift We need S 1 (σ) = trace((a σi) 1 ) and S 2 (σ) = trace((a σi) 2 ). If A σi = V V T is the Cholesky decomposition and W = V 1, then S 1 (σ) = trace(w T W ) = W 2 F, S 2 (σ) = trace(w T W W T W ) = trace(w W T W W T ) = W W T 2 F. Delvaux, Van Barel (2004): Let V = tril(giv(c, s, f)) + diag( d) be nonsingular and d i 0 for i = 1,..., n. Then W = V 1 = tril(giv( c, s, f)) + diag( d), where d 1 i = d i for i = 1,..., n. Let us assume that W = tril(giv( c, s, f)) + diag( d). The final algorithm will also be correct when W is not DPSS, which happens when d i = 0 for some i = 2,..., n 1. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
17 Computation of Laguerre s shift A 4 = c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4. Lemma: If A = Giv(c, s, f) + diag(d) is a symmetric n n DPSS matrix then A 2 F = n (c k f k + d k ) k=1 n 1 k=1 s 2 k f 2 k. Lemma: If W = tril(giv( c, s, f)) + diag( d) and c k 0 for k = 2,..., n 1, then W W T 2 F = W 2 F = ( ) n n 1 (W W T ) 2 kk + 2 (W W T 2 ) k+1,k, k=1 n (W W T ) kk. k=1 k=1 c k+1 International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
18 Algorithm for the Laguerre s shift The following algorithm computes S 1 = W 2 F and S 2 = W W T 2 F. function [S 1, S 2 ] = invtrace(c, s, f, d, y) c n = c n = 1 for k = n 1,..., 2, 1 : r 1 = 0 [ c k, s k ] = Givens(c k c k+1 y k+1, c k+1 s k dk ) for k = 1,..., n 1 β k = c k+1 s k fk /( c k+1 y k y k+1 ) = w k+1,k / c k+1 ω k = c 2 k r k + y 2 k = (W W T ) kk ξ k = c k s k r k + β k /y k = (W W T ) k+1,k / c k+1 r k+1 = s 2 k r k + β 2 k ω n = r n + y 2 n S 1 = n k=1 ω k S 2 = n k=1 ω2 k + 2 n 1 k=1 ξ2 k Flops: 31n + O(1) flops. The algorithm is also correct if d k = 0 for some k = 2,..., n 1. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
19 A 4 = Implementation c 1 f 1 + d 1 c 2 s 1 f 1 c 2 f 2 + d 2 c 3 s 2 s 1 f 1 c 3 s 2 f 2 c 3 f 3 + d 3 s 3 s 2 s 1 f 1 s 3 s 2 f 2 s 3 f 3 f 4 + d 4. If s k is small enough for some k = 1,..., n 1, then we decouple the problem into two smaller problems with matrices A(1 : k, 1 : k) and A(k + 1 : n, k + 1 : n). If s n 1 is small enough, we take f n + d n as an eigenvalue of A and continue with vectors c(1 : n 2), s(1 : n 2), f(1 : n 1), and d(1 : n 1). New initial shift is f n + d n. Even if σ k < λ n, the Cholesky factorization can fail if the difference is too small. This is a problem since Laguerre s shifts converge faster to the smallest eigenvalue than (A k ) nn. One strategy is to relax the shift with a factor τ close to 1, for instance, τ = The computation of Laguerre s shift requires more than half of the operations for one step of the Cholesky LR. We can save work by fixing the shift once the shift improvement is small enough. The algorithm for Laguerre s shift fails if c k = 0 for some k = 2,..., n 1. A simple solution is to perturb c k into some small δ (for instance, ) whenever c k = 0. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
20 Numerical examples Numerical results were obtained with Matlab 7.0 running on a Pentium4 2.6 GHz Windows XP. We compared a Matlab implementation of the Cholesky LR algorithm, a Matlab implementation of the implicit QR algorithm for DPSS matrices [Van Camp, Delvaux, Van Barel, Vandebril, Mastronardi (2005)], the Matlab function eig. Exact eigenvalues were computed in Mathematica 5 using variable precision. The cutoff criterion for both Cholesky LR and implicit QR is With the maximum relative error we denote max 1 i n λ i λ i λ i. When the initial matrix is not DPSS, the reduction into a similar DPSS matrix is done by the algorithm of Vandebril, Van Camp, Van Barel, and Mastronardi (2004). There is a connection between the Lanczos method and the reduction into a similar DPSS matrix which causes that the largest eigenvalues of A are approximated by the lower right diagonal elements of the DPSS matrix. As this is not good for the Cholesky LR method, we reverse the direction of the columns and rows of the DPSS matrix. This can be done in linear time, see, e.g., Vandebril (2004). International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
21 Numerical example 1 We take random s.p.d. DPSS matrices of the form A = diag(1,..., n) + triu(uv T, 1) + triu(uv T, 1) T + αi, where u and v are vectors of uniformly distributed random entries on [0, 1], obtained by the Matlab function rand, and the shift α is such that the smallest eigenvalue of A is 1. The condition numbers of these matrices are approximately n. Cholesky LR Implicit QR eig n t steps error t steps error t error International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
22 Numerical example 2 We take symmetric positive definite matrices A = Q diag(1 : n)q T, where Q is a random orthogonal matrix. Now we have to transform the matrix into a similar DPSS matrix before we can apply Cholesky LR or implicit QR. Cholesky LR Implicit QR eig n t steps error t steps error t error : times for Cholesky LR and Implicit QR do not include the reduction to a DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
23 Numerical example 3 We take symmetric positive definite matrices A = Q diag(λ 1,..., λ n )Q T, where Q is a random orthogonal matrix and 8+7 k 1 λ k = 10 n 1. Cholesky LR Implicit QR eig n t steps error t steps error t error : times for Cholesky LR and Implicit QR do not include the reduction to a DPSS matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
24 Conclusions Cholesky LR algorithm exploits the structure of s.p.d. DPSS matrices. The method can be combined with Laguerre s shifts. It seems natural to compare the method to the implicit QR for DPSS matrices. In Cholesky LR the eigenvalues are computed from the smallest to the largest eigenvalue, therefore the method is very appropriate for applications where one is interested in few of the smallest eigenvalues. If the complete spectrum is computed, Cholesky LR is more expensive than implicit QR, but, as it tends to be slightly more accurate, it presents an alternative. The proposed method combined with the reduction to DPSS matrices can also be applied to a general s.p.d. matrix. International Conference on Matrix Methods and Operator Equations, Moscow, June 20-25,
Katholieke Universiteit Leuven Department of Computer Science
Orthogonal similarity transformation of a symmetric matrix into a diagonal-plus-semiseparable one with free choice of the diagonal Raf Vandebril, Ellen Van Camp, Marc Van Barel, Nicola Mastronardi Report
More informationKatholieke Universiteit Leuven Department of Computer Science
The Lanczos-Ritz values appearing in an orthogonal similarity reduction of a matrix into semiseparable form M Van Barel R Vandebril N Mastronardi Report TW 36, May 23 Katholieke Universiteit Leuven Department
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationUsing semiseparable matrices to compute the SVD of a general matrix product/quotient
Using semiseparable matrices to compute the SVD of a general matrix product/quotient Marc Van Barel Yvette Vanberghen Paul Van Dooren Report TW 508, November 007 n Katholieke Universiteit Leuven Department
More informationStructured Matrix Methods for Polynomial Root-Finding
Structured Matrix Methods for Polynomial Root-Finding Luca Gemignani Department of Mathematics University of Pisa Largo B. Pontecorvo 5, 5617 Pisa, Italy gemignan@dm.unipi.it ABSTRACT In this paper we
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationDirect methods for symmetric eigenvalue problems
Direct methods for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 4, 2008 1 Theoretical background Posing the question Perturbation theory
More informationEigen-solving via reduction to DPR1 matrices
Computers and Mathematics with Applications 56 (2008) 166 171 www.elsevier.com/locate/camwa Eigen-solving via reduction to DPR1 matrices V.Y. Pan a,, B. Murphy a, R.E. Rosholt a, Y. Tang b, X. Wang c,
More informationCOMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME
COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME GARETH I HARGREAVES Abstract For an n n tridiagonal matrix we exploit the structure of its QR factorization
More informationA multiple shift QR-step for structured rank matrices
A multiple shift QR-step for structured rank matrices Raf Vandebril, Marc Van Barel and Nicola Mastronardi Raf Vandebril Department of Computer Science KU Leuven, Belgium Raf.Vandebril@cs.kuleuven.be Marc
More informationMath 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm
Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm References: Trefethen & Bau textbook Eigenvalue problem: given a matrix A, find
More informationMatrix Computations and Semiseparable Matrices
Matrix Computations and Semiseparable Matrices Volume I: Linear Systems Raf Vandebril Department of Computer Science Catholic University of Louvain Marc Van Barel Department of Computer Science Catholic
More informationECS130 Scientific Computing Handout E February 13, 2017
ECS130 Scientific Computing Handout E February 13, 2017 1. The Power Method (a) Pseudocode: Power Iteration Given an initial vector u 0, t i+1 = Au i u i+1 = t i+1 / t i+1 2 (approximate eigenvector) θ
More informationKatholieke Universiteit Leuven Department of Computer Science
A QR method for computing the singular values via semiseparable matrices. Raf Vandebril Marc Van Barel Nicola Mastronardi Report TW 366, August 2003 Katholieke Universiteit Leuven Department of Computer
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationNumerical Experiments for Finding Roots of the Polynomials in Chebyshev Basis
Available at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 12, Issue 2 (December 2017), pp. 988 1001 Applications and Applied Mathematics: An International Journal (AAM) Numerical Experiments
More informationEIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4
EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue
More informationFast computation of eigenvalues of companion, comrade, and related matrices
BIT Numer Math (204) 54:7 30 DOI 0007/s0543-03-0449-x Fast computation of eigenvalues of companion, comrade, and related matrices Jared L Aurentz Raf Vandebril David S Watkins Received: 6 December 202
More informationLecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.
Lecture # 11 The Power Method for Eigenvalues Part II The power method find the largest (in magnitude) eigenvalue of It makes two assumptions. 1. A is diagonalizable. That is, A R n n. A = XΛX 1 for some
More informationA Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix
A Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix Nicola Mastronardi Marc Van Barel Raf Vandebril Report TW 461, May 26 n Katholieke Universiteit
More informationMatrices, Moments and Quadrature, cont d
Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationKatholieke Universiteit Leuven Department of Computer Science
Rank structures preserved by the QR-algorithm: the singular case Steven Delvaux Marc Van Barel Report TW 4, August 24 Katholieke Universiteit Leuven Department of Computer Science Celestijnenlaan 2A B-31
More informationSection 4.5 Eigenvalues of Symmetric Tridiagonal Matrices
Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete
More informationAccurate eigenvalue decomposition of arrowhead matrices and applications
Accurate eigenvalue decomposition of arrowhead matrices and applications Nevena Jakovčević Stor FESB, University of Split joint work with Ivan Slapničar and Jesse Barlow IWASEP9 June 4th, 2012. 1/31 Introduction
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Logistics Notes for 2016-08-26 1. Our enrollment is at 50, and there are still a few students who want to get in. We only have 50 seats in the room, and I cannot increase the cap further. So if you are
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationMatrix decompositions
Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationDiagonal Representation of Certain Matrices
Diagonal Representation of Certain Matrices Mark Tygert Research Report YALEU/DCS/RR-33 December 2, 2004 Abstract An explicit expression is provided for the characteristic polynomial of a matrix M of the
More information13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2
The QR algorithm The most common method for solving small (dense) eigenvalue problems. The basic algorithm: QR without shifts 1. Until Convergence Do: 2. Compute the QR factorization A = QR 3. Set A :=
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationLinear algebra & Numerical Analysis
Linear algebra & Numerical Analysis Eigenvalues and Eigenvectors Marta Jarošová http://homel.vsb.cz/~dom033/ Outline Methods computing all eigenvalues Characteristic polynomial Jacobi method for symmetric
More informationTransformations to rank structures by unitary similarity
Linear Algebra and its Applications 402 (2005) 126 134 www.elsevier.com/locate/laa Transformations to rank structures by unitary similarity Roberto Bevilacqua, Enrico Bozzo, Gianna M. Del Corso Dipartimento
More informationComputation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.
Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationNumerical Methods I: Eigenvalues and eigenvectors
1/25 Numerical Methods I: Eigenvalues and eigenvectors Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu November 2, 2017 Overview 2/25 Conditioning Eigenvalues and eigenvectors How hard are they
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationOn prescribing Ritz values and GMRES residual norms generated by Arnoldi processes
On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationKrylov subspace projection methods
I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks
More informationSolving linear equations with Gaussian Elimination (I)
Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationarxiv: v1 [math.na] 24 Jan 2019
ON COMPUTING EFFICIENT DATA-SPARSE REPRESENTATIONS OF UNITARY PLUS LOW-RANK MATRICES R BEVILACQUA, GM DEL CORSO, AND L GEMIGNANI arxiv:190108411v1 [mathna 24 Jan 2019 Abstract Efficient eigenvalue solvers
More informationSolutions Problem Set 8 Math 240, Fall
Solutions Problem Set 8 Math 240, Fall 2012 5.6 T/F.2. True. If A is upper or lower diagonal, to make det(a λi) 0, we need product of the main diagonal elements of A λi to be 0, which means λ is one of
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations
More informationDirect Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le
Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization
More informationOrthogonal iteration to QR
Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation
More informationKatholieke Universiteit Leuven Department of Computer Science
Structures preserved by matrix inversion Steven Delvaux Marc Van Barel Report TW 414, December 24 Katholieke Universiteit Leuven Department of Computer Science Celestijnenlaan 2A B-31 Heverlee (Belgium)
More informationMath 504 (Fall 2011) 1. (*) Consider the matrices
Math 504 (Fall 2011) Instructor: Emre Mengi Study Guide for Weeks 11-14 This homework concerns the following topics. Basic definitions and facts about eigenvalues and eigenvectors (Trefethen&Bau, Lecture
More informationChapter 1. Matrix Algebra
ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface
More informationMATH 221, Spring Homework 10 Solutions
MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the
More informationEIGENVALUE PROBLEMS (EVP)
EIGENVALUE PROBLEMS (EVP) (Golub & Van Loan: Chaps 7-8; Watkins: Chaps 5-7) X.-W Chang and C. C. Paige PART I. EVP THEORY EIGENVALUES AND EIGENVECTORS Let A C n n. Suppose Ax = λx with x 0, then x is a
More informationCompression of unitary rank structured matrices to CMV-like shape with an application to polynomial rootfinding arxiv: v1 [math.
Compression of unitary rank structured matrices to CMV-like shape with an application to polynomial rootfinding arxiv:1307.186v1 [math.na] 8 Jul 013 Roberto Bevilacqua, Gianna M. Del Corso and Luca Gemignani
More informationQR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR
QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationA Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems
A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems Mario Arioli m.arioli@rl.ac.uk CCLRC-Rutherford Appleton Laboratory with Daniel Ruiz (E.N.S.E.E.I.H.T)
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationEigenvalue and Eigenvector Problems
Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems
More informationScientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix
Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008
More informationComputing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices
Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 9, 2008 Today
More information22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes
Numerical Determination of Eigenvalues and Eigenvectors 22.4 Introduction In Section 22. it was shown how to obtain eigenvalues and eigenvectors for low order matrices, 2 2 and. This involved firstly solving
More informationMath Matrix Algebra
Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:
More informationFundamentals of Numerical Linear Algebra
Fundamentals of Numerical Linear Algebra Seongjai Kim Department of Mathematics and Statistics Mississippi State University Mississippi State, MS 39762 USA Email: skim@math.msstate.edu Updated: November
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview
More information632 CHAP. 11 EIGENVALUES AND EIGENVECTORS. QR Method
632 CHAP 11 EIGENVALUES AND EIGENVECTORS QR Method Suppose that A is a real symmetric matrix In the preceding section we saw how Householder s method is used to construct a similar tridiagonal matrix The
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationRemark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial
More informationComputing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm
Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 19, 2010 Today
More informationMatrix Theory, Math6304 Lecture Notes from October 25, 2012
Matrix Theory, Math6304 Lecture Notes from October 25, 2012 taken by John Haas Last Time (10/23/12) Example of Low Rank Perturbation Relationship Between Eigenvalues and Principal Submatrices: We started
More informationA factorization of the inverse of the shifted companion matrix
Electronic Journal of Linear Algebra Volume 26 Volume 26 (203) Article 8 203 A factorization of the inverse of the shifted companion matrix Jared L Aurentz jaurentz@mathwsuedu Follow this and additional
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13
STAT 309: MATHEMATICAL COMPUTATIONS I FALL 208 LECTURE 3 need for pivoting we saw that under proper circumstances, we can write A LU where 0 0 0 u u 2 u n l 2 0 0 0 u 22 u 2n L l 3 l 32, U 0 0 0 l n l
More informationA CHOLESKY LR ALGORITHM FOR THE POSITIVE DEFINITE SYMMETRIC DIAGONAL-PLUS-SEMISEPARABLE EIGENPROBLEM
A CHOLESKY LR ALGORITHM FOR THE POSITIVE DEFINITE SYMMETRIC DIAGONAL-PLUS-SEMISEPARABLE EIGENPROBLEM BOR PLESTENJAK, ELLEN VAN CAMP, AND MARC VAN BAREL Abstract. We preset a Cholesky LR algorithm with
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationImproved initialization of the accelerated and robust QR-like polynomial root-finding
Improved initialization of the accelerated and robust QR-like polynomial root-finding Dario A. Bini, Luca Gemignani, and Victor Y. Pan Abstract. We approximate polynomial roots numerically as the eigenvalues
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationGaussian Elimination without/with Pivoting and Cholesky Decomposition
Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination WITHOUT pivoting Notation: For a matrix A R n n we define for k {,,n} the leading principal submatrix a a k A
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationMatrices and Linear Algebra
Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2
More informationarxiv: v2 [math.na] 21 Sep 2015
Forward stable eigenvalue decomposition of rank-one modifications of diagonal matrices arxiv:1405.7537v2 [math.na] 21 Sep 2015 N. Jakovčević Stor a,1,, I. Slapničar a,1, J. L. Barlow b,2 a Faculty of Electrical
More informationComputing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices
Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 8, 2009 Today
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationR-Linear Convergence of Limited Memory Steepest Descent
R-Linear Convergence of Limited Memory Steepest Descent Frank E. Curtis, Lehigh University joint work with Wei Guo, Lehigh University OP17 Vancouver, British Columbia, Canada 24 May 2017 R-Linear Convergence
More informationMath 577 Assignment 7
Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationFinding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October
Finding normalized and modularity cuts by spectral clustering Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics marib@math.bme.hu Ljubjana 2010, October Outline Find
More informationMATH 304 Linear Algebra Lecture 34: Review for Test 2.
MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationEE364a Review Session 7
EE364a Review Session 7 EE364a Review session outline: derivatives and chain rule (Appendix A.4) numerical linear algebra (Appendix C) factor and solve method exploiting structure and sparsity 1 Derivative
More informationSolving large scale eigenvalue problems
arge scale eigenvalue problems, Lecture 4, March 14, 2018 1/41 Lecture 4, March 14, 2018: The QR algorithm http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich E-mail:
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to
More informationIncomplete Cholesky preconditioners that exploit the low-rank property
anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More information