Cuppen s Divide and Conquer Algorithm

Size: px
Start display at page:

Download "Cuppen s Divide and Conquer Algorithm"

Transcription

1 Chapter 4 Cuppen s Divide and Conquer Algorithm In this chapter we deal with an algorithm that is designed for the efficient solution of the symmetric tridiagonal eigenvalue problem a b (4) x λx, b a bn We noticed from able 3 that the reduction of a full symmetric matrix to a similar tridiagonal matrix requires about 8 3 n3 while the tridiagonal QR algorithm needs an estimated 6n 3 floating operations (flops) to converge Because of the importance of this subproblem a considerable effort has been put into finding faster algorithms than the QR algorithms to solve the tridiagonal eigenvalue problem In the mid-98 s Dongarra and Sorensen [4] promoted an algorithm originally proposed by Cuppen [] his algorithm was based on a divide and conquer strategy However, it took ten more years until a stable variant was found by Gu and Eisenstat [5, 6] oday, a stable implementation of this latter algorithm is available in LAPACK [] 4 he divide and conquer idea Divide and conquer is an old strategy in military to defeat an enemy going back at least to Caesar In computer science, divide and conquer (D&C) is an important algorithm design paradigm It works by recursively breaking down a problem into two or more subproblems of the same (or related) type, until these become simple enough to be solved directly he solutions to the subproblems are then combined to give a solution to the original problem ranslated to our problem the strategy becomes b n Partition the tridiagonal eigenvalue problem into two (or more) smaller tridiagonal eigenvalue problems Solve the two smaller problems 3 Combine the solutions of the smaller problems to get the desired solution of the overall problem Evidently, this strategy can be applied recursively 77 a n

2 78 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM 4 Partitioning the tridiagonal matrix Partitioning the irreducible tridiagonal matrix is done in the following way We write (4) a b b a bm a b b m a m b m b m a m+ b m+ b m+ a m+ bn b a bm [ b n a n b m a m b m a m+ b m b m+ b m+ a m+ bn b n a n ] [ ] +ρuu ±em with u and ρ ±b m, e + ±b m b m b m ±b m where e m is a vector of length m n and e is a vector of length n m Notice that the most straightforward way to partition the problem without modifying the diagonal elements leads to a rank-two modification With the approach of (4) we have the original as a sum of two smaller tridiagonal systems plus a rank-one modification 43 Solving the small systems We solve the half-sized eigenvalue problems, (43) i Q i Λ i Q i, Q i Q i I, i, hese two spectral decompositions can be computed by any algorithm, in particular also by this divide and conquer algorithm by which the i would be further split It is clear that by this partitioning an large number of small problems can be generated that can be potentially solved in parallel For a parallel algorithm, however, the further phases of the algorithm must be parallelizable as well Plugging (43) into (4) gives [ ]([ )[ ] [ ] Q (44) ]+ρuu Q Λ +ρvv Q Q Λ with [ Q (45) v Q ] [ ±Q u e m Q e ] [ ±last row of Q first row of Q ]

3 44 DEFLAION 79 Now we have arrived at the eigenvalue problem (46) (D +ρvv )x λx, D Λ Λ diag(λ,,λ n ) hat is, we have to compute the spectral decomposition of a matrix that is a diagonal plus a rank-one update Let (47) D +ρvv QΛQ be this spectral decomposition hen, the spectral decomposition of the tridiagonal is [ ] [ ] Q (48) QΛQ Q Q Q Forming the product (Q Q )Q will turn out to be the most expensive step of the algorithm It costs n 3 +O(n ) floating point operations 44 Deflation here are certain solutions of (47) that can be given immediately, by just looking carefully at the equation If there are zero entries in v then we have ( (49) vi v e i ) (D +ρvv )e i d i e i hus, if an entry of v vanishes we can read the eigenvalue from the diagonal of D at once and the corresponding eigenvector is a coordinate vector If identical entries occur in the diagonal of D, say d i d j, with i < j, then we can find a plane rotation G(i,j,φ) (see (34)) such that it introduces a zero into the j-th position of v, Notice, that (for any ϕ), G v G(i,j,ϕ) v vi +v j i j G(i,j,ϕ) DG(i,j,ϕ) D, d i d j So, if there are multiple eigenvalues in D we can reduce all but one of them by introducing zeros in v and then proceed as previously in (49) When working with floating point numbers we deflate if (4) v i < Cε or d i d j < Cε, ( D +ρvv ) where C is a small constant Deflation changes the eigenvalue problem for D+ρvv into the eigenvalue problem for [ D +ρv (4) v ] O G (D+ρvv )G+E, E < Cε D O D + ρ v 4, where D has no multiple diagonal entries and v has no zero entries So, we have to compute the spectral decomposition of the matrix in (4) which is similar to a slight perturbation of the original matrix G is the product of Givens rotations

4 8 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM 44 Numerical examples Let us first consider uu hen a little Matlab experiment shows that Q Q with Q Here it is not possible to deflate Let us now look at an example with more symmetry, + + +uu

5 45 HE EIGENVALUE PROBLEM FOR D +ρvv 8 Now, Matlab gives Q Q with Q In this example we have three double eigenvalues Because the corresponding components of v (v i and v i+ ) are equal we define hen, G G(,4,π/4)G(,5,π/4)G(3,6,π/4) G Q Q G G Q Q G+G v(g v) D +G v(g v) herefore, (in this example) e 4, e 5, and e 6 are eigenvectors of D +G v(g v) D +G vv G corresponding to the eigenvalues d 4, d 5, and d 6, respectively he eigenvectors of corresponding to these three eigenvalues are the last three columns of Q G he eigenvalue problem for D+ρvv We know that ρ Otherwise there is nothing to be done Furthermore, after deflation, we know that all elements of v are nonzero and that the diagonal elements of D are all

6 8 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM distinct, in fact, d i d j > Cε We order the diagonal elements of D such that d < d < < d n Notice that this procedure permutes the elements of v as well Let (λ,x) be an eigenpair of (4) (D +ρvv )x λx hen, (43) (D λi)x ρvv x λ cannot be equal to one of the d i If λ d k then the k-th element on the left of (43) vanishes But then either v k or v x he first cannot be true for our assumption about v If on the other hand v x then (D d k I)x hus x e k and v e k v k, which cannot be true herefore D λi is nonsingular and (44) x ρ(λi D) v(v x) his equation shows that x is proportional to (λi D) v If we require x then (45) x (λi D) v (λi D) v Multiplying (44) by v from the left we get (46) v x ρv (λi D) v(v x) Since v x, λ is an eigenvalue of (4) if and only if Figure 4: Graph of + λ + λ λ λ λ λ (47) f(λ) : ρv (λi D) v ρ n k v k λ d k

7 45 HE EIGENVALUE PROBLEM FOR D +ρvv 83 his equation is called secular equation he secular equation has poles at the eigenvalues of D and zeros at the eigenvalues of D +ρvv Notice that f (λ) ρ n k v k (λ d k ) hus, the derivative of f is positive if ρ > wherever it has a finite value If ρ < the derivative of f is negative (almost) everywhere A typical graph of f with ρ > is depicted in Fig 4 (If ρ is negative the image can be flipped left to right) he secular equation implies the interlacing property of the eigenvalues of D and of D +ρvv, (48) d < λ < d < λ < < d n < λ n, ρ > or (49) λ < d < λ < d < < λ n < d n, ρ < So, we have to compute one eigenvalue in each of the intervals (d i,d i+ ), i < n, and a further eigenvalue in (d n, ) or (,d ) he corresponding eigenvector is then given by (45) Evidently, these tasks are easy to parallelize Equations (47) and (45) can also been obtained from the relations [ ] ρ v [ ] [ ] [ ρ ] ρv v λi D ρv I λi D ρvv I [ v (λi D) ] [ ] [ ρ v (λi D) v I λi D (λi D) v I hese are simply block LDL factorizations of the first matrix he first is the well-known one where the factorization is started with the (,) block he second is a backward factorization that is started with the (, ) block Because the determinants of the tridiagonal matrices are all unity, we have ] (4) ρ det(λi D ρvv ) ρ ( ρv (λi D) v)det(λi D) Denoting the eigenvalues of D +ρvv again by λ < λ < < λ n this implies (4) (λ λ j ) ( ρv (λi D) v) (λ d j ) j j ( ) n v n k ρ (λ d j ) λ d k k (λ d j ) ρ n j vk j k j k (λ d j ) Setting λ d k gives (4) (d k λ j ) ρvk j (d k d j ) j j i

8 84 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM or (43) v k ρ ρ (d k λ j ) j (d k d j ) j j i (d k λ j ) k j (d k d j ) k j ρ (d k λ j ) k j (d k d j ) k j (λ j d k ) jk jk+ > (d j d k ) (λ j d k )( ) n k+ jk jk+ herefore, the quantity on the right side is positive, so k (d k λ j ) n (λ j d k ) (44) v k j jk (d k d j ) (d j d k ) ρ k j jk+ (d j d k )( ) n k (Similar arguments hold if ρ < ) hus, we have the solution of the following inverse eigenvalue problem: Given D diag(d,,d n ) and values λ,,λ n that satisfy (48) Find a vector v [v,,v n ] with positive components v k such that the matrix D+vv has the prescribed eigenvalues λ,,λ n he solution is given by (44) he positivity of the v k makes the solution unique 46 Solving the secular equation In this section we follow closely the exposition of Demmel[3] We consider the computation of the zero of f(λ) in the interval (d i,d i+ ) We assume that ρ We may simply apply Newton s iteration to solve f(λ) However, if we look carefully at Fig 4 then we notice that the tangent at certain points in (d i,d i+ ) crosses the real axis outside this interval his happens in particular if the weights v i or v i+ are small herefore that zero finder has to be adapted in such a way that it captures the poles at the interval points It is relatively straightforward to try the ansatz (45) h(λ) c d i λ + c d i+ λ +c 3 Notice that, given the coefficients c, c, and c 3, the equation h(λ) can easily be solved by means of the equivalent quadratic equation (46) c (d i+ λ)+c (d i λ)+c 3 (d i λ)(d i+ λ) his equation has two zeros Precisly one of them is inside (d i,d i+ ) he coefficients c, c, and c 3 are computed in the following way Let us assume that we have available an approximation λ j to the zero in (d i,d i+ ) We request that h(λ j ) f(λ j ) and h (λ j ) f (λ j ) he exact procedure is as follows We write i v n k vk (47) f(λ) + + +ψ (λ)+ψ (λ) d k λ d k λ k ki+ }{{}}{{} ψ (λ) ψ (λ)

9 47 A FIRS ALGORIHM 85 ψ (λ) is a sum of positive terms and ψ (λ) is a sum of negative terms Both ψ (λ) and ψ (λ) can be computed accurately, whereas adding them would likely provoke cancellation and loss of relative accuracy We now choose c and ĉ such that (48) h (λ) : ĉ + c d i λ satisfies h (λ j ) ψ (λ j ) and h (λ j ) ψ (λ j ) his means that the graphs of h and of ψ are tangent at λ λ j his is similar to Newton s method However in Newton s method a straight line is fitted to the given function he coefficients in (48) are given by c ψ (λ j )(d i λ j ) >, ĉ ψ (λ j ) ψ (λ j )(d i λ j ) i k v k d k d i (d k λ j ) Similarly, the two constants c and ĉ are determined such that (49) h (λ) : ĉ + with the coefficients Finally, we set c d i+ λ satisfies h (λ j ) ψ (λ j ) and h (λ j ) ψ (λ j ) c ψ (λ j )(d i+ λ j ) >, n ĉ ψ (λ j ) ψ (λ j )(d i+ λ j ) vk d k d i+ (d k λ) ki+ (43) h(λ) +h (λ)+h (λ) (+ĉ +ĉ ) + }{{} d i λ + c d i+ λ c 3 his zerofinder is converging quadratically to the desired zero [7] Usually to 3 steps are sufficient to get the zero to machine precision herefore finding a zero only requires O(n) flops hus, finding all zeros costs O(n ) floating point operations c 47 A first algorithm We are now ready to give the divide and conquer algorithm, see Algorithm 4 All steps except step require O(n ) operations to complete he step costs n 3 flops hus, the full divide and conquer algorithm, requires (43) ( n ) 3 (n) n 3 + (n/) n (n/4) n 3 + n3 ( n ) (n/8) 4 3 n3 his serial complexity of the algorithm very often overestimates the computational costs of the algorithm due to significant deflation that is observed surprisingly often

10 86 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM Algorithm 4 he tridiagonal divide and conquer algorithm : Let C n n be a real symmetric tridiagonal matrix his algorithm computes the spectral decomposition of QΛQ, where the diagonal Λ is the matrix of eigenvalues and Q is orthogonal : if is then 3: return (Λ ;Q ) 4: else [ ] O 5: Partition +ρuu O according to (4) 6: Call this algorithm with as input and Q, Λ as output 7: Call this algorithm with as input and Q, Λ as output 8: Form D +ρvv from Λ,Λ,Q,Q according to (44) (46) 9: Find the eigenvalues [ Λ] and the eigenvectors Q of D +ρvv Q O : Form Q Q which are the eigenvectors of : return (Λ; Q) : if O Q 47 A numerical example Let A be a 4 4 matrix (43) A D +vv β +β 5 + β β [ β β ] In this example (that is similar to one in [8]) we want to point at a problem that the divide and conquer algorithm possesses as it is given in Algorithm 4, namely the loss of orthogonality among eigenvectors Before we do some Matlab tests let us look more closely at D and v in (43) his example becomes difficult to solve if β gets very small In Figures 4 to 45 we see graphs of the function f β (λ) that appears in the secular equation for β, β, and β he critical zeros move towards from both sides he weights v v 3 β are however not so small that they should be deflated he following Matlab code shows the problem We execute the commands for β k for k,,,4,8 v [ beta beta ] ; % rank- modification d [, -beta, +beta, 5] ; % diagonal matrix L eig(diag(d) + v*v ) % eigenvalues of the modified matrix e ones(4,); q (d*e -e*l )\(v*e ); % unnormalized eigenvectors cf (55) Q sqrt(diag(q *q)); q q/(e*q ); % normalized eigenvectors norm(q *q-eye(4)) % check for orthogonality We do not bother how we compute the eigenvalues We simply use Matlab s built-in function eig We get the results of able 4

11 47 A FIRS ALGORIHM Figure 4: Secular equation corresponding to (43) for β Figure 43: Secular equation corresponding to (43) for β We observe loss of orthogonality among the eigenvectors as the eigenvalues get closer and closer his may not be surprising as we compute the eigenvectors by formula (45) x (λi D) v (λi D) v If λ λ and λ λ 3 which are almost equal, λ λ 3 then intuitively one expects almost the same eigenvectors We have in fact Q Q I Orthogonality is lost only with respect to the vectors corresponding to the eigenvalues close to

12 88 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM Figure 44: Secular equation corresponding to (43) for β for λ Figure 45: Secular equation corresponding to (43) for β for 9 λ Already Dongarra and Sorensen [4] analyzed this problem In their formulation they normalize the vector v of D +ρvv to have norm unity, v hey stated Lemma 4 Let ( )[ ] (433) q λ v d λ, v d λ,, v n ρ / d n λ f (λ) hen for any λ,µ {d,,d n } we have (434) q λ q µ f(λ) f(µ) λ µ [f (λ)f (µ)] / Proof Observe that hen the proof is straightforward λ µ (d j λ)(d j µ) d j λ d j µ

13 48 HE ALGORIHM OF GU AND EISENSA 89 β λ λ λ 3 λ 4 Q Q I able 4: Loss of orthogonality among the eigenvectors computed by (45) Formula (434) indicates how problems may arise In exact arithmetic, if λ and µ are eigenvalues then f(λ) f(µ) However, in floating point arithmetic this values may be small but nonzero, eg, O(ε) If λ µ is very small as well then we may have trouble! So, a remedy for the problem was for a long time to compute the eigenvalues in doubled precision, so that f(λ) O(ε ) his would counteract a potential O(ε) of λ µ his solution was quite unsatisfactory because doubled precision is in general very slow since it is implemented in software It took a decade until a proper solution was found 48 he algorithm of Gu and Eisenstat Computing eigenvector according to the formula v λ d (435) x α(λi D) v α v n λ d n, α (λi D) v, is bound to fail if λ is very close to a pole d k and the difference λ d k has an error of size O(ε d k ) instead of only O(ε d k λ ) o resolve this problem Gu and Eisenstat [5] found a trick that is at the same time ingenious and simple hey observed that the v k in (44) are very accurately determined by the data d i and λ i herefore, once the eigenvalues are computed accurately a vector ˆv could be computed such that the λ i are accurate eigenvalues of D+ˆvˆv If ˆv approximates well the original v then the new eigenvectors will be the exact eigenvectors of a slightly modified eigenvalue problem, which is all we can hope for he zeros of the secular equation can be computed accurately by the method presented in section 46 However, a shift of variables is necessary In the interval (d i,d i+ ) the origin of the real axis is moved to d i if λ i is closer to d i than to d i+, ie, if f((d i +d i+ )/) > Otherwise, the origin is shifted to d i+ his shift of the origin avoids the computation of the smallest difference d i λ (or d i+ λ) in (435), thus avoiding cancellation in this most sensitive quantity Equation (46) can be rewritten as (436) (c i+ +c i +c 3 i i+ ) (c }{{} +c +c 3 ( i + i+ )) η + c }{{} 3 η }{{}, b a c where i d i λ j, i+ d i+ λ j, and λ j+ λ j +η is the next approximate zero With equations (48) (43) the coefficients in (436) get (437) a c +c +c 3 ( i + i+ ) (+Ψ +Ψ )( i + i+ ) (Ψ +Ψ ) i i+, b c i+ +c i +c 3 i i+ i i+ (+Ψ +Ψ ), c c 3 +Ψ +Ψ i Ψ i+ Ψ

14 9 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM If we are looking for a zero that is closer to d i than to d i+ then we move the origin to λ j, ie, we have eg i λ j he solution of (436) that lies inside the interval is [7] (438) η a a 4bc c, if a, b a+, a 4bc if a > he following algorithm shows how step 9 of the tridiagonal divide and conquer algorithm 4 must be implemented Algorithm 4 A stable eigensolver for D +vv : hisalgorithmstablycomputesthespectraldecompositionofd+vv QΛQ where D diag(d,d n ), v [v,,v n ] R n, Λ diag(λ,λ n ), and Q [q,,q n ] : d i+ d n + v 3: In each interval (d i,d i+ ) compute the zero λ i of the secular equation f(λ) 4: Use the formula (44) to compute the vector ˆv such that the λ i are the exact eigenvalues of D +ˆvˆv 5: In each interval (d i,d i+ ) compute the eigenvectors of D + ˆvˆv according to (45), 6: return (Λ;Q) q i (λ ii D) ˆv (λ i I D) ˆv 48 A numerical example [continued] We continue the discussion of the example on page 86 where the eigenvalue problem of (439) A D +vv β +β + β ] β [ β β 5 he Matlab code that we showed did not give orthogonal eigenvectors We show in the following script that the formulae (44) really solve the problem dlam zeros(n,); for k:n, [dlam(k), dvec(:,k)] zerodandc(d,v,k); V ones(n,); for k:n, V(k) prod(abs(dvec(k,:)))/prod(d(k) - d(:k-))/prod(d(k+:n) - d(k)); V(k) sqrt(v(k)); Q (dvec)\(v*e ); diagq sqrt(diag(q *Q)); Q Q/(e*diagq );

15 48 HE ALGORIHM OF GU AND EISENSA 9 for k:n, if dlam(k)>, dlam(k) dlam(k) + d(k); else dlam(k) d(k+) + dlam(k); norm(q *Q-eye(n)) norm((diag(d) + v*v )*Q - Q*diag(dlam )) A zero finder returns for each interval the quantity λ i d i and the vector [d λ i,,d n λ i ] to high precision hese vector elements have been computed as (d k d i ) (λ i d i ) he zerofinder of Li [7] has been employed here At the of this section we list the zerofinder written in Matlab that was used here he formulae (437) and (438) have been used to solve the quadratic equation (436) Notice that only one of the while loops is traversed, deping on if the zero is closer to the pole on the left or to the right of the interval he v k of formula (44) are computed next Q contains the eigenvectors β Algorithm Q Q I AQ QΛ I II I II I II I II able 4: Loss of orthogonality among the eigenvectors computed by the straightforward algorithm (I) and the Gu-Eisenstat approach (II) Again we ran the code for β k for k,,,4,8 he numbers in able 4 confirm that the new formulae are much more accurate than the straight forward ones he norms of the errors obtained for the Gu-Eisenstat algorithm always are in the order of machine precision, ie, 6 function [lambda,dl] zerodandc(d,v,i) % ZERODANDC - Computes eigenvalue lambda in the i-th interval % (d(i), d(i+)) with Li s middle way zero finder % dl is the n-vector [d(n) - lambda] n length(d); di d(i); v v^; if i < n, di d(i+); lambda (di + di)/; else di d(n) + norm(v)^; lambda di; eta ; psi sum((v(:i)^)/(d(:i) - lambda));

16 9 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM psi sum((v(i+:n)^)/(d(i+:n) - lambda)); if + psi + psi >, % zero is on the left half of the interval d d - di; lambda lambda - di; di di - di; di ; while abs(eta) > *eps psi sum(v(:i)/(d(:i) - lambda)); psis sum(v(:i)/((d(:i) - lambda))^); psi sum((v(i+:n))/(d(i+:n) - lambda)); psis sum(v(i+:n)/((d(i+:n) - lambda))^); % Solve for zero Di -lambda; Di di - lambda; a (Di + Di)*( + psi + psi) - Di*Di*(psis + psis); b Di*Di*( + psi + psi); c ( + psi + psi) - Di*psis - Di*psis; if a >, eta (*b)/(a + sqrt(a^ - 4*b*c)); else eta (a - sqrt(a^ - 4*b*c))/(*c); lambda lambda + eta; else % zero is on the right half of the interval d d - di; lambda lambda - di; di di - di; di ; while abs(eta) > *eps psi sum(v(:i)/(d(:i) - lambda)); psis sum(v(:i)/((d(:i) - lambda))^); psi sum((v(i+:n))/(d(i+:n) - lambda)); psis sum(v(i+:n)/((d(i+:n) - lambda))^); % Solve for zero Di di - lambda; Di - lambda; a (Di + Di)*( + psi + psi) - Di*Di*(psis + psis); b Di*Di*( + psi + psi); c ( + psi + psi) - Di*psis - Di*psis; if a >, eta (*b)/(a + sqrt(a^ - 4*b*c)); else eta (a - sqrt(a^ - 4*b*c))/(*c); lambda lambda + eta; dl d - lambda; return

17 BIBLIOGRAPHY 93 Bibliography [] E Anderson, Z Bai, C Bischof, J Demmel, J Dongarra, J D Croz, A Greenbaum, S Hammarling, A McKenney, S Ostrouchov, and D Sorensen, LAPACK Users Guide - Release, SIAM, Philadelphia, PA, 994 (Software and guide are available from Netlib at URL [] J J M Cuppen, A divide and conquer method for the symmetric tridiagonal eigenproblem, Numer Math, 36 (98), pp [3] J W Demmel, Applied Numerical Linear Algebra, SIAM, Philadelphia, PA, 997 [4] J J Dongarra and D C Sorensen, A fully parallel algorithm for the symmetric eigenvalue problem, SIAM J Sci Stat Comput, 8 (987), pp s39 s54 [5] M Gu and S C Eisenstat, A stable and efficient algorithm for the rank-one modification of the symmetric eigenproblem, SIAM J Matrix Anal Appl, 5 (994), pp [6], A divide-and-conquer algorithm for the symmetric tridiagonal eigenproblem, SIAM J Matrix Anal Appl, 6 (995), pp 7 9 [7] R-C Li, Solving secular equations stably and efficiently, echnical Report U-CS- 94-6, University of ennessee, Knoxville, N, Nov 994 LAPACK Working Note No 89 [8] D C Sorensen and P P ang, On the orthogonality of eigenvectors computed by divide-and-conquer techniques, SIAM J Numer Anal, 8 (99), pp

18 94 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM

Cuppen s Divide and Conquer Algorithm

Cuppen s Divide and Conquer Algorithm Chapter 5 Cuppen s Divide and Conquer Algorithm In this chapter we deal with an algorithm that is designed for the efficient solution of the symmetric tridiagonal eigenvalue problem a 1 b 1. (5.1) Tx =

More information

Direct methods for symmetric eigenvalue problems

Direct methods for symmetric eigenvalue problems Direct methods for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 4, 2008 1 Theoretical background Posing the question Perturbation theory

More information

Exponentials of Symmetric Matrices through Tridiagonal Reductions

Exponentials of Symmetric Matrices through Tridiagonal Reductions Exponentials of Symmetric Matrices through Tridiagonal Reductions Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A simple and efficient numerical algorithm

More information

Generalized interval arithmetic on compact matrix Lie groups

Generalized interval arithmetic on compact matrix Lie groups myjournal manuscript No. (will be inserted by the editor) Generalized interval arithmetic on compact matrix Lie groups Hermann Schichl, Mihály Csaba Markót, Arnold Neumaier Faculty of Mathematics, University

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

The QR Algorithm. Chapter The basic QR algorithm

The QR Algorithm. Chapter The basic QR algorithm Chapter 3 The QR Algorithm The QR algorithm computes a Schur decomposition of a matrix. It is certainly one of the most important algorithm in eigenvalue computations. However, it is applied to dense (or:

More information

Chapter 12 Solving secular equations

Chapter 12 Solving secular equations Chapter 12 Solving secular equations Gérard MEURANT January-February, 2012 1 Examples of Secular Equations 2 Secular equation solvers 3 Numerical experiments Examples of secular equations What are secular

More information

Computing least squares condition numbers on hybrid multicore/gpu systems

Computing least squares condition numbers on hybrid multicore/gpu systems Computing least squares condition numbers on hybrid multicore/gpu systems M. Baboulin and J. Dongarra and R. Lacroix Abstract This paper presents an efficient computation for least squares conditioning

More information

The Algorithm of Multiple Relatively Robust Representations for Multi-Core Processors

The Algorithm of Multiple Relatively Robust Representations for Multi-Core Processors Aachen Institute for Advanced Study in Computational Engineering Science Preprint: AICES-2010/09-4 23/September/2010 The Algorithm of Multiple Relatively Robust Representations for Multi-Core Processors

More information

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van

More information

A Parallel Divide and Conquer Algorithm for the Symmetric Eigenvalue Problem on Distributed Memory Architectures. F Tisseur and J Dongarra

A Parallel Divide and Conquer Algorithm for the Symmetric Eigenvalue Problem on Distributed Memory Architectures. F Tisseur and J Dongarra A Parallel Divide and Conquer Algorithm for the Symmetric Eigenvalue Problem on Distributed Memory Architectures F Tisseur and J Dongarra 999 MIMS EPrint: 2007.225 Manchester Institute for Mathematical

More information

LAPACK and the BLAS. Chapter LAPACK

LAPACK and the BLAS. Chapter LAPACK Chapter 4 LAPACK and the BLAS 4.1 LAPACK (This section is essentially compiled from the LAPACK User s Guide [1] that is available online from http://www.netlib.org/lapack/lug/.) LAPACK [1] is a library

More information

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices.

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. A.M. Matsekh E.P. Shurina 1 Introduction We present a hybrid scheme for computing singular vectors

More information

A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Akiko Fukuda

A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Akiko Fukuda Journal of Math-for-Industry Vol 3 (20A-4) pp 47 52 A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Aio Fuuda Received on October 6 200 / Revised on February 7 20 Abstract

More information

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract Computing the Logarithm of a Symmetric Positive Denite Matrix Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A numerical method for computing the logarithm

More information

arxiv: v2 [math.na] 21 Sep 2015

arxiv: v2 [math.na] 21 Sep 2015 Forward stable eigenvalue decomposition of rank-one modifications of diagonal matrices arxiv:1405.7537v2 [math.na] 21 Sep 2015 N. Jakovčević Stor a,1,, I. Slapničar a,1, J. L. Barlow b,2 a Faculty of Electrical

More information

A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION

A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION SIAM J MATRIX ANAL APPL Vol 0, No 0, pp 000 000 c XXXX Society for Industrial and Applied Mathematics A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION WEI XU AND SANZHENG QIAO Abstract This paper

More information

Accurate eigenvalue decomposition of arrowhead matrices and applications

Accurate eigenvalue decomposition of arrowhead matrices and applications Accurate eigenvalue decomposition of arrowhead matrices and applications Nevena Jakovčević Stor FESB, University of Split joint work with Ivan Slapničar and Jesse Barlow IWASEP9 June 4th, 2012. 1/31 Introduction

More information

Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems

Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems LESLIE FOSTER and RAJESH KOMMU San Jose State University Existing routines, such as xgelsy or xgelsd in LAPACK, for

More information

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Suares Problem Hongguo Xu Dedicated to Professor Erxiong Jiang on the occasion of his 7th birthday. Abstract We present

More information

Matrices, Moments and Quadrature, cont d

Matrices, Moments and Quadrature, cont d Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that

More information

Efficient and Accurate Rectangular Window Subspace Tracking

Efficient and Accurate Rectangular Window Subspace Tracking Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu

More information

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006. LAPACK-Style Codes for Pivoted Cholesky and QR Updating Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig 2007 MIMS EPrint: 2006.385 Manchester Institute for Mathematical Sciences School of Mathematics

More information

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex

More information

Solving large scale eigenvalue problems

Solving large scale eigenvalue problems arge scale eigenvalue problems, Lecture 4, March 14, 2018 1/41 Lecture 4, March 14, 2018: The QR algorithm http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich E-mail:

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview

More information

Total least squares. Gérard MEURANT. October, 2008

Total least squares. Gérard MEURANT. October, 2008 Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares

More information

A Divide-and-Conquer Method for the Takagi Factorization

A Divide-and-Conquer Method for the Takagi Factorization A Divide-and-Conquer Method for the Takagi Factorization Wei Xu 1 and Sanzheng Qiao 1, Department of Computing and Software, McMaster University Hamilton, Ont, L8S 4K1, Canada. 1 xuw5@mcmaster.ca qiao@mcmaster.ca

More information

ANONSINGULAR tridiagonal linear system of the form

ANONSINGULAR tridiagonal linear system of the form Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal

More information

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A

More information

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 9, 2008 Today

More information

ETNA Kent State University

ETNA Kent State University C 8 Electronic Transactions on Numerical Analysis. Volume 17, pp. 76-2, 2004. Copyright 2004,. ISSN 1068-613. etnamcs.kent.edu STRONG RANK REVEALING CHOLESKY FACTORIZATION M. GU AND L. MIRANIAN Abstract.

More information

Iterative Algorithm for Computing the Eigenvalues

Iterative Algorithm for Computing the Eigenvalues Iterative Algorithm for Computing the Eigenvalues LILJANA FERBAR Faculty of Economics University of Ljubljana Kardeljeva pl. 17, 1000 Ljubljana SLOVENIA Abstract: - We consider the eigenvalue problem Hx

More information

LAPACK-Style Codes for Pivoted Cholesky and QR Updating

LAPACK-Style Codes for Pivoted Cholesky and QR Updating LAPACK-Style Codes for Pivoted Cholesky and QR Updating Sven Hammarling 1, Nicholas J. Higham 2, and Craig Lucas 3 1 NAG Ltd.,Wilkinson House, Jordan Hill Road, Oxford, OX2 8DR, England, sven@nag.co.uk,

More information

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems AMS 209, Fall 205 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems. Overview We are interested in solving a well-defined linear system given

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse

More information

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 8, 2009 Today

More information

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm References: Trefethen & Bau textbook Eigenvalue problem: given a matrix A, find

More information

6 EIGENVALUES AND EIGENVECTORS

6 EIGENVALUES AND EIGENVECTORS 6 EIGENVALUES AND EIGENVECTORS INTRODUCTION TO EIGENVALUES 61 Linear equations Ax = b come from steady state problems Eigenvalues have their greatest importance in dynamic problems The solution of du/dt

More information

Eigenvalue and Eigenvector Problems

Eigenvalue and Eigenvector Problems Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

KU Leuven Department of Computer Science

KU Leuven Department of Computer Science Backward error of polynomial eigenvalue problems solved by linearization of Lagrange interpolants Piers W. Lawrence Robert M. Corless Report TW 655, September 214 KU Leuven Department of Computer Science

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to

More information

NAG Library Routine Document F08JDF (DSTEVR)

NAG Library Routine Document F08JDF (DSTEVR) F08 Least-squares and Eigenvalue Problems (LAPACK) NAG Library Routine Document (DSTEVR) Note: before using this routine, please read the Users Note for your implementation to check the interpretation

More information

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,

More information

Eigenvalue Problems and Singular Value Decomposition

Eigenvalue Problems and Singular Value Decomposition Eigenvalue Problems and Singular Value Decomposition Sanzheng Qiao Department of Computing and Software McMaster University August, 2012 Outline 1 Eigenvalue Problems 2 Singular Value Decomposition 3 Software

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue

More information

Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems. Contents

Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems. Contents Eigenvalue and Least-squares Problems Module Contents Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems nag nsym gen eig provides procedures for solving nonsymmetric generalized

More information

Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor

Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor Masami Takata 1, Hiroyuki Ishigami 2, Kini Kimura 2, and Yoshimasa Nakamura 2 1 Academic Group of Information

More information

Normalized power iterations for the computation of SVD

Normalized power iterations for the computation of SVD Normalized power iterations for the computation of SVD Per-Gunnar Martinsson Department of Applied Mathematics University of Colorado Boulder, Co. Per-gunnar.Martinsson@Colorado.edu Arthur Szlam Courant

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Block Lanczos Tridiagonalization of Complex Symmetric Matrices

Block Lanczos Tridiagonalization of Complex Symmetric Matrices Block Lanczos Tridiagonalization of Complex Symmetric Matrices Sanzheng Qiao, Guohong Liu, Wei Xu Department of Computing and Software, McMaster University, Hamilton, Ontario L8S 4L7 ABSTRACT The classic

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

Numerical Solution of Linear Eigenvalue Problems

Numerical Solution of Linear Eigenvalue Problems Numerical Solution of Linear Eigenvalue Problems Jessica Bosch and Chen Greif Abstract We review numerical methods for computing eigenvalues of matrices We start by considering the computation of the dominant

More information

Orthogonal iteration to QR

Orthogonal iteration to QR Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation

More information

D. Gimenez, M. T. Camara, P. Montilla. Aptdo Murcia. Spain. ABSTRACT

D. Gimenez, M. T. Camara, P. Montilla. Aptdo Murcia. Spain.   ABSTRACT Accelerating the Convergence of Blocked Jacobi Methods 1 D. Gimenez, M. T. Camara, P. Montilla Departamento de Informatica y Sistemas. Univ de Murcia. Aptdo 401. 0001 Murcia. Spain. e-mail: fdomingo,cpmcm,cppmmg@dif.um.es

More information

Solving linear equations with Gaussian Elimination (I)

Solving linear equations with Gaussian Elimination (I) Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian

More information

Algorithms for Solving the Polynomial Eigenvalue Problem

Algorithms for Solving the Polynomial Eigenvalue Problem Algorithms for Solving the Polynomial Eigenvalue Problem Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/ Joint work with D. Steven Mackey

More information

arxiv: v1 [math.na] 5 May 2011

arxiv: v1 [math.na] 5 May 2011 ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and

More information

COMPUTATION OF GAUSS-KRONROD QUADRATURE RULES

COMPUTATION OF GAUSS-KRONROD QUADRATURE RULES MATHEMATICS OF COMPUTATION Volume 69, Number 231, Pages 1035 1052 S 0025-5718(00)01174-1 Article electronically published on February 17, 2000 COMPUTATION OF GAUSS-KRONROD QUADRATURE RULES D.CALVETTI,G.H.GOLUB,W.B.GRAGG,ANDL.REICHEL

More information

Matrix Shapes Invariant under the Symmetric QR Algorithm

Matrix Shapes Invariant under the Symmetric QR Algorithm NUMERICAL ANALYSIS PROJECT MANUSCRIPT NA-92-12 SEPTEMBER 1992 Matrix Shapes Invariant under the Symmetric QR Algorithm Peter Arbenz and Gene H. Golub NUMERICAL ANALYSIS PROJECT COMPUTER SCIENCE DEPARTMENT

More information

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n. Lecture # 11 The Power Method for Eigenvalues Part II The power method find the largest (in magnitude) eigenvalue of It makes two assumptions. 1. A is diagonalizable. That is, A R n n. A = XΛX 1 for some

More information

Preconditioned Parallel Block Jacobi SVD Algorithm

Preconditioned Parallel Block Jacobi SVD Algorithm Parallel Numerics 5, 15-24 M. Vajteršic, R. Trobec, P. Zinterhof, A. Uhl (Eds.) Chapter 2: Matrix Algebra ISBN 961-633-67-8 Preconditioned Parallel Block Jacobi SVD Algorithm Gabriel Okša 1, Marián Vajteršic

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Recent Advances in the Numerical Solution of Quadratic Eigenvalue Problems

Recent Advances in the Numerical Solution of Quadratic Eigenvalue Problems Recent Advances in the Numerical Solution of Quadratic Eigenvalue Problems Françoise Tisseur School of Mathematics The University of Manchester ftisseur@ma.man.ac.uk http://www.ma.man.ac.uk/~ftisseur/

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis Volume 8, pp 127-137, 1999 Copyright 1999, ISSN 1068-9613 ETNA etna@mcskentedu BOUNDS FOR THE MINIMUM EIGENVALUE OF A SYMMETRIC TOEPLITZ MATRIX HEINRICH VOSS

More information

Solving large scale eigenvalue problems

Solving large scale eigenvalue problems arge scale eigenvalue problems, Lecture 5, March 23, 2016 1/30 Lecture 5, March 23, 2016: The QR algorithm II http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9 1. qr and complete orthogonal factorization poor man s svd can solve many problems on the svd list using either of these factorizations but they

More information

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as

More information

Key words. modified Cholesky factorization, optimization, Newton s method, symmetric indefinite

Key words. modified Cholesky factorization, optimization, Newton s method, symmetric indefinite SIAM J. MATRIX ANAL. APPL. c 1998 Society for Industrial and Applied Mathematics Vol. 19, No. 4, pp. 197 111, October 1998 16 A MODIFIED CHOLESKY ALGORITHM BASED ON A SYMMETRIC INDEFINITE FACTORIZATION

More information

DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM. Geunseop Lee

DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM. Geunseop Lee J. Korean Math. Soc. 0 (0), No. 0, pp. 1 0 https://doi.org/10.4134/jkms.j160152 pissn: 0304-9914 / eissn: 2234-3008 DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM

More information

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12 24 8 Matrices Determinant of 2 2 matrix Given a 2 2 matrix [ ] a a A = 2 a 2 a 22 the real number a a 22 a 2 a 2 is determinant and denoted by det(a) = a a 2 a 2 a 22 Example 8 Find determinant of 2 2

More information

Dense LU factorization and its error analysis

Dense LU factorization and its error analysis Dense LU factorization and its error analysis Laura Grigori INRIA and LJLL, UPMC February 2016 Plan Basis of floating point arithmetic and stability analysis Notation, results, proofs taken from [N.J.Higham,

More information

Eigenvalues, Eigenvectors, and Diagonalization

Eigenvalues, Eigenvectors, and Diagonalization Week12 Eigenvalues, Eigenvectors, and Diagonalization 12.1 Opening Remarks 12.1.1 Predicting the Weather, Again Let us revisit the example from Week 4, in which we had a simple model for predicting the

More information

Accuracy and Stability in Numerical Linear Algebra

Accuracy and Stability in Numerical Linear Algebra Proceedings of the 1. Conference on Applied Mathematics and Computation Dubrovnik, Croatia, September 13 18, 1999 pp. 105 112 Accuracy and Stability in Numerical Linear Algebra Zlatko Drmač Abstract. We

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 10-12 Large-Scale Eigenvalue Problems in Trust-Region Calculations Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug ISSN 1389-6520 Reports of the Department of

More information

Diagonal Representation of Certain Matrices

Diagonal Representation of Certain Matrices Diagonal Representation of Certain Matrices Mark Tygert Research Report YALEU/DCS/RR-33 December 2, 2004 Abstract An explicit expression is provided for the characteristic polynomial of a matrix M of the

More information

Math 471 (Numerical methods) Chapter 4. Eigenvalues and Eigenvectors

Math 471 (Numerical methods) Chapter 4. Eigenvalues and Eigenvectors Math 471 (Numerical methods) Chapter 4. Eigenvalues and Eigenvectors Overlap 4.1 4.3 of Bradie 4.0 Review on eigenvalues and eigenvectors Definition. A nonzero vector v is called an eigenvector of A if

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

NAG Library Routine Document F08FAF (DSYEV)

NAG Library Routine Document F08FAF (DSYEV) NAG Library Routine Document (DSYEV) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT -09 Computational and Sensitivity Aspects of Eigenvalue-Based Methods for the Large-Scale Trust-Region Subproblem Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug

More information

Eigenproblems II: Computation

Eigenproblems II: Computation Eigenproblems II: Computation CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Eigenproblems II: Computation 1 / 31 Setup A

More information

Notes on Householder QR Factorization

Notes on Householder QR Factorization Notes on Householder QR Factorization Robert A van de Geijn Department of Computer Science he University of exas at Austin Austin, X 7872 rvdg@csutexasedu September 2, 24 Motivation A fundamental problem

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

NAG Library Routine Document F08VAF (DGGSVD)

NAG Library Routine Document F08VAF (DGGSVD) NAG Library Routine Document (DGGSVD) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

On an Inverse Problem for a Quadratic Eigenvalue Problem

On an Inverse Problem for a Quadratic Eigenvalue Problem International Journal of Difference Equations ISSN 0973-6069, Volume 12, Number 1, pp. 13 26 (2017) http://campus.mst.edu/ijde On an Inverse Problem for a Quadratic Eigenvalue Problem Ebru Ergun and Adil

More information

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 19, 2010 Today

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

Outline Background Schur-Horn Theorem Mirsky Theorem Sing-Thompson Theorem Weyl-Horn Theorem A Recursive Algorithm The Building Block Case The Origina

Outline Background Schur-Horn Theorem Mirsky Theorem Sing-Thompson Theorem Weyl-Horn Theorem A Recursive Algorithm The Building Block Case The Origina A Fast Recursive Algorithm for Constructing Matrices with Prescribed Eigenvalues and Singular Values by Moody T. Chu North Carolina State University Outline Background Schur-Horn Theorem Mirsky Theorem

More information

arxiv: v1 [cs.lg] 26 Jul 2017

arxiv: v1 [cs.lg] 26 Jul 2017 Updating Singular Value Decomposition for Rank One Matrix Perturbation Ratnik Gandhi, Amoli Rajgor School of Engineering & Applied Science, Ahmedabad University, Ahmedabad-380009, India arxiv:70708369v

More information

NAG Library Routine Document F08UBF (DSBGVX)

NAG Library Routine Document F08UBF (DSBGVX) NAG Library Routine Document (DSBGVX) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information