Cuppen s Divide and Conquer Algorithm
|
|
- Kelly Harmon
- 5 years ago
- Views:
Transcription
1 Chapter 4 Cuppen s Divide and Conquer Algorithm In this chapter we deal with an algorithm that is designed for the efficient solution of the symmetric tridiagonal eigenvalue problem a b (4) x λx, b a bn We noticed from able 3 that the reduction of a full symmetric matrix to a similar tridiagonal matrix requires about 8 3 n3 while the tridiagonal QR algorithm needs an estimated 6n 3 floating operations (flops) to converge Because of the importance of this subproblem a considerable effort has been put into finding faster algorithms than the QR algorithms to solve the tridiagonal eigenvalue problem In the mid-98 s Dongarra and Sorensen [4] promoted an algorithm originally proposed by Cuppen [] his algorithm was based on a divide and conquer strategy However, it took ten more years until a stable variant was found by Gu and Eisenstat [5, 6] oday, a stable implementation of this latter algorithm is available in LAPACK [] 4 he divide and conquer idea Divide and conquer is an old strategy in military to defeat an enemy going back at least to Caesar In computer science, divide and conquer (D&C) is an important algorithm design paradigm It works by recursively breaking down a problem into two or more subproblems of the same (or related) type, until these become simple enough to be solved directly he solutions to the subproblems are then combined to give a solution to the original problem ranslated to our problem the strategy becomes b n Partition the tridiagonal eigenvalue problem into two (or more) smaller tridiagonal eigenvalue problems Solve the two smaller problems 3 Combine the solutions of the smaller problems to get the desired solution of the overall problem Evidently, this strategy can be applied recursively 77 a n
2 78 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM 4 Partitioning the tridiagonal matrix Partitioning the irreducible tridiagonal matrix is done in the following way We write (4) a b b a bm a b b m a m b m b m a m+ b m+ b m+ a m+ bn b a bm [ b n a n b m a m b m a m+ b m b m+ b m+ a m+ bn b n a n ] [ ] +ρuu ±em with u and ρ ±b m, e + ±b m b m b m ±b m where e m is a vector of length m n and e is a vector of length n m Notice that the most straightforward way to partition the problem without modifying the diagonal elements leads to a rank-two modification With the approach of (4) we have the original as a sum of two smaller tridiagonal systems plus a rank-one modification 43 Solving the small systems We solve the half-sized eigenvalue problems, (43) i Q i Λ i Q i, Q i Q i I, i, hese two spectral decompositions can be computed by any algorithm, in particular also by this divide and conquer algorithm by which the i would be further split It is clear that by this partitioning an large number of small problems can be generated that can be potentially solved in parallel For a parallel algorithm, however, the further phases of the algorithm must be parallelizable as well Plugging (43) into (4) gives [ ]([ )[ ] [ ] Q (44) ]+ρuu Q Λ +ρvv Q Q Λ with [ Q (45) v Q ] [ ±Q u e m Q e ] [ ±last row of Q first row of Q ]
3 44 DEFLAION 79 Now we have arrived at the eigenvalue problem (46) (D +ρvv )x λx, D Λ Λ diag(λ,,λ n ) hat is, we have to compute the spectral decomposition of a matrix that is a diagonal plus a rank-one update Let (47) D +ρvv QΛQ be this spectral decomposition hen, the spectral decomposition of the tridiagonal is [ ] [ ] Q (48) QΛQ Q Q Q Forming the product (Q Q )Q will turn out to be the most expensive step of the algorithm It costs n 3 +O(n ) floating point operations 44 Deflation here are certain solutions of (47) that can be given immediately, by just looking carefully at the equation If there are zero entries in v then we have ( (49) vi v e i ) (D +ρvv )e i d i e i hus, if an entry of v vanishes we can read the eigenvalue from the diagonal of D at once and the corresponding eigenvector is a coordinate vector If identical entries occur in the diagonal of D, say d i d j, with i < j, then we can find a plane rotation G(i,j,φ) (see (34)) such that it introduces a zero into the j-th position of v, Notice, that (for any ϕ), G v G(i,j,ϕ) v vi +v j i j G(i,j,ϕ) DG(i,j,ϕ) D, d i d j So, if there are multiple eigenvalues in D we can reduce all but one of them by introducing zeros in v and then proceed as previously in (49) When working with floating point numbers we deflate if (4) v i < Cε or d i d j < Cε, ( D +ρvv ) where C is a small constant Deflation changes the eigenvalue problem for D+ρvv into the eigenvalue problem for [ D +ρv (4) v ] O G (D+ρvv )G+E, E < Cε D O D + ρ v 4, where D has no multiple diagonal entries and v has no zero entries So, we have to compute the spectral decomposition of the matrix in (4) which is similar to a slight perturbation of the original matrix G is the product of Givens rotations
4 8 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM 44 Numerical examples Let us first consider uu hen a little Matlab experiment shows that Q Q with Q Here it is not possible to deflate Let us now look at an example with more symmetry, + + +uu
5 45 HE EIGENVALUE PROBLEM FOR D +ρvv 8 Now, Matlab gives Q Q with Q In this example we have three double eigenvalues Because the corresponding components of v (v i and v i+ ) are equal we define hen, G G(,4,π/4)G(,5,π/4)G(3,6,π/4) G Q Q G G Q Q G+G v(g v) D +G v(g v) herefore, (in this example) e 4, e 5, and e 6 are eigenvectors of D +G v(g v) D +G vv G corresponding to the eigenvalues d 4, d 5, and d 6, respectively he eigenvectors of corresponding to these three eigenvalues are the last three columns of Q G he eigenvalue problem for D+ρvv We know that ρ Otherwise there is nothing to be done Furthermore, after deflation, we know that all elements of v are nonzero and that the diagonal elements of D are all
6 8 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM distinct, in fact, d i d j > Cε We order the diagonal elements of D such that d < d < < d n Notice that this procedure permutes the elements of v as well Let (λ,x) be an eigenpair of (4) (D +ρvv )x λx hen, (43) (D λi)x ρvv x λ cannot be equal to one of the d i If λ d k then the k-th element on the left of (43) vanishes But then either v k or v x he first cannot be true for our assumption about v If on the other hand v x then (D d k I)x hus x e k and v e k v k, which cannot be true herefore D λi is nonsingular and (44) x ρ(λi D) v(v x) his equation shows that x is proportional to (λi D) v If we require x then (45) x (λi D) v (λi D) v Multiplying (44) by v from the left we get (46) v x ρv (λi D) v(v x) Since v x, λ is an eigenvalue of (4) if and only if Figure 4: Graph of + λ + λ λ λ λ λ (47) f(λ) : ρv (λi D) v ρ n k v k λ d k
7 45 HE EIGENVALUE PROBLEM FOR D +ρvv 83 his equation is called secular equation he secular equation has poles at the eigenvalues of D and zeros at the eigenvalues of D +ρvv Notice that f (λ) ρ n k v k (λ d k ) hus, the derivative of f is positive if ρ > wherever it has a finite value If ρ < the derivative of f is negative (almost) everywhere A typical graph of f with ρ > is depicted in Fig 4 (If ρ is negative the image can be flipped left to right) he secular equation implies the interlacing property of the eigenvalues of D and of D +ρvv, (48) d < λ < d < λ < < d n < λ n, ρ > or (49) λ < d < λ < d < < λ n < d n, ρ < So, we have to compute one eigenvalue in each of the intervals (d i,d i+ ), i < n, and a further eigenvalue in (d n, ) or (,d ) he corresponding eigenvector is then given by (45) Evidently, these tasks are easy to parallelize Equations (47) and (45) can also been obtained from the relations [ ] ρ v [ ] [ ] [ ρ ] ρv v λi D ρv I λi D ρvv I [ v (λi D) ] [ ] [ ρ v (λi D) v I λi D (λi D) v I hese are simply block LDL factorizations of the first matrix he first is the well-known one where the factorization is started with the (,) block he second is a backward factorization that is started with the (, ) block Because the determinants of the tridiagonal matrices are all unity, we have ] (4) ρ det(λi D ρvv ) ρ ( ρv (λi D) v)det(λi D) Denoting the eigenvalues of D +ρvv again by λ < λ < < λ n this implies (4) (λ λ j ) ( ρv (λi D) v) (λ d j ) j j ( ) n v n k ρ (λ d j ) λ d k k (λ d j ) ρ n j vk j k j k (λ d j ) Setting λ d k gives (4) (d k λ j ) ρvk j (d k d j ) j j i
8 84 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM or (43) v k ρ ρ (d k λ j ) j (d k d j ) j j i (d k λ j ) k j (d k d j ) k j ρ (d k λ j ) k j (d k d j ) k j (λ j d k ) jk jk+ > (d j d k ) (λ j d k )( ) n k+ jk jk+ herefore, the quantity on the right side is positive, so k (d k λ j ) n (λ j d k ) (44) v k j jk (d k d j ) (d j d k ) ρ k j jk+ (d j d k )( ) n k (Similar arguments hold if ρ < ) hus, we have the solution of the following inverse eigenvalue problem: Given D diag(d,,d n ) and values λ,,λ n that satisfy (48) Find a vector v [v,,v n ] with positive components v k such that the matrix D+vv has the prescribed eigenvalues λ,,λ n he solution is given by (44) he positivity of the v k makes the solution unique 46 Solving the secular equation In this section we follow closely the exposition of Demmel[3] We consider the computation of the zero of f(λ) in the interval (d i,d i+ ) We assume that ρ We may simply apply Newton s iteration to solve f(λ) However, if we look carefully at Fig 4 then we notice that the tangent at certain points in (d i,d i+ ) crosses the real axis outside this interval his happens in particular if the weights v i or v i+ are small herefore that zero finder has to be adapted in such a way that it captures the poles at the interval points It is relatively straightforward to try the ansatz (45) h(λ) c d i λ + c d i+ λ +c 3 Notice that, given the coefficients c, c, and c 3, the equation h(λ) can easily be solved by means of the equivalent quadratic equation (46) c (d i+ λ)+c (d i λ)+c 3 (d i λ)(d i+ λ) his equation has two zeros Precisly one of them is inside (d i,d i+ ) he coefficients c, c, and c 3 are computed in the following way Let us assume that we have available an approximation λ j to the zero in (d i,d i+ ) We request that h(λ j ) f(λ j ) and h (λ j ) f (λ j ) he exact procedure is as follows We write i v n k vk (47) f(λ) + + +ψ (λ)+ψ (λ) d k λ d k λ k ki+ }{{}}{{} ψ (λ) ψ (λ)
9 47 A FIRS ALGORIHM 85 ψ (λ) is a sum of positive terms and ψ (λ) is a sum of negative terms Both ψ (λ) and ψ (λ) can be computed accurately, whereas adding them would likely provoke cancellation and loss of relative accuracy We now choose c and ĉ such that (48) h (λ) : ĉ + c d i λ satisfies h (λ j ) ψ (λ j ) and h (λ j ) ψ (λ j ) his means that the graphs of h and of ψ are tangent at λ λ j his is similar to Newton s method However in Newton s method a straight line is fitted to the given function he coefficients in (48) are given by c ψ (λ j )(d i λ j ) >, ĉ ψ (λ j ) ψ (λ j )(d i λ j ) i k v k d k d i (d k λ j ) Similarly, the two constants c and ĉ are determined such that (49) h (λ) : ĉ + with the coefficients Finally, we set c d i+ λ satisfies h (λ j ) ψ (λ j ) and h (λ j ) ψ (λ j ) c ψ (λ j )(d i+ λ j ) >, n ĉ ψ (λ j ) ψ (λ j )(d i+ λ j ) vk d k d i+ (d k λ) ki+ (43) h(λ) +h (λ)+h (λ) (+ĉ +ĉ ) + }{{} d i λ + c d i+ λ c 3 his zerofinder is converging quadratically to the desired zero [7] Usually to 3 steps are sufficient to get the zero to machine precision herefore finding a zero only requires O(n) flops hus, finding all zeros costs O(n ) floating point operations c 47 A first algorithm We are now ready to give the divide and conquer algorithm, see Algorithm 4 All steps except step require O(n ) operations to complete he step costs n 3 flops hus, the full divide and conquer algorithm, requires (43) ( n ) 3 (n) n 3 + (n/) n (n/4) n 3 + n3 ( n ) (n/8) 4 3 n3 his serial complexity of the algorithm very often overestimates the computational costs of the algorithm due to significant deflation that is observed surprisingly often
10 86 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM Algorithm 4 he tridiagonal divide and conquer algorithm : Let C n n be a real symmetric tridiagonal matrix his algorithm computes the spectral decomposition of QΛQ, where the diagonal Λ is the matrix of eigenvalues and Q is orthogonal : if is then 3: return (Λ ;Q ) 4: else [ ] O 5: Partition +ρuu O according to (4) 6: Call this algorithm with as input and Q, Λ as output 7: Call this algorithm with as input and Q, Λ as output 8: Form D +ρvv from Λ,Λ,Q,Q according to (44) (46) 9: Find the eigenvalues [ Λ] and the eigenvectors Q of D +ρvv Q O : Form Q Q which are the eigenvectors of : return (Λ; Q) : if O Q 47 A numerical example Let A be a 4 4 matrix (43) A D +vv β +β 5 + β β [ β β ] In this example (that is similar to one in [8]) we want to point at a problem that the divide and conquer algorithm possesses as it is given in Algorithm 4, namely the loss of orthogonality among eigenvectors Before we do some Matlab tests let us look more closely at D and v in (43) his example becomes difficult to solve if β gets very small In Figures 4 to 45 we see graphs of the function f β (λ) that appears in the secular equation for β, β, and β he critical zeros move towards from both sides he weights v v 3 β are however not so small that they should be deflated he following Matlab code shows the problem We execute the commands for β k for k,,,4,8 v [ beta beta ] ; % rank- modification d [, -beta, +beta, 5] ; % diagonal matrix L eig(diag(d) + v*v ) % eigenvalues of the modified matrix e ones(4,); q (d*e -e*l )\(v*e ); % unnormalized eigenvectors cf (55) Q sqrt(diag(q *q)); q q/(e*q ); % normalized eigenvectors norm(q *q-eye(4)) % check for orthogonality We do not bother how we compute the eigenvalues We simply use Matlab s built-in function eig We get the results of able 4
11 47 A FIRS ALGORIHM Figure 4: Secular equation corresponding to (43) for β Figure 43: Secular equation corresponding to (43) for β We observe loss of orthogonality among the eigenvectors as the eigenvalues get closer and closer his may not be surprising as we compute the eigenvectors by formula (45) x (λi D) v (λi D) v If λ λ and λ λ 3 which are almost equal, λ λ 3 then intuitively one expects almost the same eigenvectors We have in fact Q Q I Orthogonality is lost only with respect to the vectors corresponding to the eigenvalues close to
12 88 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM Figure 44: Secular equation corresponding to (43) for β for λ Figure 45: Secular equation corresponding to (43) for β for 9 λ Already Dongarra and Sorensen [4] analyzed this problem In their formulation they normalize the vector v of D +ρvv to have norm unity, v hey stated Lemma 4 Let ( )[ ] (433) q λ v d λ, v d λ,, v n ρ / d n λ f (λ) hen for any λ,µ {d,,d n } we have (434) q λ q µ f(λ) f(µ) λ µ [f (λ)f (µ)] / Proof Observe that hen the proof is straightforward λ µ (d j λ)(d j µ) d j λ d j µ
13 48 HE ALGORIHM OF GU AND EISENSA 89 β λ λ λ 3 λ 4 Q Q I able 4: Loss of orthogonality among the eigenvectors computed by (45) Formula (434) indicates how problems may arise In exact arithmetic, if λ and µ are eigenvalues then f(λ) f(µ) However, in floating point arithmetic this values may be small but nonzero, eg, O(ε) If λ µ is very small as well then we may have trouble! So, a remedy for the problem was for a long time to compute the eigenvalues in doubled precision, so that f(λ) O(ε ) his would counteract a potential O(ε) of λ µ his solution was quite unsatisfactory because doubled precision is in general very slow since it is implemented in software It took a decade until a proper solution was found 48 he algorithm of Gu and Eisenstat Computing eigenvector according to the formula v λ d (435) x α(λi D) v α v n λ d n, α (λi D) v, is bound to fail if λ is very close to a pole d k and the difference λ d k has an error of size O(ε d k ) instead of only O(ε d k λ ) o resolve this problem Gu and Eisenstat [5] found a trick that is at the same time ingenious and simple hey observed that the v k in (44) are very accurately determined by the data d i and λ i herefore, once the eigenvalues are computed accurately a vector ˆv could be computed such that the λ i are accurate eigenvalues of D+ˆvˆv If ˆv approximates well the original v then the new eigenvectors will be the exact eigenvectors of a slightly modified eigenvalue problem, which is all we can hope for he zeros of the secular equation can be computed accurately by the method presented in section 46 However, a shift of variables is necessary In the interval (d i,d i+ ) the origin of the real axis is moved to d i if λ i is closer to d i than to d i+, ie, if f((d i +d i+ )/) > Otherwise, the origin is shifted to d i+ his shift of the origin avoids the computation of the smallest difference d i λ (or d i+ λ) in (435), thus avoiding cancellation in this most sensitive quantity Equation (46) can be rewritten as (436) (c i+ +c i +c 3 i i+ ) (c }{{} +c +c 3 ( i + i+ )) η + c }{{} 3 η }{{}, b a c where i d i λ j, i+ d i+ λ j, and λ j+ λ j +η is the next approximate zero With equations (48) (43) the coefficients in (436) get (437) a c +c +c 3 ( i + i+ ) (+Ψ +Ψ )( i + i+ ) (Ψ +Ψ ) i i+, b c i+ +c i +c 3 i i+ i i+ (+Ψ +Ψ ), c c 3 +Ψ +Ψ i Ψ i+ Ψ
14 9 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM If we are looking for a zero that is closer to d i than to d i+ then we move the origin to λ j, ie, we have eg i λ j he solution of (436) that lies inside the interval is [7] (438) η a a 4bc c, if a, b a+, a 4bc if a > he following algorithm shows how step 9 of the tridiagonal divide and conquer algorithm 4 must be implemented Algorithm 4 A stable eigensolver for D +vv : hisalgorithmstablycomputesthespectraldecompositionofd+vv QΛQ where D diag(d,d n ), v [v,,v n ] R n, Λ diag(λ,λ n ), and Q [q,,q n ] : d i+ d n + v 3: In each interval (d i,d i+ ) compute the zero λ i of the secular equation f(λ) 4: Use the formula (44) to compute the vector ˆv such that the λ i are the exact eigenvalues of D +ˆvˆv 5: In each interval (d i,d i+ ) compute the eigenvectors of D + ˆvˆv according to (45), 6: return (Λ;Q) q i (λ ii D) ˆv (λ i I D) ˆv 48 A numerical example [continued] We continue the discussion of the example on page 86 where the eigenvalue problem of (439) A D +vv β +β + β ] β [ β β 5 he Matlab code that we showed did not give orthogonal eigenvectors We show in the following script that the formulae (44) really solve the problem dlam zeros(n,); for k:n, [dlam(k), dvec(:,k)] zerodandc(d,v,k); V ones(n,); for k:n, V(k) prod(abs(dvec(k,:)))/prod(d(k) - d(:k-))/prod(d(k+:n) - d(k)); V(k) sqrt(v(k)); Q (dvec)\(v*e ); diagq sqrt(diag(q *Q)); Q Q/(e*diagq );
15 48 HE ALGORIHM OF GU AND EISENSA 9 for k:n, if dlam(k)>, dlam(k) dlam(k) + d(k); else dlam(k) d(k+) + dlam(k); norm(q *Q-eye(n)) norm((diag(d) + v*v )*Q - Q*diag(dlam )) A zero finder returns for each interval the quantity λ i d i and the vector [d λ i,,d n λ i ] to high precision hese vector elements have been computed as (d k d i ) (λ i d i ) he zerofinder of Li [7] has been employed here At the of this section we list the zerofinder written in Matlab that was used here he formulae (437) and (438) have been used to solve the quadratic equation (436) Notice that only one of the while loops is traversed, deping on if the zero is closer to the pole on the left or to the right of the interval he v k of formula (44) are computed next Q contains the eigenvectors β Algorithm Q Q I AQ QΛ I II I II I II I II able 4: Loss of orthogonality among the eigenvectors computed by the straightforward algorithm (I) and the Gu-Eisenstat approach (II) Again we ran the code for β k for k,,,4,8 he numbers in able 4 confirm that the new formulae are much more accurate than the straight forward ones he norms of the errors obtained for the Gu-Eisenstat algorithm always are in the order of machine precision, ie, 6 function [lambda,dl] zerodandc(d,v,i) % ZERODANDC - Computes eigenvalue lambda in the i-th interval % (d(i), d(i+)) with Li s middle way zero finder % dl is the n-vector [d(n) - lambda] n length(d); di d(i); v v^; if i < n, di d(i+); lambda (di + di)/; else di d(n) + norm(v)^; lambda di; eta ; psi sum((v(:i)^)/(d(:i) - lambda));
16 9 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM psi sum((v(i+:n)^)/(d(i+:n) - lambda)); if + psi + psi >, % zero is on the left half of the interval d d - di; lambda lambda - di; di di - di; di ; while abs(eta) > *eps psi sum(v(:i)/(d(:i) - lambda)); psis sum(v(:i)/((d(:i) - lambda))^); psi sum((v(i+:n))/(d(i+:n) - lambda)); psis sum(v(i+:n)/((d(i+:n) - lambda))^); % Solve for zero Di -lambda; Di di - lambda; a (Di + Di)*( + psi + psi) - Di*Di*(psis + psis); b Di*Di*( + psi + psi); c ( + psi + psi) - Di*psis - Di*psis; if a >, eta (*b)/(a + sqrt(a^ - 4*b*c)); else eta (a - sqrt(a^ - 4*b*c))/(*c); lambda lambda + eta; else % zero is on the right half of the interval d d - di; lambda lambda - di; di di - di; di ; while abs(eta) > *eps psi sum(v(:i)/(d(:i) - lambda)); psis sum(v(:i)/((d(:i) - lambda))^); psi sum((v(i+:n))/(d(i+:n) - lambda)); psis sum(v(i+:n)/((d(i+:n) - lambda))^); % Solve for zero Di di - lambda; Di - lambda; a (Di + Di)*( + psi + psi) - Di*Di*(psis + psis); b Di*Di*( + psi + psi); c ( + psi + psi) - Di*psis - Di*psis; if a >, eta (*b)/(a + sqrt(a^ - 4*b*c)); else eta (a - sqrt(a^ - 4*b*c))/(*c); lambda lambda + eta; dl d - lambda; return
17 BIBLIOGRAPHY 93 Bibliography [] E Anderson, Z Bai, C Bischof, J Demmel, J Dongarra, J D Croz, A Greenbaum, S Hammarling, A McKenney, S Ostrouchov, and D Sorensen, LAPACK Users Guide - Release, SIAM, Philadelphia, PA, 994 (Software and guide are available from Netlib at URL [] J J M Cuppen, A divide and conquer method for the symmetric tridiagonal eigenproblem, Numer Math, 36 (98), pp [3] J W Demmel, Applied Numerical Linear Algebra, SIAM, Philadelphia, PA, 997 [4] J J Dongarra and D C Sorensen, A fully parallel algorithm for the symmetric eigenvalue problem, SIAM J Sci Stat Comput, 8 (987), pp s39 s54 [5] M Gu and S C Eisenstat, A stable and efficient algorithm for the rank-one modification of the symmetric eigenproblem, SIAM J Matrix Anal Appl, 5 (994), pp [6], A divide-and-conquer algorithm for the symmetric tridiagonal eigenproblem, SIAM J Matrix Anal Appl, 6 (995), pp 7 9 [7] R-C Li, Solving secular equations stably and efficiently, echnical Report U-CS- 94-6, University of ennessee, Knoxville, N, Nov 994 LAPACK Working Note No 89 [8] D C Sorensen and P P ang, On the orthogonality of eigenvectors computed by divide-and-conquer techniques, SIAM J Numer Anal, 8 (99), pp
18 94 CHAPER 4 CUPPEN S DIVIDE AND CONQUER ALGORIHM
Cuppen s Divide and Conquer Algorithm
Chapter 5 Cuppen s Divide and Conquer Algorithm In this chapter we deal with an algorithm that is designed for the efficient solution of the symmetric tridiagonal eigenvalue problem a 1 b 1. (5.1) Tx =
More informationDirect methods for symmetric eigenvalue problems
Direct methods for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 4, 2008 1 Theoretical background Posing the question Perturbation theory
More informationExponentials of Symmetric Matrices through Tridiagonal Reductions
Exponentials of Symmetric Matrices through Tridiagonal Reductions Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A simple and efficient numerical algorithm
More informationGeneralized interval arithmetic on compact matrix Lie groups
myjournal manuscript No. (will be inserted by the editor) Generalized interval arithmetic on compact matrix Lie groups Hermann Schichl, Mihály Csaba Markót, Arnold Neumaier Faculty of Mathematics, University
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationThe QR Algorithm. Chapter The basic QR algorithm
Chapter 3 The QR Algorithm The QR algorithm computes a Schur decomposition of a matrix. It is certainly one of the most important algorithm in eigenvalue computations. However, it is applied to dense (or:
More informationChapter 12 Solving secular equations
Chapter 12 Solving secular equations Gérard MEURANT January-February, 2012 1 Examples of Secular Equations 2 Secular equation solvers 3 Numerical experiments Examples of secular equations What are secular
More informationComputing least squares condition numbers on hybrid multicore/gpu systems
Computing least squares condition numbers on hybrid multicore/gpu systems M. Baboulin and J. Dongarra and R. Lacroix Abstract This paper presents an efficient computation for least squares conditioning
More informationThe Algorithm of Multiple Relatively Robust Representations for Multi-Core Processors
Aachen Institute for Advanced Study in Computational Engineering Science Preprint: AICES-2010/09-4 23/September/2010 The Algorithm of Multiple Relatively Robust Representations for Multi-Core Processors
More informationA Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem
A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van
More informationA Parallel Divide and Conquer Algorithm for the Symmetric Eigenvalue Problem on Distributed Memory Architectures. F Tisseur and J Dongarra
A Parallel Divide and Conquer Algorithm for the Symmetric Eigenvalue Problem on Distributed Memory Architectures F Tisseur and J Dongarra 999 MIMS EPrint: 2007.225 Manchester Institute for Mathematical
More informationLAPACK and the BLAS. Chapter LAPACK
Chapter 4 LAPACK and the BLAS 4.1 LAPACK (This section is essentially compiled from the LAPACK User s Guide [1] that is available online from http://www.netlib.org/lapack/lug/.) LAPACK [1] is a library
More informationUsing Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices.
Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. A.M. Matsekh E.P. Shurina 1 Introduction We present a hybrid scheme for computing singular vectors
More informationA note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Akiko Fukuda
Journal of Math-for-Industry Vol 3 (20A-4) pp 47 52 A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Aio Fuuda Received on October 6 200 / Revised on February 7 20 Abstract
More informationPositive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract
Computing the Logarithm of a Symmetric Positive Denite Matrix Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A numerical method for computing the logarithm
More informationarxiv: v2 [math.na] 21 Sep 2015
Forward stable eigenvalue decomposition of rank-one modifications of diagonal matrices arxiv:1405.7537v2 [math.na] 21 Sep 2015 N. Jakovčević Stor a,1,, I. Slapničar a,1, J. L. Barlow b,2 a Faculty of Electrical
More informationA DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION
SIAM J MATRIX ANAL APPL Vol 0, No 0, pp 000 000 c XXXX Society for Industrial and Applied Mathematics A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION WEI XU AND SANZHENG QIAO Abstract This paper
More informationAccurate eigenvalue decomposition of arrowhead matrices and applications
Accurate eigenvalue decomposition of arrowhead matrices and applications Nevena Jakovčević Stor FESB, University of Split joint work with Ivan Slapničar and Jesse Barlow IWASEP9 June 4th, 2012. 1/31 Introduction
More informationAlgorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems
Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems LESLIE FOSTER and RAJESH KOMMU San Jose State University Existing routines, such as xgelsy or xgelsd in LAPACK, for
More informationA Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem
A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Suares Problem Hongguo Xu Dedicated to Professor Erxiong Jiang on the occasion of his 7th birthday. Abstract We present
More informationMatrices, Moments and Quadrature, cont d
Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that
More informationEfficient and Accurate Rectangular Window Subspace Tracking
Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu
More informationLAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.
LAPACK-Style Codes for Pivoted Cholesky and QR Updating Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig 2007 MIMS EPrint: 2006.385 Manchester Institute for Mathematical Sciences School of Mathematics
More informationthat determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that
Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex
More informationSolving large scale eigenvalue problems
arge scale eigenvalue problems, Lecture 4, March 14, 2018 1/41 Lecture 4, March 14, 2018: The QR algorithm http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich E-mail:
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview
More informationTotal least squares. Gérard MEURANT. October, 2008
Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares
More informationA Divide-and-Conquer Method for the Takagi Factorization
A Divide-and-Conquer Method for the Takagi Factorization Wei Xu 1 and Sanzheng Qiao 1, Department of Computing and Software, McMaster University Hamilton, Ont, L8S 4K1, Canada. 1 xuw5@mcmaster.ca qiao@mcmaster.ca
More informationANONSINGULAR tridiagonal linear system of the form
Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal
More informationECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems
ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A
More informationComputing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices
Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 9, 2008 Today
More informationETNA Kent State University
C 8 Electronic Transactions on Numerical Analysis. Volume 17, pp. 76-2, 2004. Copyright 2004,. ISSN 1068-613. etnamcs.kent.edu STRONG RANK REVEALING CHOLESKY FACTORIZATION M. GU AND L. MIRANIAN Abstract.
More informationIterative Algorithm for Computing the Eigenvalues
Iterative Algorithm for Computing the Eigenvalues LILJANA FERBAR Faculty of Economics University of Ljubljana Kardeljeva pl. 17, 1000 Ljubljana SLOVENIA Abstract: - We consider the eigenvalue problem Hx
More informationLAPACK-Style Codes for Pivoted Cholesky and QR Updating
LAPACK-Style Codes for Pivoted Cholesky and QR Updating Sven Hammarling 1, Nicholas J. Higham 2, and Craig Lucas 3 1 NAG Ltd.,Wilkinson House, Jordan Hill Road, Oxford, OX2 8DR, England, sven@nag.co.uk,
More informationAMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems
AMS 209, Fall 205 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems. Overview We are interested in solving a well-defined linear system given
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse
More informationComputing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices
Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 8, 2009 Today
More informationMath 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm
Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm References: Trefethen & Bau textbook Eigenvalue problem: given a matrix A, find
More information6 EIGENVALUES AND EIGENVECTORS
6 EIGENVALUES AND EIGENVECTORS INTRODUCTION TO EIGENVALUES 61 Linear equations Ax = b come from steady state problems Eigenvalues have their greatest importance in dynamic problems The solution of du/dt
More informationEigenvalue and Eigenvector Problems
Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems
More informationScientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix
Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationSection 4.5 Eigenvalues of Symmetric Tridiagonal Matrices
Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationKU Leuven Department of Computer Science
Backward error of polynomial eigenvalue problems solved by linearization of Lagrange interpolants Piers W. Lawrence Robert M. Corless Report TW 655, September 214 KU Leuven Department of Computer Science
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to
More informationNAG Library Routine Document F08JDF (DSTEVR)
F08 Least-squares and Eigenvalue Problems (LAPACK) NAG Library Routine Document (DSTEVR) Note: before using this routine, please read the Users Note for your implementation to check the interpretation
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More informationEigenvalue Problems and Singular Value Decomposition
Eigenvalue Problems and Singular Value Decomposition Sanzheng Qiao Department of Computing and Software McMaster University August, 2012 Outline 1 Eigenvalue Problems 2 Singular Value Decomposition 3 Software
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationEIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4
EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue
More informationModule 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems. Contents
Eigenvalue and Least-squares Problems Module Contents Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems nag nsym gen eig provides procedures for solving nonsymmetric generalized
More informationPerformance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor
Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor Masami Takata 1, Hiroyuki Ishigami 2, Kini Kimura 2, and Yoshimasa Nakamura 2 1 Academic Group of Information
More informationNormalized power iterations for the computation of SVD
Normalized power iterations for the computation of SVD Per-Gunnar Martinsson Department of Applied Mathematics University of Colorado Boulder, Co. Per-gunnar.Martinsson@Colorado.edu Arthur Szlam Courant
More informationArnoldi Methods in SLEPc
Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationBlock Lanczos Tridiagonalization of Complex Symmetric Matrices
Block Lanczos Tridiagonalization of Complex Symmetric Matrices Sanzheng Qiao, Guohong Liu, Wei Xu Department of Computing and Software, McMaster University, Hamilton, Ontario L8S 4L7 ABSTRACT The classic
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More informationNumerical Solution of Linear Eigenvalue Problems
Numerical Solution of Linear Eigenvalue Problems Jessica Bosch and Chen Greif Abstract We review numerical methods for computing eigenvalues of matrices We start by considering the computation of the dominant
More informationOrthogonal iteration to QR
Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation
More informationD. Gimenez, M. T. Camara, P. Montilla. Aptdo Murcia. Spain. ABSTRACT
Accelerating the Convergence of Blocked Jacobi Methods 1 D. Gimenez, M. T. Camara, P. Montilla Departamento de Informatica y Sistemas. Univ de Murcia. Aptdo 401. 0001 Murcia. Spain. e-mail: fdomingo,cpmcm,cppmmg@dif.um.es
More informationSolving linear equations with Gaussian Elimination (I)
Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian
More informationAlgorithms for Solving the Polynomial Eigenvalue Problem
Algorithms for Solving the Polynomial Eigenvalue Problem Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/ Joint work with D. Steven Mackey
More informationarxiv: v1 [math.na] 5 May 2011
ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and
More informationCOMPUTATION OF GAUSS-KRONROD QUADRATURE RULES
MATHEMATICS OF COMPUTATION Volume 69, Number 231, Pages 1035 1052 S 0025-5718(00)01174-1 Article electronically published on February 17, 2000 COMPUTATION OF GAUSS-KRONROD QUADRATURE RULES D.CALVETTI,G.H.GOLUB,W.B.GRAGG,ANDL.REICHEL
More informationMatrix Shapes Invariant under the Symmetric QR Algorithm
NUMERICAL ANALYSIS PROJECT MANUSCRIPT NA-92-12 SEPTEMBER 1992 Matrix Shapes Invariant under the Symmetric QR Algorithm Peter Arbenz and Gene H. Golub NUMERICAL ANALYSIS PROJECT COMPUTER SCIENCE DEPARTMENT
More informationLecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.
Lecture # 11 The Power Method for Eigenvalues Part II The power method find the largest (in magnitude) eigenvalue of It makes two assumptions. 1. A is diagonalizable. That is, A R n n. A = XΛX 1 for some
More informationPreconditioned Parallel Block Jacobi SVD Algorithm
Parallel Numerics 5, 15-24 M. Vajteršic, R. Trobec, P. Zinterhof, A. Uhl (Eds.) Chapter 2: Matrix Algebra ISBN 961-633-67-8 Preconditioned Parallel Block Jacobi SVD Algorithm Gabriel Okša 1, Marián Vajteršic
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationRecent Advances in the Numerical Solution of Quadratic Eigenvalue Problems
Recent Advances in the Numerical Solution of Quadratic Eigenvalue Problems Françoise Tisseur School of Mathematics The University of Manchester ftisseur@ma.man.ac.uk http://www.ma.man.ac.uk/~ftisseur/
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis Volume 8, pp 127-137, 1999 Copyright 1999, ISSN 1068-9613 ETNA etna@mcskentedu BOUNDS FOR THE MINIMUM EIGENVALUE OF A SYMMETRIC TOEPLITZ MATRIX HEINRICH VOSS
More informationSolving large scale eigenvalue problems
arge scale eigenvalue problems, Lecture 5, March 23, 2016 1/30 Lecture 5, March 23, 2016: The QR algorithm II http://people.inf.ethz.ch/arbenz/ewp/ Peter Arbenz Computer Science Department, ETH Zürich
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9
STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9 1. qr and complete orthogonal factorization poor man s svd can solve many problems on the svd list using either of these factorizations but they
More informationSparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices
Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as
More informationKey words. modified Cholesky factorization, optimization, Newton s method, symmetric indefinite
SIAM J. MATRIX ANAL. APPL. c 1998 Society for Industrial and Applied Mathematics Vol. 19, No. 4, pp. 197 111, October 1998 16 A MODIFIED CHOLESKY ALGORITHM BASED ON A SYMMETRIC INDEFINITE FACTORIZATION
More informationDUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM. Geunseop Lee
J. Korean Math. Soc. 0 (0), No. 0, pp. 1 0 https://doi.org/10.4134/jkms.j160152 pissn: 0304-9914 / eissn: 2234-3008 DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM
More informationa 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12
24 8 Matrices Determinant of 2 2 matrix Given a 2 2 matrix [ ] a a A = 2 a 2 a 22 the real number a a 22 a 2 a 2 is determinant and denoted by det(a) = a a 2 a 2 a 22 Example 8 Find determinant of 2 2
More informationDense LU factorization and its error analysis
Dense LU factorization and its error analysis Laura Grigori INRIA and LJLL, UPMC February 2016 Plan Basis of floating point arithmetic and stability analysis Notation, results, proofs taken from [N.J.Higham,
More informationEigenvalues, Eigenvectors, and Diagonalization
Week12 Eigenvalues, Eigenvectors, and Diagonalization 12.1 Opening Remarks 12.1.1 Predicting the Weather, Again Let us revisit the example from Week 4, in which we had a simple model for predicting the
More informationAccuracy and Stability in Numerical Linear Algebra
Proceedings of the 1. Conference on Applied Mathematics and Computation Dubrovnik, Croatia, September 13 18, 1999 pp. 105 112 Accuracy and Stability in Numerical Linear Algebra Zlatko Drmač Abstract. We
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT 10-12 Large-Scale Eigenvalue Problems in Trust-Region Calculations Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug ISSN 1389-6520 Reports of the Department of
More informationDiagonal Representation of Certain Matrices
Diagonal Representation of Certain Matrices Mark Tygert Research Report YALEU/DCS/RR-33 December 2, 2004 Abstract An explicit expression is provided for the characteristic polynomial of a matrix M of the
More informationMath 471 (Numerical methods) Chapter 4. Eigenvalues and Eigenvectors
Math 471 (Numerical methods) Chapter 4. Eigenvalues and Eigenvectors Overlap 4.1 4.3 of Bradie 4.0 Review on eigenvalues and eigenvectors Definition. A nonzero vector v is called an eigenvector of A if
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationNAG Library Routine Document F08FAF (DSYEV)
NAG Library Routine Document (DSYEV) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT -09 Computational and Sensitivity Aspects of Eigenvalue-Based Methods for the Large-Scale Trust-Region Subproblem Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug
More informationEigenproblems II: Computation
Eigenproblems II: Computation CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Eigenproblems II: Computation 1 / 31 Setup A
More informationNotes on Householder QR Factorization
Notes on Householder QR Factorization Robert A van de Geijn Department of Computer Science he University of exas at Austin Austin, X 7872 rvdg@csutexasedu September 2, 24 Motivation A fundamental problem
More informationLinear Algebra and Matrices
Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.
More informationNAG Library Routine Document F08VAF (DGGSVD)
NAG Library Routine Document (DGGSVD) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationOn an Inverse Problem for a Quadratic Eigenvalue Problem
International Journal of Difference Equations ISSN 0973-6069, Volume 12, Number 1, pp. 13 26 (2017) http://campus.mst.edu/ijde On an Inverse Problem for a Quadratic Eigenvalue Problem Ebru Ergun and Adil
More informationComputing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm
Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 19, 2010 Today
More informationThe quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying
I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given
More informationOutline Background Schur-Horn Theorem Mirsky Theorem Sing-Thompson Theorem Weyl-Horn Theorem A Recursive Algorithm The Building Block Case The Origina
A Fast Recursive Algorithm for Constructing Matrices with Prescribed Eigenvalues and Singular Values by Moody T. Chu North Carolina State University Outline Background Schur-Horn Theorem Mirsky Theorem
More informationarxiv: v1 [cs.lg] 26 Jul 2017
Updating Singular Value Decomposition for Rank One Matrix Perturbation Ratnik Gandhi, Amoli Rajgor School of Engineering & Applied Science, Ahmedabad University, Ahmedabad-380009, India arxiv:70708369v
More informationNAG Library Routine Document F08UBF (DSBGVX)
NAG Library Routine Document (DSBGVX) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works
CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The
More information