Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems
|
|
- Ann Nichols
- 5 years ago
- Views:
Transcription
1 Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan April 28, 2011 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
2 Outline 1 Jacobi s orthogonal component correction 2 Davidson s method 3 Jacobi-Davidson method 4 Polynomial Jacobi-Davidson method 5 Non-equivalence deflation T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
3 References SIAM J. Matrix Anal. Vol. 17, No. 2, PP , 1996, Van der Vorst etc. BIT. Vol. 36, No. 3, PP , 1996, Van der Vorst etc. T.-M. Hwang, W.-W. Lin and V. Mehrmann, SIAM J. Sci. Comput. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
4 Some basic theorems Theorem 1 Let X be an eigenspace of A and let X be a basis for X. Then there is a unique matrix L such that The matrix L is given by AX = XL. L = X I AX, where X I is a matrix satisfying X I X = I. If (λ, x) is an eigenpair of A with x X, then (λ, X I x) is an eigenpair of L. Conversely, if (λ, u) is an eigenpair of L, then (λ, Xu) is an eigenpair of A. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
5 Proof: Let X = [x 1 x k ] and Y = AX = [y 1 y k ]. Since y i X and X is a basis for X, there is a unique vector l i such that y i = Xl i. If we set L = [l 1 l k ], then AX = XL and L = X I XL = X I AX. Now let (λ, x) be an eigenpair of A with x X. Then there is a unique vector u such that x = Xu. However, u = X I x. Hence λx = Ax = AXu = XLu λu = λx I x = Lu. Conversely, if Lu = λu, then A(Xu) = (AX)u = (XL)u = X(Lu) = λ(xu), so that (λ, Xu) is an eigenpair of A. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
6 Theorem 2 (Optimal residuals) Let [X X ] be unitary. Let R = AX XL and S H = X H A LX H. Then R and S are minimized when L = X H AX, in which case (a) R = X H AX, (b) S = X H AX, (c) X H R = 0. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
7 Proof: Set Then [ X H X H [ X H X H ] ] [ ˆL H R = G M A [ [ ] ˆL H X X = G M ] [ X H X H X H ] [ X H X X H It implies that [ ] [ ] R = X H R ˆL L =, G which is minimized when L = ˆL = X H AX and min R = G = X H AX. The proof for S is similar. If L = X H AX, then ]. ] [ ˆL L XL = G ]. X H R = X H AX X H XL = X H AX L = 0. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
8 Definition 3 Let X be of full column rank and let X I be a left inverse of X. Then X I AX is a Rayleigh quotient of A. Theorem 4 Let X be orthonormal, A be Hermitian and R = AX XL. If l 1,..., l k are the eigenvalues of L, then there are eigenvalues λ j1,..., λ jk of A such that l i λ ji R 2 and k (l i λ ji ) 2 2 R F. i=1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
9 Jacobi s orthogonal component correction, 1846 Consider the eigenvalue problem [ ] [ 1 α c T A z b F ] [ 1 z ] [ 1 = λ z ], (1) where A is diagonal dominant and α is the largest diagonal element. (1) is equivalent to { λ = α + c T z, (F λi)z = b. Jacobi iteration : (with z 1 = 0) { θk = α + c T z k, (D θ k I)z k+1 = (D F )z k b (2) where D = diag(f ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
10 Davidson s method (1975) Algorithm 1 (Davidson s method) Given unit vector v, set V = [v] Iterate until convergence Compute desired eigenpair (θ, s) of V T AV. Compute u = V s and r = Au θu. If ( r 2 < ε), stop. Solve (D A θi)t = r. Orthog. t V v, V = [V, v] end T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
11 Let u k = (1, z T k )T. Then r k = (A θ k I)u k = [ α θk + c T z k (F θ k I)z k + b Substituting the residual vector r k into linear systems [ α 0 (D A θ k I)t k = r k, where D A = 0 D we get (D θ k I)y k = (F θ k I)z k b From (2) and above equality, we see that ] ], = (D F )z k (D θ k I)z k b (D θ k I)(z k + y k ) = (D F )z k b = (D θ k I)z k+1. This implies that z k+1 = z k + y k as one step of JOCC starting with z k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
12 Jacobi-Davidson method (1996) (θ k, u k ): approx. eigenpair of A, θ k λ, with u k = V k s k, V T k AV ks k = θ k s k and s k 2 = 1. Definition 5 (θ k, u k ) is called a Ritz pair of A. θ k is called a Ritz value and u k is a Ritz vector. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
13 Jacobi-Davidson method (1996) (θ k, u k ): approx. eigenpair of A, θ k λ, with Then u k = V k s k, V T k AV ks k = θ k s k and s k 2 = 1. u T k r k u T k (A θ ki)u k = s T k V T k AV ks k θ k s T k V T k V ks k = 0 r k u k Find the correction t u k such that That is A(u k + t) = λ(u k + t). (A λi)t = λu k Au k = r k + (λ θ k )u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
14 Since (I u k u T k )r k = r k, we have ( I uk u T ) k (A λi)t = rk. Correction equation (I u k u T k )(A θ ki) ( I u k u T k ) t = rk and t u k, T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
15 Solving correction vector t Correction equation: ( I uk u T ) k (A θk I)(I u k u T k )t = r k, t u k. (3) Scheme S OneLS : Use preconditioning iterative approximations, e.g., GMRES, to solve (3). Use a preconditioner M p ( I u k u T k ) M ( I uk u T k ), where M is an approximation of A θ k I. In each of the iterative steps, it needs to solve M p y = b, y u k (4) for a given b. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
16 Since y u k, Eq. (4) can be rewritten as ( I uk u T ) ( k My = b My = u T k My ) u k + b η k u k + b. Hence where y = M 1 b + η k M 1 u k, η k = ut k M 1 b u T k M 1 u k. SSOR preconditioner: Let A θ k I = L + D + U. Then M = (D + ωl)d 1 (D + ωu). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
17 Scheme S T wols : Since t u k, Eq. (3) can be rewritten as (A θ k I)t = ( u T k (A θ ki)t ) u k r k εu k r k. (5) Let t 1 and t 2 be approximated solutions of the following linear systems: (A θ k I)t = r k and (A θ k I)t = u k, respectively. Then the approximated solution t for (5) is t = t 1 + εt 2 for ε = ut k t 1 u T k t. 2 Scheme S OneStep : The approximated solution t for (5) is t = M 1 r k + εm 1 u k for ε = u k M 1 r k u k M 1 u k, where M is an approximation of A θ k I. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
18 Algorithm 2 (Jacobi-Davidson Method) Choose an n-by-m orthonormal matrix V 0 Do k = 0, 1, 2, Compute all the eigenpairs of Vk T AV ks = λs. Select the desired (target) eigenpair (θ k, s k ) with s k 2 = 1. Compute u k = V k s k and r k = (A θ k I)u k. If ( r k 2 < ε), λ = θ k, x = u k, Stop Solve (approximately) a t k u k from (I u k u T k )(A θ ki)(i u k u T k )t = r k. Orthogonalize t k V k v k+1, V k+1 = [V k, v k+1 ] T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
19 Locking: V k with V k V = I k are convergent Schur vectors, i.e., AV k = V k T k for some upper triangular T k. Set V = [V k, V q ] with V V = I k+q in k + 1-th iteration of Jacobi-Davidson Algorithm. Then [ ] [ ] V V AV = = k AV k Vq AV k [ Tk V k AV q 0 V q AV q Vk AV q Vq AV q ]. = T k V q V k T k Vk AV q V q AV q Restarting: Keep the locked Schur vectors as well as the Schur vectors of interest in the subspace and throw away those we are not interested. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
20 Remark 1 If ε = 0, we obtain Davidson method with ( t is NOT orthogonal to u k ) t 1 = (D A θ k I) 1 r. If the linear systems in (6) are exactly solved, then the vector t becomes t = ε(a θ k I) 1 u k u k. (6) Since t is made orthogonal to u k, (6) is equivalent to t = (A θ k I) 1 u k which is equivalent to shift-invert power iteration. Hence it is quadratic convergence. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
21 Consider Ax = λx and assume that λ is simple. Lemma 6 Consider w with w T x 0. Then the map ) ) F p (I xwt w T (A λi) (I xwt x w T x is a bijection from w to w. Proof: Suppose y w and F p y = 0. That is Then it holds that ) ) (I xwt w T (A λi) (I xwt x w T y = 0. x (A λi)y = εx. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
22 Therefore, both y and x belong to the kernel of (A λi) 2. The simplicity of λ implies that y is a scale multiple of x. The fact that y w and x T w 0 implies y = 0, which proves the injectivity of F p. An obvious dimension argument implies bijectivity. Extension F p t Then ) ) (I uut u T (A θi) (I uut u u T t = r, t u, r u. u t u Fp r u. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
23 Theorem 7 Assume that the correction equation ( I uu T ) (A θi) ( I uu T ) t = r, t u (7) is solved exactly in each step of Jacobi-Davidson Algorithm. Assume u k = u x and u T k x has non-trivial limit. Then if u k is sufficiently chosen to x, then u k x with locally quadratical convergence and θ k = u T k Au k λ. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
24 Proof: Suppose Ax = λx with x such that x = u + z for z u. Then (A θi) z = (A θi) u + (λ θ) x = r + (λ θ) x. (8) Consider the exact solution z 1 u of (7): (I P ) (A θi) z 1 = (I P ) r, (9) where P = uu T. Note that (I P ) r = r since u r. Since x (u + z 1 ) = z z 1 and z = x u, for quadratic convergence, it suffices to show that x (u + z 1 ) = z z 1 = O ( z 2). (10) Multiplying (8) by (I P ) and subtracting the result from (9) yields (I P ) (A θi) (z z 1 ) = (λ θ) (I P ) z + (λ θ) (I P ) u. (11) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
25 Multiplying (8) by u T and using r u leads to λ θ = ut (A θi) z u T. (12) x Since u T k x has non-trivial limit, we obtain (λ θ) (I P ) z = u T (A θi) z u T x (I P ) z. (13) From (11), Lemma 6 and (I P ) u = 0, we have [ ] 1 z z 1 = (I P ) (A θ k I) u (λ θ) (I P ) z k [ ] 1 = u T k (I P ) (A θ k I) (A θ ki) z u k u T k x (I P ) z = O ( z 2). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
26 Jacobi Davidson method as on accelerated Newton Scheme Consider Ax = λx and assume that λ is simple. Choose w T x = 1. Consider nonlinear equation F (u) = Au θ(u)u = 0 with θ (u) = wt Au w T u, where u = 1 or w T u = 1. Then F : {u w T u = 1} w. In particular, r F (u) = Au θ (u) u w. Suppose u k x and the next Newton approximation u k+1 : ( 1 F u k+1 = u k u F (u k ) u=uk) is given by u k+1 = u k + t, i.e., t satisfies that ( ) F t = F (u k ) = r. u u=uk T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
27 Since 1 = u T k+1 w = (u k + t) T w = 1 + t T w, it implies that w T t = 0. By the definition of F, we have F u = A θ (u) I ( w T Au ) uw T + ( w T u ) uw T A (w T u) 2 = A θi + wt Au (w T u) 2 uwt uwt A (I w T u = u kw T ) w T (A θ k I). u k Consequently, the Jacobian of F acts on w and is given by ( ) F t = (I u kw T ) u w T (A θ k I) t, t w. u k u=uk Hence the correction equation of Newton method read as t w, (I u kw T ) w T (A θ k I) t = r, u k which is the correction equation of Jacobi-Davidson method in (7) with w = u. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
28 Polynomial Jacobi-Davidson method (θ k, u k ): approx. eigenpair of A(λ) τ k=0 λk A k, θ k λ, with Let Then u k = V k s k, V k A(λ)V ks k = 0 and s k 2 = 1. r k = A(θ k )u k. u k r k = u k A(θ k)u k = s k V k A(θ k)v k s k = 0 r k u k Find the correction t such that That is A(λ)(u k + t) = 0. A(λ)t = A(λ)u k = r k + (A(θ k ) A(λ))u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
29 Let p k = A (θ k )u k ( τ i=1 iθ i 1 k A i ) u k. A(λ) = A λi: A(λ) = A λb: p k = u k, (A(θ k ) A(λ))u k = (λ θ k )u k = (θ k λ k )p k p k = Bu k, (A(θ k ) A(λ))u k = (λ θ k )Bu k = (θ k λ)p k A(λ) = τ i=0 λi A i with τ 2: [ (A(θ k ) A(λ))u k = (θ k λ)a (θ k ) 1 ] 2 (θ k λ) 2 A (ξ k ) u k = (θ k λ)p k 1 2 (θ k λ) 2 A (ξ k )u k T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
30 Hence A(λ)t = r k + (θ k λ)p k 1 2 (θ k λ) 2 A (ξ k )u k Since r k u k, we have ( I p ku ) k u k p A(λ)t = r k 12 ( (θ k λ) 2 I p ku ) k k u k p A (ξ k )u k. k Correction equation: ( I p ku ) k u k p A(θ k )(I u k u k )t = r k and t u k, k or ( I p ku ) ( k u k p (A θ k B) I u kp ) k k p k u t = r k and t B u k, k with symmetric positive definite matrix B. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
31 Algorithm 3 (Jacobi-Davidson Algorithm for solving A(λ)x = 0) Choose an n-by-m orthonormal matrix V 0 Do k = 0, 1, 2, Compute all the eigenpairs of Vk A(λ)V k = 0. Select the desired (target) eigenpair (θ k, s k ) with s k 2 = 1. Compute u k = V k s k, r k = A(θ k )u k and p k = A (θ k )u k. If ( r k 2 < ε), λ = θ k, x = u k, Stop Solve (approximately) a t k u k from (I p ku T k u T k p )A(θ k )(I u k u T k k )t = r k. Orthogonalize t k V k v k+1, V k+1 = [V k, v k+1 ] p k = A (θ k )u k ( τ i=1 iθ i 1 k A i ) u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
32 Non-equivalence deflation of quadratic eigenproblems Let λ 1 be a real eigenvalue of Q(λ) λ 2 M + λc + K and x 1, z 1 be the associated right and left eigenvectors, respectively, with z T 1 Kx 1 = 1. Let θ 1 = (z T 1 Mx 1 ) 1. We introduce a deflated quadratic eigenproblem [ ] Q(λ)x λ 2 M + λ C + K x = 0, where M = M θ 1 Mx 1 z T 1 M, C = C + θ 1 λ 1 (Mx 1 z T 1 K + Kx 1 z T 1 M), K = K θ 1 λ 2 Kx 1 z1 T K. 1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
33 Complex deflation Let λ 1 = α 1 + iβ 1 be a complex eigenvalue of Q(λ) and x 1 = x 1R + ix 1I, z 1 = z 1R + iz 1I be the associated right and left eigenvectors, respectively, such that Z T 1 KX 1 = I 2, where X 1 = [x 1R, x 1I ] and Z 1 = [z 1R, z 1I ]. Let Θ 1 = (Z T 1 MX 1 ) 1. Then we introduce a deflated quadratic eigenproblem with M = M MX 1 Θ 1 Z1 T M, C = C + MX 1 Θ 1 Λ T 1 Z1 T K + KX 1 Λ 1 1 ΘT 1 Z1 T M, K = K KX 1 Λ 1 1 Θ 1Λ T 1 Z1 T K [ ] α1 β in which Λ 1 = 1. β 1 α 1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
34 Theorem 8 (i) Let λ 1 be a simple real eigenvalue of Q(λ). Then the spectrum of Q(λ) is given by provided that λ 2 1 θ 1. (σ(q(λ)) {λ 1 }) { } (ii) Let λ 1 be a simple complex eigenvalue of Q(λ). Then the spectrum of Q(λ) is given by ( σ(q(λ)) {λ1, λ 1 } ) {, } provided that Λ 1 Λ T 1 Θ 1. Furthermore, in both cases (i) and (ii), if λ 2 λ 1 and (λ 2, x 2 ) is an eigenpair of Q(λ) then the pair (λ 2, x 2 ) is also an eigenpair of Q(λ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
35 Suppose that M, C, K are symmetric. Given an eigenmatrix pair (Λ 1, X 1 ) R r r R n r of Q(λ), where Λ 1 is nonsingular and X 1 satisfies X T 1 KX 1 = I r, Θ 1 := (X T 1 MX 1 ) 1. We define Q(λ) := λ 2 M + λ C + K, where M := M MX 1 Θ 1 X1 T M, C := C + MX 1 Θ 1 Λ T 1 X1 T K + KX 1 Λ 1 1 Θ 1X1 T M, K := K KX 1 Λ 1 1 Θ 1Λ T 1 X1 T K. Theorem 9 Suppose that Θ 1 Λ 1 Λ T 1 is nonsingular. Then the eigenvalues of the real symmetric quadratic pencil Q(λ) are the same as those of Q(λ) except that the eigenvalues of Λ 1, which are closed under complex conjugation, are replaced by r infinities. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
36 Proof: Since (Λ 1, X 1 ) is an eigenmatrix pair of Q(λ), i.e., we have MX 1 Λ CX 1 Λ 1 + KX 1 = 0, Q(λ) = Q(λ) + [MX 1 (λi r + Λ 1 ) + CX 1 ] Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M) = Q(λ) + Q(λ)X 1 (λi r Λ 1 ) 1 Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M). By using the identity where R, S T R n m, we have det[ Q(λ)] det(i n + RS) = det(i m + SR), = det[q(λ)]det[i + X 1 (λi r Λ 1 ) 1 Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M)] = det[q(λ)]det[i r + (λi r Λ 1 ) 1 Θ 1 Λ T 1 (I r λλ T 1 Θ 1 1 )] det[q(λ)] = det(λi r Λ 1 ) det(θ 1Λ T 1 Λ 1 ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
37 Since (Θ 1 Λ 1 Λ T 1 ) Rr r is nonsingular, we have det(θ 1 Λ T 1 Λ 1 ) 0. Therefore, Q(λ) has the same eigenvalues as Q(λ) except that r eigenvalues of Λ 1 are replaced by r infinities. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
38 Non-equiv. deflation for cubic poly. eigenproblems Let (Λ, V u ) R r r R n r be an eigenmatrix pair of A(λ) λ 3 A 3 + λ 2 A 2 + λa 1 + A 0 (14) with V T u V u = I r and 0 / σ(λ), i.e., A 3 V u Λ 3 + A 2 V u Λ 2 + A 1 V u Λ + A 0 V u = 0. (15) Define a new deflated cubic eigenvalue problem by Ã(λ)u = (λ 3 à 3 + λ 2 à 2 + λã1 + Ã0)u = 0, (16) where à 0 = A 0, à 1 = A 1 (A 1 V u V T u + A 2 V u ΛV T u + A 3 V u Λ 2 V T u ), à 2 = A 2 (A 2 V u V T u + A 3 V u ΛV T u ), à 3 = A 3 A 3 V u V T u. (17) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
39 Lemma 10 Let A(λ) and Ã(λ) be cubic pencils given by (14) and (16), respectively. Then it holds Ã(λ) = A(λ) ( I n λv u (λi r Λ) 1 V T u ). (18) Theorem 11 Let (Λ, V u ) be an eigenmatrix pair of A(λ) with V T u V u = I r. Then (i) (σ(a(λ)) σ(λ)) { } = σ(ã(λ)). (ii) Let (µ, z) be an eigenpair of A(λ) with z 2 = 1 and µ / σ(λ). Define z = (I n µv u Λ 1 V T u )z T (µ)z. (19) Then (µ, z) is an eigenpair of Ã(λ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
40 Proof of Lemma: Using (17) and (15), and the fundamental matrix calculation, we have Ã(λ) = A(λ) λ ( λ 2 A 3 V F VF T + λa 2 V F VF T + λa 3 V F ΛVF T +A 1 V F VF T + A 2 V F ΛVF T + A 3 V F Λ 2 VF T ) = A(λ) λ ( A 3 V F (λi r Λ) 3 (λi r Λ) 1 VF T +3A 3 V F Λ(λI r Λ) 2 (λi r Λ) 1 VF T +3A 3 V F Λ 2 (λi r Λ)(λI r Λ) 1 VF T +A 2 V F (λi r Λ) 2 (λi r Λ) 1 VF T +2A 2 V F Λ(λI r Λ)(λI r Λ) 1 VF T +A 1 V F (λi r Λ)(λI r Λ) 1 VF T ) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
41 Ã(λ) = A(λ) λ {[ A 3 V F (λ 3 I r Λ 3 ) + A 2 V F (λ 2 I r Λ 2 ) +A 1 V F (λi r Λ) + A 0 V F A 0 V F ] (λi r Λ) 1 VF T } = A(λ) λ [ A(λ)V F (λi r Λ) 1 VF T ] = A(λ) [ I n λv F (λi r Λ) 1 VF T ]. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
42 Proof of Theorem: (i) Using the identity det(i n + RS) = det(i m + SR) and Lemma 10, we have det(ã(λ)) = det(a(λ)) det ( I n λv F (λi r Λ) 1 VF T ) = det(a(λ)) det ( I n λ(λi r Λ) 1) = det(a(λ)) det(λi r Λ) 1 det( Λ). Since 0 / σ(λ), det( Λ) 0. Thus, Ã(λ) and A(λ) have the same finite spectrum except the eigenvalues in σ(λ). Furthermore, dividing Eq. (16) by λ 3 and using the fact that à 3 V F = (A 3 A 3 V F V T F )V F = 0, we see that (diag r {,, }, V F ) is an eigenmatrix pair of Ã(λ) corresponding to infinite eigenvalues. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
43 (ii) Since µ / σ(λ), the matrix T (µ) = (I µv F Λ 1 VF T ) in (19) is invertible with the inverse T (µ) 1 = I n µv F (µi r Λ) 1 VF T. (20) From Lemma 10, we have Ã(µ) z = A(µ) [ I n µv F (µi r Λ) 1 VF T ] [ In µv F Λ 1 VF T ] z = 0. This completes the proof. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42
Jacobi Davidson Methods for Cubic Eigenvalue Problems (March 11, 2004)
Jacobi Davidson Methods for Cubic Eigenvalue Problems (March 11, 24) Tsung-Min Hwang 1, Wen-Wei Lin 2, Jinn-Liang Liu 3, Weichung Wang 4 1 Department of Mathematics, National Taiwan Normal University,
More informationITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD
ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation
More informationA Jacobi Davidson Method for Nonlinear Eigenproblems
A Jacobi Davidson Method for Nonlinear Eigenproblems Heinrich Voss Section of Mathematics, Hamburg University of Technology, D 21071 Hamburg voss @ tu-harburg.de http://www.tu-harburg.de/mat/hp/voss Abstract.
More informationECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems
ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A
More informationDavidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD
Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology The Davidson method is a popular technique to compute a few of the smallest (or largest)
More informationKrylov Subspace Methods for Large/Sparse Eigenvalue Problems
Krylov Subspace Methods for Large/Sparse Eigenvalue Problems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan April 17, 2012 T.-M. Huang (Taiwan Normal University) Krylov
More informationReduction of nonlinear eigenproblems with JD
Reduction of nonlinear eigenproblems with JD Henk van der Vorst H.A.vanderVorst@math.uu.nl Mathematical Institute Utrecht University July 13, 2005, SIAM Annual New Orleans p.1/16 Outline Polynomial Eigenproblems
More informationEigenvalues and Eigenvectors
Eigenvalues and Eigenvectors week -2 Fall 26 Eigenvalues and eigenvectors The most simple linear transformation from R n to R n may be the transformation of the form: T (x,,, x n ) (λ x, λ 2,, λ n x n
More informationLINEAR ALGEBRA REVIEW
LINEAR ALGEBRA REVIEW JC Stuff you should know for the exam. 1. Basics on vector spaces (1) F n is the set of all n-tuples (a 1,... a n ) with a i F. It forms a VS with the operations of + and scalar multiplication
More informationEIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..
EIGENVALUE PROBLEMS Background on eigenvalues/ eigenvectors / decompositions Perturbation analysis, condition numbers.. Power method The QR algorithm Practical QR algorithms: use of Hessenberg form and
More informationQR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR
QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique
More informationEIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4
EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationHARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM
HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter
More informationEigenvalues and eigenvectors
Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors
More informationof dimension n 1 n 2, one defines the matrix determinants
HARMONIC RAYLEIGH RITZ FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter eigenvalue
More informationOVERVIEW. Basics Generalized Schur Form Shifts / Generalized Shifts Partial Generalized Schur Form Jacobi-Davidson QZ Algorithm
OVERVIEW Basics Generalized Schur Form Shifts / Generalized Shifts Partial Generalized Schur Form Jacobi-Davidson QZ Algorithm BASICS Generalized Eigenvalue Problem Ax lbx We call (, x) Left eigenpair
More informationA Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation
A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition
More informationKrylov subspace projection methods
I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks
More informationHOMOGENEOUS JACOBI DAVIDSON. 1. Introduction. We study a homogeneous Jacobi Davidson variant for the polynomial eigenproblem
HOMOGENEOUS JACOBI DAVIDSON MICHIEL E. HOCHSTENBACH AND YVAN NOTAY Abstract. We study a homogeneous variant of the Jacobi Davidson method for the generalized and polynomial eigenvalue problem. Special
More informationPreconditioned inverse iteration and shift-invert Arnoldi method
Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical
More informationThe quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying
I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture
More informationLec 2: Mathematical Economics
Lec 2: Mathematical Economics to Spectral Theory Sugata Bag Delhi School of Economics 24th August 2012 [SB] (Delhi School of Economics) Introductory Math Econ 24th August 2012 1 / 17 Definition: Eigen
More informationEigenvalue Problems CHAPTER 1 : PRELIMINARIES
Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14
More informationMATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.
MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:
More informationAlternative correction equations in the Jacobi-Davidson method
Chapter 2 Alternative correction equations in the Jacobi-Davidson method Menno Genseberger and Gerard Sleijpen Abstract The correction equation in the Jacobi-Davidson method is effective in a subspace
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationEigenvalues and Eigenvectors
/88 Chia-Ping Chen Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Eigenvalue Problem /88 Eigenvalue Equation By definition, the eigenvalue equation for matrix
More information(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.
1 Which of the following statements is always true? I The null space of an m n matrix is a subspace of R m II If the set B = {v 1,, v n } spans a vector space V and dimv = n, then B is a basis for V III
More informationA Jacobi-Davidson method for two real parameter nonlinear eigenvalue problems arising from delay differential equations
A Jacobi-Davidson method for two real parameter nonlinear eigenvalue problems arising from delay differential equations Heinrich Voss voss@tuhh.de Joint work with Karl Meerbergen (KU Leuven) and Christian
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationMultiparameter eigenvalue problem as a structured eigenproblem
Multiparameter eigenvalue problem as a structured eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana This is joint work with M Hochstenbach Będlewo, 2932007 1/28 Overview Introduction
More informationScaling, Sensitivity and Stability in Numerical Solution of the Quadratic Eigenvalue Problem
Scaling, Sensitivity and Stability in Numerical Solution of the Quadratic Eigenvalue Problem Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/
More informationNonlinear Eigenvalue Problems
115 Nonlinear Eigenvalue Problems Heinrich Voss Hamburg University of Technology 115.1 Basic Properties........................................ 115-2 115.2 Analytic matrix functions.............................
More informationIterative techniques in matrix algebra
Iterative techniques in matrix algebra Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan September 12, 2015 Outline 1 Norms of vectors and matrices 2 Eigenvalues and
More informationMICHIEL E. HOCHSTENBACH
VARIATIONS ON HARMONIC RAYLEIGH RITZ FOR STANDARD AND GENERALIZED EIGENPROBLEMS MICHIEL E. HOCHSTENBACH Abstract. We present several variations on the harmonic Rayleigh Ritz method. First, we introduce
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 16: Eigenvalue Problems; Similarity Transformations Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 18 Eigenvalue
More informationA Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation
A Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation Zih-Hao Wei 1, Feng-Nan Hwang 1, Tsung-Ming Huang 2, and Weichung Wang 3 1 Department of
More informationNumerical Methods for Solving Large Scale Eigenvalue Problems
Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems
More informationJUST THE MATHS UNIT NUMBER 9.8. MATRICES 8 (Characteristic properties) & (Similarity transformations) A.J.Hobson
JUST THE MATHS UNIT NUMBER 9.8 MATRICES 8 (Characteristic properties) & (Similarity transformations) by A.J.Hobson 9.8. Properties of eigenvalues and eigenvectors 9.8. Similar matrices 9.8.3 Exercises
More informationOn the Modification of an Eigenvalue Problem that Preserves an Eigenspace
Purdue University Purdue e-pubs Department of Computer Science Technical Reports Department of Computer Science 2009 On the Modification of an Eigenvalue Problem that Preserves an Eigenspace Maxim Maumov
More informationInexact inverse iteration with preconditioning
Department of Mathematical Sciences Computational Methods with Applications Harrachov, Czech Republic 24th August 2007 (joint work with M. Robbé and M. Sadkane (Brest)) 1 Introduction 2 Preconditioned
More informationIterative projection methods for sparse nonlinear eigenvalue problems
Iterative projection methods for sparse nonlinear eigenvalue problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Iterative projection
More informationLecture 3: Inexact inverse iteration with preconditioning
Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method
More informationEfficient Methods For Nonlinear Eigenvalue Problems. Diploma Thesis
Efficient Methods For Nonlinear Eigenvalue Problems Diploma Thesis Timo Betcke Technical University of Hamburg-Harburg Department of Mathematics (Prof. Dr. H. Voß) August 2002 Abstract During the last
More informationEigenvalue and Eigenvector Problems
Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems
More informationFAME-matlab Package: Fast Algorithm for Maxwell Equations Tsung-Ming Huang
FAME-matlab Package: Fast Algorithm for Maxwell Equations Tsung-Ming Huang Modelling, Simulation and Analysis of Nonlinear Optics, NUK, September, 4-8, 2015 1 2 FAME group Wen-Wei Lin Department of Applied
More informationJacobi-Davidson Algorithm for Locating Resonances in a Few-Body Tunneling System
Jacobi-Davidson Algorithm for Locating Resonances in a Few-Body Tunneling System Bachelor Thesis in Physics and Engineering Physics Gustav Hjelmare Jonathan Larsson David Lidberg Sebastian Östnell Department
More informationResearch Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics
Research Matters February 25, 2009 The Nonlinear Eigenvalue Problem Nick Higham Part III Director of Research School of Mathematics Françoise Tisseur School of Mathematics The University of Manchester
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationMatrix Theory. A.Holst, V.Ufnarovski
Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different
More informationMATH 221, Spring Homework 10 Solutions
MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the
More informationNonlinear Eigenvalue Problems: An Introduction
Nonlinear Eigenvalue Problems: An Introduction Cedric Effenberger Seminar for Applied Mathematics ETH Zurich Pro*Doc Workshop Disentis, August 18 21, 2010 Cedric Effenberger (SAM, ETHZ) NLEVPs: An Introduction
More informationModel order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers
MAX PLANCK INSTITUTE International Conference on Communications, Computing and Control Applications March 3-5, 2011, Hammamet, Tunisia. Model order reduction of large-scale dynamical systems with Jacobi-Davidson
More informationA Jacobi Davidson-type projection method for nonlinear eigenvalue problems
A Jacobi Davidson-type projection method for nonlinear eigenvalue problems Timo Betce and Heinrich Voss Technical University of Hamburg-Harburg, Department of Mathematics, Schwarzenbergstrasse 95, D-21073
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationLinear Algebra Lecture Notes-II
Linear Algebra Lecture Notes-II Vikas Bist Department of Mathematics Panjab University, Chandigarh-64 email: bistvikas@gmail.com Last revised on March 5, 8 This text is based on the lectures delivered
More informationGoal: to construct some general-purpose algorithms for solving systems of linear Equations
Chapter IV Solving Systems of Linear Equations Goal: to construct some general-purpose algorithms for solving systems of linear Equations 4.6 Solution of Equations by Iterative Methods 4.6 Solution of
More informationThe Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix
The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix Chun-Yueh Chiang Center for General Education, National Formosa University, Huwei 632, Taiwan. Matthew M. Lin 2, Department of
More informationDIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix
DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationUniversity of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm
University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm Name: The proctor will let you read the following conditions
More informationEigenvalues and Eigenvectors
Chapter 1 Eigenvalues and Eigenvectors Among problems in numerical linear algebra, the determination of the eigenvalues and eigenvectors of matrices is second in importance only to the solution of linear
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More information235 Final exam review questions
5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an
More informationMath 215 HW #9 Solutions
Math 5 HW #9 Solutions. Problem 4.4.. If A is a 5 by 5 matrix with all a ij, then det A. Volumes or the big formula or pivots should give some upper bound on the determinant. Answer: Let v i be the ith
More informationEigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.
Chapter 3 Linear Algebra In this Chapter we provide a review of some basic concepts from Linear Algebra which will be required in order to compute solutions of LTI systems in state space form, discuss
More informationDetermining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form. Jay Daigle Advised by Stephan Garcia
Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form Jay Daigle Advised by Stephan Garcia April 4, 2008 2 Contents Introduction 5 2 Technical Background 9 3
More informationLU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU
LU Factorization A m n matri A admits an LU factorization if it can be written in the form of Where, A = LU L : is a m m lower triangular matri with s on the diagonal. The matri L is invertible and is
More informationMath 489AB Exercises for Chapter 2 Fall Section 2.3
Math 489AB Exercises for Chapter 2 Fall 2008 Section 2.3 2.3.3. Let A M n (R). Then the eigenvalues of A are the roots of the characteristic polynomial p A (t). Since A is real, p A (t) is a polynomial
More informationMath 310 Final Exam Solutions
Math 3 Final Exam Solutions. ( pts) Consider the system of equations Ax = b where: A, b (a) Compute deta. Is A singular or nonsingular? (b) Compute A, if possible. (c) Write the row reduced echelon form
More informationProblems of Eigenvalues/Eigenvectors
5 Problems of Eigenvalues/Eigenvectors Reveiw of Eigenvalues and Eigenvectors Gerschgorin s Disk Theorem Power and Inverse Power Methods Jacobi Transform for Symmetric Matrices Singular Value Decomposition
More informationComputable error bounds for nonlinear eigenvalue problems allowing for a minmax characterization
Computable error bounds for nonlinear eigenvalue problems allowing for a minmax characterization Heinrich Voss voss@tuhh.de Joint work with Kemal Yildiztekin Hamburg University of Technology Institute
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationLinear System Theory
Linear System Theory Wonhee Kim Lecture 4 Apr. 4, 2018 1 / 40 Recap Vector space, linear space, linear vector space Subspace Linearly independence and dependence Dimension, Basis, Change of Basis 2 / 40
More informationEfficient computation of transfer function dominant poles of large second-order dynamical systems
Chapter 6 Efficient computation of transfer function dominant poles of large second-order dynamical systems Abstract This chapter presents a new algorithm for the computation of dominant poles of transfer
More informationEE5120 Linear Algebra: Tutorial 7, July-Dec Covers sec 5.3 (only powers of a matrix part), 5.5,5.6 of GS
EE5 Linear Algebra: Tutorial 7, July-Dec 7-8 Covers sec 5. (only powers of a matrix part), 5.5,5. of GS. Prove that the eigenvectors corresponding to different eigenvalues are orthonormal for unitary matrices.
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationTsung-Ming Huang. Matrix Computation, 2016, NTNU
Tsung-Ming Huang Matrix Computation, 2016, NTNU 1 Plan Gradient method Conjugate gradient method Preconditioner 2 Gradient method 3 Theorem Ax = b, A : s.p.d Definition A : symmetric positive definite
More informationLARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems
More informationProblems of Eigenvalues/Eigenvectors
67 Problems of Eigenvalues/Eigenvectors Reveiw of Eigenvalues and Eigenvectors Gerschgorin s Disk Theorem Power and Inverse Power Methods Jacobi Transform for Symmetric Matrices Spectrum Decomposition
More informationMATH Spring 2011 Sample problems for Test 2: Solutions
MATH 304 505 Spring 011 Sample problems for Test : Solutions Any problem may be altered or replaced by a different one! Problem 1 (15 pts) Let M, (R) denote the vector space of matrices with real entries
More informationMath 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018
1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)
More informationProofs for Quizzes. Proof. Suppose T is a linear transformation, and let A be a matrix such that T (x) = Ax for all x R m. Then
Proofs for Quizzes 1 Linear Equations 2 Linear Transformations Theorem 1 (2.1.3, linearity criterion). A function T : R m R n is a linear transformation if and only if a) T (v + w) = T (v) + T (w), for
More informationPart IA. Vectors and Matrices. Year
Part IA Vectors and Matrices Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2018 Paper 1, Section I 1C Vectors and Matrices For z, w C define the principal value of z w. State de Moivre s
More informationReview Notes for Linear Algebra True or False Last Updated: January 25, 2010
Review Notes for Linear Algebra True or False Last Updated: January 25, 2010 Chapter 3 [ Eigenvalues and Eigenvectors ] 31 If A is an n n matrix, then A can have at most n eigenvalues The characteristic
More informationACM 104. Homework Set 5 Solutions. February 21, 2001
ACM 04 Homework Set 5 Solutions February, 00 Franklin Chapter 4, Problem 4, page 0 Let A be an n n non-hermitian matrix Suppose that A has distinct eigenvalues λ,, λ n Show that A has the eigenvalues λ,,
More informationLARGE SPARSE EIGENVALUE PROBLEMS
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems
More informationthat determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that
Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex
More informationInverse Eigenvalue Problem with Non-simple Eigenvalues for Damped Vibration Systems
Journal of Informatics Mathematical Sciences Volume 1 (2009), Numbers 2 & 3, pp. 91 97 RGN Publications (Invited paper) Inverse Eigenvalue Problem with Non-simple Eigenvalues for Damped Vibration Systems
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationMAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:
MAC Module Eigenvalues and Eigenvectors Learning Objectives Upon completing this module, you should be able to: Solve the eigenvalue problem by finding the eigenvalues and the corresponding eigenvectors
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationABSTRACT. Professor G.W. Stewart
ABSTRACT Title of dissertation: Residual Arnoldi Methods : Theory, Package, and Experiments Che-Rung Lee, Doctor of Philosophy, 2007 Dissertation directed by: Professor G.W. Stewart Department of Computer
More informationA Note on Inverse Iteration
A Note on Inverse Iteration Klaus Neymeyr Universität Rostock, Fachbereich Mathematik, Universitätsplatz 1, 18051 Rostock, Germany; SUMMARY Inverse iteration, if applied to a symmetric positive definite
More informationMatrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland
Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii
More information