Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems

Size: px
Start display at page:

Download "Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems"

Transcription

1 Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan April 28, 2011 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

2 Outline 1 Jacobi s orthogonal component correction 2 Davidson s method 3 Jacobi-Davidson method 4 Polynomial Jacobi-Davidson method 5 Non-equivalence deflation T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

3 References SIAM J. Matrix Anal. Vol. 17, No. 2, PP , 1996, Van der Vorst etc. BIT. Vol. 36, No. 3, PP , 1996, Van der Vorst etc. T.-M. Hwang, W.-W. Lin and V. Mehrmann, SIAM J. Sci. Comput. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

4 Some basic theorems Theorem 1 Let X be an eigenspace of A and let X be a basis for X. Then there is a unique matrix L such that The matrix L is given by AX = XL. L = X I AX, where X I is a matrix satisfying X I X = I. If (λ, x) is an eigenpair of A with x X, then (λ, X I x) is an eigenpair of L. Conversely, if (λ, u) is an eigenpair of L, then (λ, Xu) is an eigenpair of A. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

5 Proof: Let X = [x 1 x k ] and Y = AX = [y 1 y k ]. Since y i X and X is a basis for X, there is a unique vector l i such that y i = Xl i. If we set L = [l 1 l k ], then AX = XL and L = X I XL = X I AX. Now let (λ, x) be an eigenpair of A with x X. Then there is a unique vector u such that x = Xu. However, u = X I x. Hence λx = Ax = AXu = XLu λu = λx I x = Lu. Conversely, if Lu = λu, then A(Xu) = (AX)u = (XL)u = X(Lu) = λ(xu), so that (λ, Xu) is an eigenpair of A. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

6 Theorem 2 (Optimal residuals) Let [X X ] be unitary. Let R = AX XL and S H = X H A LX H. Then R and S are minimized when L = X H AX, in which case (a) R = X H AX, (b) S = X H AX, (c) X H R = 0. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

7 Proof: Set Then [ X H X H [ X H X H ] ] [ ˆL H R = G M A [ [ ] ˆL H X X = G M ] [ X H X H X H ] [ X H X X H It implies that [ ] [ ] R = X H R ˆL L =, G which is minimized when L = ˆL = X H AX and min R = G = X H AX. The proof for S is similar. If L = X H AX, then ]. ] [ ˆL L XL = G ]. X H R = X H AX X H XL = X H AX L = 0. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

8 Definition 3 Let X be of full column rank and let X I be a left inverse of X. Then X I AX is a Rayleigh quotient of A. Theorem 4 Let X be orthonormal, A be Hermitian and R = AX XL. If l 1,..., l k are the eigenvalues of L, then there are eigenvalues λ j1,..., λ jk of A such that l i λ ji R 2 and k (l i λ ji ) 2 2 R F. i=1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

9 Jacobi s orthogonal component correction, 1846 Consider the eigenvalue problem [ ] [ 1 α c T A z b F ] [ 1 z ] [ 1 = λ z ], (1) where A is diagonal dominant and α is the largest diagonal element. (1) is equivalent to { λ = α + c T z, (F λi)z = b. Jacobi iteration : (with z 1 = 0) { θk = α + c T z k, (D θ k I)z k+1 = (D F )z k b (2) where D = diag(f ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

10 Davidson s method (1975) Algorithm 1 (Davidson s method) Given unit vector v, set V = [v] Iterate until convergence Compute desired eigenpair (θ, s) of V T AV. Compute u = V s and r = Au θu. If ( r 2 < ε), stop. Solve (D A θi)t = r. Orthog. t V v, V = [V, v] end T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

11 Let u k = (1, z T k )T. Then r k = (A θ k I)u k = [ α θk + c T z k (F θ k I)z k + b Substituting the residual vector r k into linear systems [ α 0 (D A θ k I)t k = r k, where D A = 0 D we get (D θ k I)y k = (F θ k I)z k b From (2) and above equality, we see that ] ], = (D F )z k (D θ k I)z k b (D θ k I)(z k + y k ) = (D F )z k b = (D θ k I)z k+1. This implies that z k+1 = z k + y k as one step of JOCC starting with z k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

12 Jacobi-Davidson method (1996) (θ k, u k ): approx. eigenpair of A, θ k λ, with u k = V k s k, V T k AV ks k = θ k s k and s k 2 = 1. Definition 5 (θ k, u k ) is called a Ritz pair of A. θ k is called a Ritz value and u k is a Ritz vector. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

13 Jacobi-Davidson method (1996) (θ k, u k ): approx. eigenpair of A, θ k λ, with Then u k = V k s k, V T k AV ks k = θ k s k and s k 2 = 1. u T k r k u T k (A θ ki)u k = s T k V T k AV ks k θ k s T k V T k V ks k = 0 r k u k Find the correction t u k such that That is A(u k + t) = λ(u k + t). (A λi)t = λu k Au k = r k + (λ θ k )u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

14 Since (I u k u T k )r k = r k, we have ( I uk u T ) k (A λi)t = rk. Correction equation (I u k u T k )(A θ ki) ( I u k u T k ) t = rk and t u k, T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

15 Solving correction vector t Correction equation: ( I uk u T ) k (A θk I)(I u k u T k )t = r k, t u k. (3) Scheme S OneLS : Use preconditioning iterative approximations, e.g., GMRES, to solve (3). Use a preconditioner M p ( I u k u T k ) M ( I uk u T k ), where M is an approximation of A θ k I. In each of the iterative steps, it needs to solve M p y = b, y u k (4) for a given b. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

16 Since y u k, Eq. (4) can be rewritten as ( I uk u T ) ( k My = b My = u T k My ) u k + b η k u k + b. Hence where y = M 1 b + η k M 1 u k, η k = ut k M 1 b u T k M 1 u k. SSOR preconditioner: Let A θ k I = L + D + U. Then M = (D + ωl)d 1 (D + ωu). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

17 Scheme S T wols : Since t u k, Eq. (3) can be rewritten as (A θ k I)t = ( u T k (A θ ki)t ) u k r k εu k r k. (5) Let t 1 and t 2 be approximated solutions of the following linear systems: (A θ k I)t = r k and (A θ k I)t = u k, respectively. Then the approximated solution t for (5) is t = t 1 + εt 2 for ε = ut k t 1 u T k t. 2 Scheme S OneStep : The approximated solution t for (5) is t = M 1 r k + εm 1 u k for ε = u k M 1 r k u k M 1 u k, where M is an approximation of A θ k I. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

18 Algorithm 2 (Jacobi-Davidson Method) Choose an n-by-m orthonormal matrix V 0 Do k = 0, 1, 2, Compute all the eigenpairs of Vk T AV ks = λs. Select the desired (target) eigenpair (θ k, s k ) with s k 2 = 1. Compute u k = V k s k and r k = (A θ k I)u k. If ( r k 2 < ε), λ = θ k, x = u k, Stop Solve (approximately) a t k u k from (I u k u T k )(A θ ki)(i u k u T k )t = r k. Orthogonalize t k V k v k+1, V k+1 = [V k, v k+1 ] T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

19 Locking: V k with V k V = I k are convergent Schur vectors, i.e., AV k = V k T k for some upper triangular T k. Set V = [V k, V q ] with V V = I k+q in k + 1-th iteration of Jacobi-Davidson Algorithm. Then [ ] [ ] V V AV = = k AV k Vq AV k [ Tk V k AV q 0 V q AV q Vk AV q Vq AV q ]. = T k V q V k T k Vk AV q V q AV q Restarting: Keep the locked Schur vectors as well as the Schur vectors of interest in the subspace and throw away those we are not interested. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

20 Remark 1 If ε = 0, we obtain Davidson method with ( t is NOT orthogonal to u k ) t 1 = (D A θ k I) 1 r. If the linear systems in (6) are exactly solved, then the vector t becomes t = ε(a θ k I) 1 u k u k. (6) Since t is made orthogonal to u k, (6) is equivalent to t = (A θ k I) 1 u k which is equivalent to shift-invert power iteration. Hence it is quadratic convergence. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

21 Consider Ax = λx and assume that λ is simple. Lemma 6 Consider w with w T x 0. Then the map ) ) F p (I xwt w T (A λi) (I xwt x w T x is a bijection from w to w. Proof: Suppose y w and F p y = 0. That is Then it holds that ) ) (I xwt w T (A λi) (I xwt x w T y = 0. x (A λi)y = εx. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

22 Therefore, both y and x belong to the kernel of (A λi) 2. The simplicity of λ implies that y is a scale multiple of x. The fact that y w and x T w 0 implies y = 0, which proves the injectivity of F p. An obvious dimension argument implies bijectivity. Extension F p t Then ) ) (I uut u T (A θi) (I uut u u T t = r, t u, r u. u t u Fp r u. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

23 Theorem 7 Assume that the correction equation ( I uu T ) (A θi) ( I uu T ) t = r, t u (7) is solved exactly in each step of Jacobi-Davidson Algorithm. Assume u k = u x and u T k x has non-trivial limit. Then if u k is sufficiently chosen to x, then u k x with locally quadratical convergence and θ k = u T k Au k λ. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

24 Proof: Suppose Ax = λx with x such that x = u + z for z u. Then (A θi) z = (A θi) u + (λ θ) x = r + (λ θ) x. (8) Consider the exact solution z 1 u of (7): (I P ) (A θi) z 1 = (I P ) r, (9) where P = uu T. Note that (I P ) r = r since u r. Since x (u + z 1 ) = z z 1 and z = x u, for quadratic convergence, it suffices to show that x (u + z 1 ) = z z 1 = O ( z 2). (10) Multiplying (8) by (I P ) and subtracting the result from (9) yields (I P ) (A θi) (z z 1 ) = (λ θ) (I P ) z + (λ θ) (I P ) u. (11) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

25 Multiplying (8) by u T and using r u leads to λ θ = ut (A θi) z u T. (12) x Since u T k x has non-trivial limit, we obtain (λ θ) (I P ) z = u T (A θi) z u T x (I P ) z. (13) From (11), Lemma 6 and (I P ) u = 0, we have [ ] 1 z z 1 = (I P ) (A θ k I) u (λ θ) (I P ) z k [ ] 1 = u T k (I P ) (A θ k I) (A θ ki) z u k u T k x (I P ) z = O ( z 2). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

26 Jacobi Davidson method as on accelerated Newton Scheme Consider Ax = λx and assume that λ is simple. Choose w T x = 1. Consider nonlinear equation F (u) = Au θ(u)u = 0 with θ (u) = wt Au w T u, where u = 1 or w T u = 1. Then F : {u w T u = 1} w. In particular, r F (u) = Au θ (u) u w. Suppose u k x and the next Newton approximation u k+1 : ( 1 F u k+1 = u k u F (u k ) u=uk) is given by u k+1 = u k + t, i.e., t satisfies that ( ) F t = F (u k ) = r. u u=uk T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

27 Since 1 = u T k+1 w = (u k + t) T w = 1 + t T w, it implies that w T t = 0. By the definition of F, we have F u = A θ (u) I ( w T Au ) uw T + ( w T u ) uw T A (w T u) 2 = A θi + wt Au (w T u) 2 uwt uwt A (I w T u = u kw T ) w T (A θ k I). u k Consequently, the Jacobian of F acts on w and is given by ( ) F t = (I u kw T ) u w T (A θ k I) t, t w. u k u=uk Hence the correction equation of Newton method read as t w, (I u kw T ) w T (A θ k I) t = r, u k which is the correction equation of Jacobi-Davidson method in (7) with w = u. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

28 Polynomial Jacobi-Davidson method (θ k, u k ): approx. eigenpair of A(λ) τ k=0 λk A k, θ k λ, with Let Then u k = V k s k, V k A(λ)V ks k = 0 and s k 2 = 1. r k = A(θ k )u k. u k r k = u k A(θ k)u k = s k V k A(θ k)v k s k = 0 r k u k Find the correction t such that That is A(λ)(u k + t) = 0. A(λ)t = A(λ)u k = r k + (A(θ k ) A(λ))u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

29 Let p k = A (θ k )u k ( τ i=1 iθ i 1 k A i ) u k. A(λ) = A λi: A(λ) = A λb: p k = u k, (A(θ k ) A(λ))u k = (λ θ k )u k = (θ k λ k )p k p k = Bu k, (A(θ k ) A(λ))u k = (λ θ k )Bu k = (θ k λ)p k A(λ) = τ i=0 λi A i with τ 2: [ (A(θ k ) A(λ))u k = (θ k λ)a (θ k ) 1 ] 2 (θ k λ) 2 A (ξ k ) u k = (θ k λ)p k 1 2 (θ k λ) 2 A (ξ k )u k T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

30 Hence A(λ)t = r k + (θ k λ)p k 1 2 (θ k λ) 2 A (ξ k )u k Since r k u k, we have ( I p ku ) k u k p A(λ)t = r k 12 ( (θ k λ) 2 I p ku ) k k u k p A (ξ k )u k. k Correction equation: ( I p ku ) k u k p A(θ k )(I u k u k )t = r k and t u k, k or ( I p ku ) ( k u k p (A θ k B) I u kp ) k k p k u t = r k and t B u k, k with symmetric positive definite matrix B. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

31 Algorithm 3 (Jacobi-Davidson Algorithm for solving A(λ)x = 0) Choose an n-by-m orthonormal matrix V 0 Do k = 0, 1, 2, Compute all the eigenpairs of Vk A(λ)V k = 0. Select the desired (target) eigenpair (θ k, s k ) with s k 2 = 1. Compute u k = V k s k, r k = A(θ k )u k and p k = A (θ k )u k. If ( r k 2 < ε), λ = θ k, x = u k, Stop Solve (approximately) a t k u k from (I p ku T k u T k p )A(θ k )(I u k u T k k )t = r k. Orthogonalize t k V k v k+1, V k+1 = [V k, v k+1 ] p k = A (θ k )u k ( τ i=1 iθ i 1 k A i ) u k. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

32 Non-equivalence deflation of quadratic eigenproblems Let λ 1 be a real eigenvalue of Q(λ) λ 2 M + λc + K and x 1, z 1 be the associated right and left eigenvectors, respectively, with z T 1 Kx 1 = 1. Let θ 1 = (z T 1 Mx 1 ) 1. We introduce a deflated quadratic eigenproblem [ ] Q(λ)x λ 2 M + λ C + K x = 0, where M = M θ 1 Mx 1 z T 1 M, C = C + θ 1 λ 1 (Mx 1 z T 1 K + Kx 1 z T 1 M), K = K θ 1 λ 2 Kx 1 z1 T K. 1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

33 Complex deflation Let λ 1 = α 1 + iβ 1 be a complex eigenvalue of Q(λ) and x 1 = x 1R + ix 1I, z 1 = z 1R + iz 1I be the associated right and left eigenvectors, respectively, such that Z T 1 KX 1 = I 2, where X 1 = [x 1R, x 1I ] and Z 1 = [z 1R, z 1I ]. Let Θ 1 = (Z T 1 MX 1 ) 1. Then we introduce a deflated quadratic eigenproblem with M = M MX 1 Θ 1 Z1 T M, C = C + MX 1 Θ 1 Λ T 1 Z1 T K + KX 1 Λ 1 1 ΘT 1 Z1 T M, K = K KX 1 Λ 1 1 Θ 1Λ T 1 Z1 T K [ ] α1 β in which Λ 1 = 1. β 1 α 1 T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

34 Theorem 8 (i) Let λ 1 be a simple real eigenvalue of Q(λ). Then the spectrum of Q(λ) is given by provided that λ 2 1 θ 1. (σ(q(λ)) {λ 1 }) { } (ii) Let λ 1 be a simple complex eigenvalue of Q(λ). Then the spectrum of Q(λ) is given by ( σ(q(λ)) {λ1, λ 1 } ) {, } provided that Λ 1 Λ T 1 Θ 1. Furthermore, in both cases (i) and (ii), if λ 2 λ 1 and (λ 2, x 2 ) is an eigenpair of Q(λ) then the pair (λ 2, x 2 ) is also an eigenpair of Q(λ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

35 Suppose that M, C, K are symmetric. Given an eigenmatrix pair (Λ 1, X 1 ) R r r R n r of Q(λ), where Λ 1 is nonsingular and X 1 satisfies X T 1 KX 1 = I r, Θ 1 := (X T 1 MX 1 ) 1. We define Q(λ) := λ 2 M + λ C + K, where M := M MX 1 Θ 1 X1 T M, C := C + MX 1 Θ 1 Λ T 1 X1 T K + KX 1 Λ 1 1 Θ 1X1 T M, K := K KX 1 Λ 1 1 Θ 1Λ T 1 X1 T K. Theorem 9 Suppose that Θ 1 Λ 1 Λ T 1 is nonsingular. Then the eigenvalues of the real symmetric quadratic pencil Q(λ) are the same as those of Q(λ) except that the eigenvalues of Λ 1, which are closed under complex conjugation, are replaced by r infinities. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

36 Proof: Since (Λ 1, X 1 ) is an eigenmatrix pair of Q(λ), i.e., we have MX 1 Λ CX 1 Λ 1 + KX 1 = 0, Q(λ) = Q(λ) + [MX 1 (λi r + Λ 1 ) + CX 1 ] Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M) = Q(λ) + Q(λ)X 1 (λi r Λ 1 ) 1 Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M). By using the identity where R, S T R n m, we have det[ Q(λ)] det(i n + RS) = det(i m + SR), = det[q(λ)]det[i + X 1 (λi r Λ 1 ) 1 Θ 1 Λ T 1 (X T 1 K λλ T 1 X T 1 M)] = det[q(λ)]det[i r + (λi r Λ 1 ) 1 Θ 1 Λ T 1 (I r λλ T 1 Θ 1 1 )] det[q(λ)] = det(λi r Λ 1 ) det(θ 1Λ T 1 Λ 1 ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

37 Since (Θ 1 Λ 1 Λ T 1 ) Rr r is nonsingular, we have det(θ 1 Λ T 1 Λ 1 ) 0. Therefore, Q(λ) has the same eigenvalues as Q(λ) except that r eigenvalues of Λ 1 are replaced by r infinities. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

38 Non-equiv. deflation for cubic poly. eigenproblems Let (Λ, V u ) R r r R n r be an eigenmatrix pair of A(λ) λ 3 A 3 + λ 2 A 2 + λa 1 + A 0 (14) with V T u V u = I r and 0 / σ(λ), i.e., A 3 V u Λ 3 + A 2 V u Λ 2 + A 1 V u Λ + A 0 V u = 0. (15) Define a new deflated cubic eigenvalue problem by Ã(λ)u = (λ 3 à 3 + λ 2 à 2 + λã1 + Ã0)u = 0, (16) where à 0 = A 0, à 1 = A 1 (A 1 V u V T u + A 2 V u ΛV T u + A 3 V u Λ 2 V T u ), à 2 = A 2 (A 2 V u V T u + A 3 V u ΛV T u ), à 3 = A 3 A 3 V u V T u. (17) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

39 Lemma 10 Let A(λ) and Ã(λ) be cubic pencils given by (14) and (16), respectively. Then it holds Ã(λ) = A(λ) ( I n λv u (λi r Λ) 1 V T u ). (18) Theorem 11 Let (Λ, V u ) be an eigenmatrix pair of A(λ) with V T u V u = I r. Then (i) (σ(a(λ)) σ(λ)) { } = σ(ã(λ)). (ii) Let (µ, z) be an eigenpair of A(λ) with z 2 = 1 and µ / σ(λ). Define z = (I n µv u Λ 1 V T u )z T (µ)z. (19) Then (µ, z) is an eigenpair of Ã(λ). T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

40 Proof of Lemma: Using (17) and (15), and the fundamental matrix calculation, we have Ã(λ) = A(λ) λ ( λ 2 A 3 V F VF T + λa 2 V F VF T + λa 3 V F ΛVF T +A 1 V F VF T + A 2 V F ΛVF T + A 3 V F Λ 2 VF T ) = A(λ) λ ( A 3 V F (λi r Λ) 3 (λi r Λ) 1 VF T +3A 3 V F Λ(λI r Λ) 2 (λi r Λ) 1 VF T +3A 3 V F Λ 2 (λi r Λ)(λI r Λ) 1 VF T +A 2 V F (λi r Λ) 2 (λi r Λ) 1 VF T +2A 2 V F Λ(λI r Λ)(λI r Λ) 1 VF T +A 1 V F (λi r Λ)(λI r Λ) 1 VF T ) T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

41 Ã(λ) = A(λ) λ {[ A 3 V F (λ 3 I r Λ 3 ) + A 2 V F (λ 2 I r Λ 2 ) +A 1 V F (λi r Λ) + A 0 V F A 0 V F ] (λi r Λ) 1 VF T } = A(λ) λ [ A(λ)V F (λi r Λ) 1 VF T ] = A(λ) [ I n λv F (λi r Λ) 1 VF T ]. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

42 Proof of Theorem: (i) Using the identity det(i n + RS) = det(i m + SR) and Lemma 10, we have det(ã(λ)) = det(a(λ)) det ( I n λv F (λi r Λ) 1 VF T ) = det(a(λ)) det ( I n λ(λi r Λ) 1) = det(a(λ)) det(λi r Λ) 1 det( Λ). Since 0 / σ(λ), det( Λ) 0. Thus, Ã(λ) and A(λ) have the same finite spectrum except the eigenvalues in σ(λ). Furthermore, dividing Eq. (16) by λ 3 and using the fact that à 3 V F = (A 3 A 3 V F V T F )V F = 0, we see that (diag r {,, }, V F ) is an eigenmatrix pair of Ã(λ) corresponding to infinite eigenvalues. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

43 (ii) Since µ / σ(λ), the matrix T (µ) = (I µv F Λ 1 VF T ) in (19) is invertible with the inverse T (µ) 1 = I n µv F (µi r Λ) 1 VF T. (20) From Lemma 10, we have Ã(µ) z = A(µ) [ I n µv F (µi r Λ) 1 VF T ] [ In µv F Λ 1 VF T ] z = 0. This completes the proof. T.M. Huang (Taiwan Normal Univ.) Poly. JD method for PEP April 28, / 42

Jacobi Davidson Methods for Cubic Eigenvalue Problems (March 11, 2004)

Jacobi Davidson Methods for Cubic Eigenvalue Problems (March 11, 2004) Jacobi Davidson Methods for Cubic Eigenvalue Problems (March 11, 24) Tsung-Min Hwang 1, Wen-Wei Lin 2, Jinn-Liang Liu 3, Weichung Wang 4 1 Department of Mathematics, National Taiwan Normal University,

More information

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation

More information

A Jacobi Davidson Method for Nonlinear Eigenproblems

A Jacobi Davidson Method for Nonlinear Eigenproblems A Jacobi Davidson Method for Nonlinear Eigenproblems Heinrich Voss Section of Mathematics, Hamburg University of Technology, D 21071 Hamburg voss @ tu-harburg.de http://www.tu-harburg.de/mat/hp/voss Abstract.

More information

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A

More information

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology The Davidson method is a popular technique to compute a few of the smallest (or largest)

More information

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems Krylov Subspace Methods for Large/Sparse Eigenvalue Problems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan April 17, 2012 T.-M. Huang (Taiwan Normal University) Krylov

More information

Reduction of nonlinear eigenproblems with JD

Reduction of nonlinear eigenproblems with JD Reduction of nonlinear eigenproblems with JD Henk van der Vorst H.A.vanderVorst@math.uu.nl Mathematical Institute Utrecht University July 13, 2005, SIAM Annual New Orleans p.1/16 Outline Polynomial Eigenproblems

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors week -2 Fall 26 Eigenvalues and eigenvectors The most simple linear transformation from R n to R n may be the transformation of the form: T (x,,, x n ) (λ x, λ 2,, λ n x n

More information

LINEAR ALGEBRA REVIEW

LINEAR ALGEBRA REVIEW LINEAR ALGEBRA REVIEW JC Stuff you should know for the exam. 1. Basics on vector spaces (1) F n is the set of all n-tuples (a 1,... a n ) with a i F. It forms a VS with the operations of + and scalar multiplication

More information

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers.. EIGENVALUE PROBLEMS Background on eigenvalues/ eigenvectors / decompositions Perturbation analysis, condition numbers.. Power method The QR algorithm Practical QR algorithms: use of Hessenberg form and

More information

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique

More information

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter

More information

Eigenvalues and eigenvectors

Eigenvalues and eigenvectors Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors

More information

of dimension n 1 n 2, one defines the matrix determinants

of dimension n 1 n 2, one defines the matrix determinants HARMONIC RAYLEIGH RITZ FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter eigenvalue

More information

OVERVIEW. Basics Generalized Schur Form Shifts / Generalized Shifts Partial Generalized Schur Form Jacobi-Davidson QZ Algorithm

OVERVIEW. Basics Generalized Schur Form Shifts / Generalized Shifts Partial Generalized Schur Form Jacobi-Davidson QZ Algorithm OVERVIEW Basics Generalized Schur Form Shifts / Generalized Shifts Partial Generalized Schur Form Jacobi-Davidson QZ Algorithm BASICS Generalized Eigenvalue Problem Ax lbx We call (, x) Left eigenpair

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

Krylov subspace projection methods

Krylov subspace projection methods I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks

More information

HOMOGENEOUS JACOBI DAVIDSON. 1. Introduction. We study a homogeneous Jacobi Davidson variant for the polynomial eigenproblem

HOMOGENEOUS JACOBI DAVIDSON. 1. Introduction. We study a homogeneous Jacobi Davidson variant for the polynomial eigenproblem HOMOGENEOUS JACOBI DAVIDSON MICHIEL E. HOCHSTENBACH AND YVAN NOTAY Abstract. We study a homogeneous variant of the Jacobi Davidson method for the generalized and polynomial eigenvalue problem. Special

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

Lec 2: Mathematical Economics

Lec 2: Mathematical Economics Lec 2: Mathematical Economics to Spectral Theory Sugata Bag Delhi School of Economics 24th August 2012 [SB] (Delhi School of Economics) Introductory Math Econ 24th August 2012 1 / 17 Definition: Eigen

More information

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14

More information

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:

More information

Alternative correction equations in the Jacobi-Davidson method

Alternative correction equations in the Jacobi-Davidson method Chapter 2 Alternative correction equations in the Jacobi-Davidson method Menno Genseberger and Gerard Sleijpen Abstract The correction equation in the Jacobi-Davidson method is effective in a subspace

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors /88 Chia-Ping Chen Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Eigenvalue Problem /88 Eigenvalue Equation By definition, the eigenvalue equation for matrix

More information

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true. 1 Which of the following statements is always true? I The null space of an m n matrix is a subspace of R m II If the set B = {v 1,, v n } spans a vector space V and dimv = n, then B is a basis for V III

More information

A Jacobi-Davidson method for two real parameter nonlinear eigenvalue problems arising from delay differential equations

A Jacobi-Davidson method for two real parameter nonlinear eigenvalue problems arising from delay differential equations A Jacobi-Davidson method for two real parameter nonlinear eigenvalue problems arising from delay differential equations Heinrich Voss voss@tuhh.de Joint work with Karl Meerbergen (KU Leuven) and Christian

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Multiparameter eigenvalue problem as a structured eigenproblem

Multiparameter eigenvalue problem as a structured eigenproblem Multiparameter eigenvalue problem as a structured eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana This is joint work with M Hochstenbach Będlewo, 2932007 1/28 Overview Introduction

More information

Scaling, Sensitivity and Stability in Numerical Solution of the Quadratic Eigenvalue Problem

Scaling, Sensitivity and Stability in Numerical Solution of the Quadratic Eigenvalue Problem Scaling, Sensitivity and Stability in Numerical Solution of the Quadratic Eigenvalue Problem Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/

More information

Nonlinear Eigenvalue Problems

Nonlinear Eigenvalue Problems 115 Nonlinear Eigenvalue Problems Heinrich Voss Hamburg University of Technology 115.1 Basic Properties........................................ 115-2 115.2 Analytic matrix functions.............................

More information

Iterative techniques in matrix algebra

Iterative techniques in matrix algebra Iterative techniques in matrix algebra Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan September 12, 2015 Outline 1 Norms of vectors and matrices 2 Eigenvalues and

More information

MICHIEL E. HOCHSTENBACH

MICHIEL E. HOCHSTENBACH VARIATIONS ON HARMONIC RAYLEIGH RITZ FOR STANDARD AND GENERALIZED EIGENPROBLEMS MICHIEL E. HOCHSTENBACH Abstract. We present several variations on the harmonic Rayleigh Ritz method. First, we introduce

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 16: Eigenvalue Problems; Similarity Transformations Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 18 Eigenvalue

More information

A Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation

A Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation A Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation Zih-Hao Wei 1, Feng-Nan Hwang 1, Tsung-Ming Huang 2, and Weichung Wang 3 1 Department of

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

JUST THE MATHS UNIT NUMBER 9.8. MATRICES 8 (Characteristic properties) & (Similarity transformations) A.J.Hobson

JUST THE MATHS UNIT NUMBER 9.8. MATRICES 8 (Characteristic properties) & (Similarity transformations) A.J.Hobson JUST THE MATHS UNIT NUMBER 9.8 MATRICES 8 (Characteristic properties) & (Similarity transformations) by A.J.Hobson 9.8. Properties of eigenvalues and eigenvectors 9.8. Similar matrices 9.8.3 Exercises

More information

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace Purdue University Purdue e-pubs Department of Computer Science Technical Reports Department of Computer Science 2009 On the Modification of an Eigenvalue Problem that Preserves an Eigenspace Maxim Maumov

More information

Inexact inverse iteration with preconditioning

Inexact inverse iteration with preconditioning Department of Mathematical Sciences Computational Methods with Applications Harrachov, Czech Republic 24th August 2007 (joint work with M. Robbé and M. Sadkane (Brest)) 1 Introduction 2 Preconditioned

More information

Iterative projection methods for sparse nonlinear eigenvalue problems

Iterative projection methods for sparse nonlinear eigenvalue problems Iterative projection methods for sparse nonlinear eigenvalue problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Iterative projection

More information

Lecture 3: Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method

More information

Efficient Methods For Nonlinear Eigenvalue Problems. Diploma Thesis

Efficient Methods For Nonlinear Eigenvalue Problems. Diploma Thesis Efficient Methods For Nonlinear Eigenvalue Problems Diploma Thesis Timo Betcke Technical University of Hamburg-Harburg Department of Mathematics (Prof. Dr. H. Voß) August 2002 Abstract During the last

More information

Eigenvalue and Eigenvector Problems

Eigenvalue and Eigenvector Problems Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems

More information

FAME-matlab Package: Fast Algorithm for Maxwell Equations Tsung-Ming Huang

FAME-matlab Package: Fast Algorithm for Maxwell Equations Tsung-Ming Huang FAME-matlab Package: Fast Algorithm for Maxwell Equations Tsung-Ming Huang Modelling, Simulation and Analysis of Nonlinear Optics, NUK, September, 4-8, 2015 1 2 FAME group Wen-Wei Lin Department of Applied

More information

Jacobi-Davidson Algorithm for Locating Resonances in a Few-Body Tunneling System

Jacobi-Davidson Algorithm for Locating Resonances in a Few-Body Tunneling System Jacobi-Davidson Algorithm for Locating Resonances in a Few-Body Tunneling System Bachelor Thesis in Physics and Engineering Physics Gustav Hjelmare Jonathan Larsson David Lidberg Sebastian Östnell Department

More information

Research Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics

Research Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics Research Matters February 25, 2009 The Nonlinear Eigenvalue Problem Nick Higham Part III Director of Research School of Mathematics Françoise Tisseur School of Mathematics The University of Manchester

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Matrix Theory. A.Holst, V.Ufnarovski

Matrix Theory. A.Holst, V.Ufnarovski Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

Nonlinear Eigenvalue Problems: An Introduction

Nonlinear Eigenvalue Problems: An Introduction Nonlinear Eigenvalue Problems: An Introduction Cedric Effenberger Seminar for Applied Mathematics ETH Zurich Pro*Doc Workshop Disentis, August 18 21, 2010 Cedric Effenberger (SAM, ETHZ) NLEVPs: An Introduction

More information

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers MAX PLANCK INSTITUTE International Conference on Communications, Computing and Control Applications March 3-5, 2011, Hammamet, Tunisia. Model order reduction of large-scale dynamical systems with Jacobi-Davidson

More information

A Jacobi Davidson-type projection method for nonlinear eigenvalue problems

A Jacobi Davidson-type projection method for nonlinear eigenvalue problems A Jacobi Davidson-type projection method for nonlinear eigenvalue problems Timo Betce and Heinrich Voss Technical University of Hamburg-Harburg, Department of Mathematics, Schwarzenbergstrasse 95, D-21073

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Linear Algebra Lecture Notes-II

Linear Algebra Lecture Notes-II Linear Algebra Lecture Notes-II Vikas Bist Department of Mathematics Panjab University, Chandigarh-64 email: bistvikas@gmail.com Last revised on March 5, 8 This text is based on the lectures delivered

More information

Goal: to construct some general-purpose algorithms for solving systems of linear Equations

Goal: to construct some general-purpose algorithms for solving systems of linear Equations Chapter IV Solving Systems of Linear Equations Goal: to construct some general-purpose algorithms for solving systems of linear Equations 4.6 Solution of Equations by Iterative Methods 4.6 Solution of

More information

The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix

The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix Chun-Yueh Chiang Center for General Education, National Formosa University, Huwei 632, Taiwan. Matthew M. Lin 2, Department of

More information

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm Name: The proctor will let you read the following conditions

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Chapter 1 Eigenvalues and Eigenvectors Among problems in numerical linear algebra, the determination of the eigenvalues and eigenvectors of matrices is second in importance only to the solution of linear

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

Math 215 HW #9 Solutions

Math 215 HW #9 Solutions Math 5 HW #9 Solutions. Problem 4.4.. If A is a 5 by 5 matrix with all a ij, then det A. Volumes or the big formula or pivots should give some upper bound on the determinant. Answer: Let v i be the ith

More information

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems. Chapter 3 Linear Algebra In this Chapter we provide a review of some basic concepts from Linear Algebra which will be required in order to compute solutions of LTI systems in state space form, discuss

More information

Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form. Jay Daigle Advised by Stephan Garcia

Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form. Jay Daigle Advised by Stephan Garcia Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form Jay Daigle Advised by Stephan Garcia April 4, 2008 2 Contents Introduction 5 2 Technical Background 9 3

More information

LU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU

LU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU LU Factorization A m n matri A admits an LU factorization if it can be written in the form of Where, A = LU L : is a m m lower triangular matri with s on the diagonal. The matri L is invertible and is

More information

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Math 489AB Exercises for Chapter 2 Fall Section 2.3 Math 489AB Exercises for Chapter 2 Fall 2008 Section 2.3 2.3.3. Let A M n (R). Then the eigenvalues of A are the roots of the characteristic polynomial p A (t). Since A is real, p A (t) is a polynomial

More information

Math 310 Final Exam Solutions

Math 310 Final Exam Solutions Math 3 Final Exam Solutions. ( pts) Consider the system of equations Ax = b where: A, b (a) Compute deta. Is A singular or nonsingular? (b) Compute A, if possible. (c) Write the row reduced echelon form

More information

Problems of Eigenvalues/Eigenvectors

Problems of Eigenvalues/Eigenvectors 5 Problems of Eigenvalues/Eigenvectors Reveiw of Eigenvalues and Eigenvectors Gerschgorin s Disk Theorem Power and Inverse Power Methods Jacobi Transform for Symmetric Matrices Singular Value Decomposition

More information

Computable error bounds for nonlinear eigenvalue problems allowing for a minmax characterization

Computable error bounds for nonlinear eigenvalue problems allowing for a minmax characterization Computable error bounds for nonlinear eigenvalue problems allowing for a minmax characterization Heinrich Voss voss@tuhh.de Joint work with Kemal Yildiztekin Hamburg University of Technology Institute

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Linear System Theory

Linear System Theory Linear System Theory Wonhee Kim Lecture 4 Apr. 4, 2018 1 / 40 Recap Vector space, linear space, linear vector space Subspace Linearly independence and dependence Dimension, Basis, Change of Basis 2 / 40

More information

Efficient computation of transfer function dominant poles of large second-order dynamical systems

Efficient computation of transfer function dominant poles of large second-order dynamical systems Chapter 6 Efficient computation of transfer function dominant poles of large second-order dynamical systems Abstract This chapter presents a new algorithm for the computation of dominant poles of transfer

More information

EE5120 Linear Algebra: Tutorial 7, July-Dec Covers sec 5.3 (only powers of a matrix part), 5.5,5.6 of GS

EE5120 Linear Algebra: Tutorial 7, July-Dec Covers sec 5.3 (only powers of a matrix part), 5.5,5.6 of GS EE5 Linear Algebra: Tutorial 7, July-Dec 7-8 Covers sec 5. (only powers of a matrix part), 5.5,5. of GS. Prove that the eigenvectors corresponding to different eigenvalues are orthonormal for unitary matrices.

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Tsung-Ming Huang. Matrix Computation, 2016, NTNU

Tsung-Ming Huang. Matrix Computation, 2016, NTNU Tsung-Ming Huang Matrix Computation, 2016, NTNU 1 Plan Gradient method Conjugate gradient method Preconditioner 2 Gradient method 3 Theorem Ax = b, A : s.p.d Definition A : symmetric positive definite

More information

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems

More information

Problems of Eigenvalues/Eigenvectors

Problems of Eigenvalues/Eigenvectors 67 Problems of Eigenvalues/Eigenvectors Reveiw of Eigenvalues and Eigenvectors Gerschgorin s Disk Theorem Power and Inverse Power Methods Jacobi Transform for Symmetric Matrices Spectrum Decomposition

More information

MATH Spring 2011 Sample problems for Test 2: Solutions

MATH Spring 2011 Sample problems for Test 2: Solutions MATH 304 505 Spring 011 Sample problems for Test : Solutions Any problem may be altered or replaced by a different one! Problem 1 (15 pts) Let M, (R) denote the vector space of matrices with real entries

More information

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018 1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)

More information

Proofs for Quizzes. Proof. Suppose T is a linear transformation, and let A be a matrix such that T (x) = Ax for all x R m. Then

Proofs for Quizzes. Proof. Suppose T is a linear transformation, and let A be a matrix such that T (x) = Ax for all x R m. Then Proofs for Quizzes 1 Linear Equations 2 Linear Transformations Theorem 1 (2.1.3, linearity criterion). A function T : R m R n is a linear transformation if and only if a) T (v + w) = T (v) + T (w), for

More information

Part IA. Vectors and Matrices. Year

Part IA. Vectors and Matrices. Year Part IA Vectors and Matrices Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2018 Paper 1, Section I 1C Vectors and Matrices For z, w C define the principal value of z w. State de Moivre s

More information

Review Notes for Linear Algebra True or False Last Updated: January 25, 2010

Review Notes for Linear Algebra True or False Last Updated: January 25, 2010 Review Notes for Linear Algebra True or False Last Updated: January 25, 2010 Chapter 3 [ Eigenvalues and Eigenvectors ] 31 If A is an n n matrix, then A can have at most n eigenvalues The characteristic

More information

ACM 104. Homework Set 5 Solutions. February 21, 2001

ACM 104. Homework Set 5 Solutions. February 21, 2001 ACM 04 Homework Set 5 Solutions February, 00 Franklin Chapter 4, Problem 4, page 0 Let A be an n n non-hermitian matrix Suppose that A has distinct eigenvalues λ,, λ n Show that A has the eigenvalues λ,,

More information

LARGE SPARSE EIGENVALUE PROBLEMS

LARGE SPARSE EIGENVALUE PROBLEMS LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems

More information

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex

More information

Inverse Eigenvalue Problem with Non-simple Eigenvalues for Damped Vibration Systems

Inverse Eigenvalue Problem with Non-simple Eigenvalues for Damped Vibration Systems Journal of Informatics Mathematical Sciences Volume 1 (2009), Numbers 2 & 3, pp. 91 97 RGN Publications (Invited paper) Inverse Eigenvalue Problem with Non-simple Eigenvalues for Damped Vibration Systems

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to: MAC Module Eigenvalues and Eigenvectors Learning Objectives Upon completing this module, you should be able to: Solve the eigenvalue problem by finding the eigenvalues and the corresponding eigenvectors

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

ABSTRACT. Professor G.W. Stewart

ABSTRACT. Professor G.W. Stewart ABSTRACT Title of dissertation: Residual Arnoldi Methods : Theory, Package, and Experiments Che-Rung Lee, Doctor of Philosophy, 2007 Dissertation directed by: Professor G.W. Stewart Department of Computer

More information

A Note on Inverse Iteration

A Note on Inverse Iteration A Note on Inverse Iteration Klaus Neymeyr Universität Rostock, Fachbereich Mathematik, Universitätsplatz 1, 18051 Rostock, Germany; SUMMARY Inverse iteration, if applied to a symmetric positive definite

More information

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii

More information