Convergence Theory for Iterative Eigensolvers

Size: px
Start display at page:

Download "Convergence Theory for Iterative Eigensolvers"

Transcription

1 Convergence Theory for Iterative Eigensolvers Mark Embree Virginia Tech with Chris Beattie, Russell Carden, John Rossi, Dan Sorensen RandNLA Workshop Simons Institute September 28

2 setting for the talk Let A C n n be a large square matrix, potentially non-hermitian (A A ). Computing all eigenvalues of A is too expensive (and usually not needed). Thus we seek m n distinguished eigenvalues relevant to our application (largest, smallest, rightmost, etc.)

3 setting for the talk Let A C n n be a large square matrix, potentially non-hermitian (A A ). Computing all eigenvalues of A is too expensive (and usually not needed). Thus we seek m n distinguished eigenvalues relevant to our application (largest, smallest, rightmost, etc.) Projection Methods V C n = k-dimensional subspace of C n, the projection subspace for A The columns of V C n k for an orthonormal basis for V: V V = I, V AV C k k We hope some eigenvalues of V AV σ(v AV) = {θ,..., θ k } approximate some eigenvalues of A. σ(a) = {λ,..., λ n} For example, θ λ,... θ m λ m for some m k.

4 setting for the talk Let A C n n be a large square matrix, potentially non-hermitian (A A ). Computing all eigenvalues of A is too expensive (and usually not needed). Thus we seek m n distinguished eigenvalues relevant to our application (largest, smallest, rightmost, etc.) Projection Methods V C n = k-dimensional subspace of C n, the projection subspace for A The columns of V C n k for an orthonormal basis for V: V V = I, V AV C k k We hope some eigenvalues of V AV σ(v AV) = {θ,..., θ k } approximate some eigenvalues of A. σ(a) = {λ,..., λ n} For example, θ λ,... θ m λ m for some m k. This talk mainly describes established results for the deterministic case, with some thoughts from a RandNLA perspective along the way.

5 why consider A A? While Hermitian problems are common (SVD, quantum mechanics, etc.), many important applications lead to non-hermitian problems and subtler issues of spectral perturbation theory. Many examples: [Trefethen, E. 25]. atmospheric science Farrell... fluid flow stability Trefethen; Schmid & Henningson;... damped mechanical systems Cox & Zuazua,... control theory Hinrichsen & Pritchard,... data-driven modeling Antoulas, Beattie, Gugercin,... lasers Landau, Siegman,... ecology May, Caswell,... Markov chains Diaconis, computed rightmost eigenvalues for stability of 2d flow over a backward facing step, n = 38,539 [E., Keeler 27]

6 An Overview of Projection-Based Eigensolvers

7 projection-based eigensolvers V C n = k-dimensional subspace of C n, the projection subspace for A. The columns of V C n k form an orthonormal basis for V: V V = I, V AV C k k We hope some eigenvalues of V AV σ(v AV) = {θ,..., θ k } approximate some eigenvalues of A. σ(a) = {λ,..., λ n} For example, θ λ,... θ m λ m for some m k.

8 projection-based eigensolvers V C n = k-dimensional subspace of C n, the projection subspace for A. The columns of V C n k form an orthonormal basis for V: V V = I, V AV C k k We hope some eigenvalues of V AV σ(v AV) = {θ,..., θ k } approximate some eigenvalues of A. σ(a) = {λ,..., λ n} For example, θ λ,... θ m λ m for some m k. Power method V = span{a p x} (minimal storage, easy to implement, can be slow) Subspace iteration (more storage, subtler to implement, computes repeated eigs) V = Range(A p X) for X C n k [Halko, Martinsson, Tropp 2] et al.

9 projection-based eigensolvers: krylov methods Power method V = span{a p x} (minimal storage, easy to implement, can be slow) Subspace iteration (more storage, subtler to implement, computes multiple eigs) V = Range(A p X) for X C n b Krylov subspace methods (growing subspace dimension; higher powers of A) V = span{x, Ax, A 2 x,..., A k x} Block Krylov methods (subspace dimension grows quickly: dim(v) kb) V = Range([X AX A 2 X A k X]) for X C n b SVD: [Musco & Musco 25], [Drineas et al. 28] Must balance benefit of large k with block size b, storage.

10 projection-based eigensolvers: krylov methods (extensions) Krylov subspace methods (growing subspace dimension; higher powers of A) V = span{x, Ax, A 2 x,..., A k x} Restarted Krylov (used in eigs: filter ψ improves starting vector) V = span{ψ(a)x, Aψ(A)x, A 2 ψ(a)x,..., A k ψ(a)x} Polynomial Preconditioned Krylov V = span{x, π(a)x, π(a) 2 x,..., π(a) k x} (very high degree polys, care needed) Shift-Invert Krylov (used in eigs: ideal for eigenvalues near µ) V = span{x, (A µi) )x, (A µi) 2 x,..., (A µi) (k ) x} Rational Krylov (helps for finding eigenvalues in a region) V = span{x, (A µ ) x, (A µ 2) x,..., (A µ k I) x}

11 preliminaries: spectral structure of A Distinct eigenvalues of A: λ, λ 2,..., λ n Spectral projectors P j and invariant subspaces U j : P j := (zi A) dz, U j := Range(P j ), 2πi Γ j Γ j is a contour in C containing λ j but no other distinct eigenvalues. If A = A and λ j is simple with unit eigenvector u j, then P j = u j u j. P j is a projector onto the invariant subspace U j, but P j need not be an orthogonal projector when A A. The spectral projectors give a resolution of the identity: n j= P j = I.

12 preliminaries: spectral structure of A Distinct eigenvalues of A: λ, λ 2,..., λ n Spectral projectors P j and invariant subspaces U j : P j := (zi A) dz, U j := Range(P j ), 2πi Γ j Γ j is a contour in C containing λ j but no other distinct eigenvalues. If A = A and λ j is simple with unit eigenvector u j, then P j = u j u j. P j is a projector onto the invariant subspace U j, but P j need not be an orthogonal projector when A A. The spectral projectors give a resolution of the identity: P g := P + + P m, U g := Range(P g), m = dim(u g). n j= P j = I. P b := I P g, U b := Range(P b ), dim(u b ) = n m.

13 preliminaries: angles between subspaces V = approximating subspace. For our problems, V = K k (A, x) := span{x, Ax, A 2 x,..., A k x}. U g = desired invariant subspace Measure convergence via the containment gap: δ(u g, V) = max u U g sin (u, V) = max u U g u v min v V u. u min u v v V (u, V) V We will monitor how δ(u g, K k (A, x)) develops as k increases.

14 example convergence behavior, A A δ(ug, Kk(A, x)) k < m warm-up -5 - sublinear convergence (starting vector bias) (nonorthogonal eigenvectors) steady convergence (eigenvalue location) subspace dimension, k cf. GMRES convergence model of [Nevanlinna 993] example adapted from [Beattie, E., Rossi 24]

15 basic convergence model Building on [Saad 98, 983], [Jia 995], [Sorensen 22], [Beattie, E., Rossi 24], and others... Theorem [Beattie, E., Sorensen 25]. Suppose U g is reachable from the Krylov space K k (A, x). Then for k 2m, δ(u g, K k (A, x)) C C 2 min φ P k 2m max z Ω b α(z)φ(z). C = C (A, x) = measure of starting vector bias. C 2 = C 2(A, Ω b ) = measure of eigenvector departure from orthogonality. P k 2m = set of polynomials of degree k 2m or less. Ω b C contains the undesired eigenvalues. α(z) = (z λ ) (z λ m).

16 Invariant Subspaces reachable by Krylov Subspaces

17 reachable invariant subspaces If x C n lacks a component in the desired eigenvector, e.g., P x =, the desired eigenvalue/eigenvector is invisible to Krylov methods (in exact arithmetic). For example, in the power method A p x = n n λ p j P j x = + λ p j P j x, j= j=2 the eigenvalue λ has no influence. (We will address this more later.) A different problem arises when A has repeated eigenvalues with linearly indpendent eigenvectors (derogatory eigenvalues). A simple example: A = I (identity matrix). K k (A, x) = span{x, Ax,..., A k x} = span{x}. The Krylov method converges in one step (happy breakdown), exactly finding one copy of the eigenvalue λ = and eigenvector x.

18 reachable invariant subspaces A more perplexing example from Chris Beattie [Beattie, E., Rossi 24]: A =, x = c. By taking c large, we bias x toward the eigenvector [,,,, ] T. For any c the Krylov method breaks down (happily) at iteration k = 3, discovering the Jordan block and invariant subspace c V AV = S S, Range. The Krylov method finds the 3 3 Jordan block with eigenvector [,,,, ] T. Only a set of measure zero x can discover [,,,, ] T.

19 reachable invariant subspaces Unlucky choices of x C n can (in principle) prevent the Krylov method from seeing a desired (simple) eigenvector. This behavior is fragile to numerical computations, since infinitesimal perturbations to x will add a small component in the desired eigenvector. Single-vector Krylov methods can (in principle) find one Jordan block associated with each eigenvalue. This behavior is fragile to numerical computations, since infinitesimal perturbations split multiple eigenvalues. Block Krylov methods (with block size b, X C n b ) K k (A, X) = Range([X AX A 2 X A k X]), can find b linearly independent eigenvectors for a single eigenvalue.

20 How does the starting vector x affect convergence?

21 effect of starting vector on convergence Henceforth assume the desired invariant subspace U is reachable from x: U g K n(a, x). How does x influence convergence? For a single eigenpair (λ, u ) with spectral projector P, Saad [98] gives sin (u, K k (A, x)) P x min φ P k φ(λ )= (I P )ψ(a), The leading constant grows as the orientation of x toward u diminishes.

22 effect of starting vector on convergence Henceforth assume the desired invariant subspace U is reachable from x: U g K n(a, x). How does x influence convergence? For a single eigenpair (λ, u ) with spectral projector P, Saad [98] gives sin (u, K k (A, x)) P x min φ P k φ(λ )= (I P )ψ(a), The leading constant grows as the orientation of x toward u diminishes. For m-dimensional invariant subspaces U g, our bounds replace / P x with ψ(a)p b x C := max ψ P m ψ(a)p = max P b v gx v K m(a,x) P, gv the ratio of the bad to the good component in the worst approximation to U g from the m-dimensional Krylov space.

23 effect of starting vector on convergence A = symmetric matrix (n = 28) with 8 desired eigenvalues in [, 2]; 2 undesired eigenvalues in [, ]. θ = (x, U g), the angle between x and its best approximation in U g. δ(ug, Kk(A, x)) θ = 6 θ = 9 θ = 2 θ = 5 θ = θ = 3 iteration, k

24 effect of starting vector on convergence A = symmetric matrix (n = 28) with 8 desired eigenvalues in [, 2]; 2 undesired eigenvalues in [, ]. θ = (x, U g), the angle between x and its best approximation in U g. θ = 5 θ = 3 P b v C = max v K m(a,x) P gv δ(ug, Kk(A, x)) -5 - θ = 6 θ = 9 θ = 2 θ = [BER 24] bound iteration, k

25 How do the eigenvalues of A affect convergence?

26 asymptotic convergence rate determined by eigenvalues min φ P k 2m max z Ω b α(z)φ(z) P k 2m = set of polynomials of degree k 2m or less. Ω b C contains the undesired eigenvalues. α(z) = (z λ ) (z λ m). The polynomial approximation problem gives convergence like C γ k for some constant C and rate γ. When A = A, Ω b = [λ m+, λ n], and use Chebyshev polynomials to compute the convergence rate γ. When Ω b is a simply connected open subset of C, use conformal mapping to approach the approximation problem.

27 potential theoretic determination of the convergence rate Step : Begin by identifying undesired ( ) and desired ( ) eigenvalues.

28 potential theoretic determination of the convergence rate Step 2: Bound bad eigenvalues with Ω b.

29 potential theoretic determination of the convergence rate Step 3: Conformally map C \ Ω b to the exterior of the unit disk.

30 potential theoretic determination of the convergence rate Step 4: Find the lowest level curve of the Green s function intersecting a good eigenvalues.

31 potential theoretic determination of the convergence rate Step 5: Invert map to obtain curves of constant convergence rate in the original domain. Black circles on final figure are Fejér points, asymptotically optimal interpolation points for φ.

32 convergence rate: and granularity of the spectrum If convergence is very slow, perhaps you are solving the wrong problem.

33 convergence rate: and granularity of the spectrum If convergence is very slow, perhaps you are solving the wrong problem. Consider m =, where we can use the elementary bound [Saad 98] sin (u, K k (A, x)) P x min φ P k φ(λ )= (I P )ψ(a). Suppose A = A and we seek leftmost eigenvalue λ, where λ < λ 2 λ n. The error bound suggests the progress made at each iteration is like γ := κ κ +, where κ := λn λ λ 2 λ. When A is a discretization of an unbounded operator, we expect λ n = A as n. The convergence rate goes to one as n. Thus Krylov subspace methods often perform poorly for PDE eigenvalue problems unless the set-up is modified.

34 convergence slows as the discretization improves Apply the Krylov method to discretizations of the Laplacian in one dimension. How does the convergence rate change as the discretization improves? n =

35 convergence slows as the discretization improves Apply the Krylov method to discretizations of the Laplacian in one dimension. How does the convergence rate change as the discretization improves? n = 52 n =

36 convergence slows as the discretization improves Apply the Krylov method to discretizations of the Laplacian in one dimension. How does the convergence rate change as the discretization improves? n = 496 n = 248 n = 24 n = 52 n =

37 Convergence of Krylov Subspace Projection The problem becomes immediately apparent if we attempt to run Krylov subspace projection on the operator itself, K k (L, f ) = span{f, Lf,..., L k f }. For Lu = u with Dirichlet boundary conditions, u() = u() =, take some starting vector f Dom(L), i.e., f () = f () =. In general Lf Dom(L), so we cannot build the next Krylov direction L 2 f = L(Lf ). The Krylov algorithm breaks down at the third step.

38 Convergence of Krylov Subspace Projection The problem becomes immediately apparent if we attempt to run Krylov subspace projection on the operator itself, K k (L, f ) = span{f, Lf,..., L k f }. For Lu = u with Dirichlet boundary conditions, u() = u() =, take some starting vector f Dom(L), i.e., f () = f () =. In general Lf Dom(L), so we cannot build the next Krylov direction L 2 f = L(Lf ). The Krylov algorithm breaks down at the third step. The operator setting suggests that we instead apply Krylov to L : K k (L, f ) = span{f, L f,..., L (k ) f }. In this case, L is a beautiful compact operator: (L f )(x) = f + C + C x, where we choose C and C so that (L f )() = (L f )() =.

39 krylov projection applied to the operator We run the Krylov method on L exactly in Mathematica. Denote the eigenvalue estimates at the kth iteration as θ (k) θ (k) 2 θ (k) k. j3 (k) j! 6jj 4 j = j = k Observe superlinear convergence as k increases. For CG and GMRES applied to operators, see [Winther 98], [Nevanlinna 993], [Moret 997], [Olver 29], [Kirby 2]. For superlinear convergence in finite dimensional settings, see [van der Sluis, van der Vorst, 986], [van der Vorst, Vuik, 992], [Beattie, E., Rossi 24], [Simoncini, Szyld 25].

40 krylov projection applied to the operator We run the Krylov method on L exactly in Mathematica. Denote the eigenvalue estimates at the kth iteration as θ (k) θ (k) 2 θ (k) k. j3 (k) j! 6jj 4 j = j = k Observe superlinear convergence as k increases. For CG and GMRES applied to operators, see [Winther 98], [Nevanlinna 993], [Moret 997], [Olver 29], [Kirby 2]. For superlinear convergence in finite dimensional settings, see [van der Sluis, van der Vorst, 986], [van der Vorst, Vuik, 992], [Beattie, E., Rossi 24], [Simoncini, Szyld 25]. This mode of computation is preferred for discretization matrices as well: the shift-invert Arnoldi method uses K k ((A µi), x).

41 polynomial preconditioning: a cheap spectral transformation Replace the conventional Krylov space K k (A, x) = span{x, Ax, A 2 x,..., A k x} with the polynomial preconditioned transformation K k (π(a), x) = span{x, π(a)x, π(a) 2 x,..., π(a) k x}. [Thornquist 26], [E., Loe, Morgan arxiv:86.82] Use the polynomial π to separate the interesting eigenvalues. Often increases matvecs, but decreases iterations (hence orthogonalization). For example, with Hermitian A, take π to be the degree-d MINRES residual polynomial [Paige & Saunders 975], which attains min p(a)b. p P d p()= This polynomial tends to separate smallest-magnitude eigenvalues.

42 polynomial preconditioning: a cheap spectral transformation Polynomial preconditioning: Hermitian A, MINRES polynomial horizontal axis: σ(a) vertical axis σ(π(a))

43 How do the eigenvectors of A affect convergence?

44 bounding functions of a matrix If A is normal (A A = AA ), eigenvectors are orthogonal. For f analytic on σ(a), f (A) = max f (λ). λ σ(a) For nonnormal A, the situation is considerably more complicated. If A is diagonalizable, A = UΛU, then f (A) U U max f (λ). λ σ(a) For the numerical range (field of values) W (A) = {v Av : v = }, ( f (A) + ) 2 max f (z). z W (A) For the ε-pseudospectrum σ ε(a) = {z σ(a + E) for some E < ε}, f (A) L ε 2πε max f (z), z σ ε(a) where L ε is the boundary length of σ ε(a).

45 constant to account for nonnormality C 2 = C 2(A, Ω b ) comes from bounding f (A Ub ), f (z) = α(z)φ(z). Theorem [Beattie, E., Sorensen 25]. Suppose U g is reachable from the Krylov space K k (A, x). Then for k 2m, δ(u g, K k (A, x)) C C 2 min φ P k 2m max z Ω b α(z)φ(z). If Ω b = σ(a Ub ) (no defective eigenvalues), then C 2 = U b U + b, where the columns of U b C n (n m) are eigenvectors of A Ub. If Ω b = W (A Ub ) then C 2 = + 2. If Ω b = σ ε(a Ub ) then C 2 = L ε/(2πε). Tension: balance C 2 verses size of Ω b.

46 transient behavior of the power method Large coefficients in the expansion of x in the eigenvector basis can lead to cancellation effects in x k = A k x. Example: here different choices of α and β affect eigenvalue conditioning, α 4α 8αβ/2 A = 3/4 β, u =, u 2 =, u 3 = 2β/3 3/4. u3 u3 6 u2 x x 2 x 4 x 6 x 8 x x 7 5 x 3 x u x x,..., x 8 u2 u u2 x 3 x 5 x 7 x u3 x x 2 x 4 x 6 x 8 u α = β = α =, β = α =, β = [Trefethen & E. 25]

47 restarting krylov suspace methods an essential tool for controlling subspace dimension

48 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired eigenvalues desired eigenvalues To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace.5 Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x To understand convergence, one must understand how the Ritz values are distributed over σ(a).

49 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired eigenvalues desired eigenvalues To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace.5 Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x Ritz values To understand convergence, one must understand how the Ritz values are distributed over σ(a).

50 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired eigenvalues desired eigenvalues To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace.5 Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x filter root approx. eig To understand convergence, one must understand how the Ritz values are distributed over σ(a).

51 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired desired >=.75 To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x filter root approx. eig <= Shade log (magnitude of filter polynomial) To understand convergence, one must understand how the Ritz values are distributed over σ(a).

52 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired desired >=.75 To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x. To understand convergence, one must understand how the Ritz values are distributed over σ(a) filter root approx. eig <=-2 Shade log (magnitude of filter polynomial) Pushing the language Haim Avron used the previous talk, a standard Krylov method (fixed k) is a sketch-and-solve method, while restarted Krylov methods sketch-to-precondition.

53 restarting krylov methods restarted Arnoldi algorithm (eigs).5 undesired >=.75 To compute m < k eigenvalues of A C n n, Arnoldi methods restrict A to act on the k-dimensional Krylov subspace Ran(V) = span{x, Ax,..., A k x}. Compute eigenvalues of V AV (Ritz values), and order them by relevance; e.g., if seeking the rightmost eigenvalue of A, let Re θ Re θ 2 Re θ k. Exact shifts [Sorensen 992] restart the method, attempting to improve v with a polynomial filter having the unwanted Ritz values as roots: x + = (A θ m+ I) (A θ k I)x. To understand convergence, one must understand how the Ritz values are distributed over σ(a) filter root approx. eig <=-2 Shade log (magnitude of filter polynomial) [Sorensen 992] proved convergence for A = A. The process fails for some A A [E. 29], [Duintjer Tebbens, Meurant 22]. Stringent sufficient conditions are known [Carden 2].

54 Do nonsymmetric matrices enjoy any kind of interlacing?

55 interlacing is a key to convergence theory for A = A Cauchy s interlacing theorem assures us that Ritz values cannot bunch up at the ends of the spectrum. iteration, k eigenvalues (red lines) and Ritz values (black dots)

56 interlacing does not hold for A = A The absence of interlacing for non-hermitian problems is the major impediment to a full convergence theory and is the mechanism that allows the method to fail (in theory). iteration, k eigenvalues (red lines) and Ritz values (black dots = real parts)

57 a pathologically terrible example A monster, built using the construction of [Duintjer Tebbens, Meurant 22]: A = , x =. arxiv:

58 ritz value localization for non-hermitian matrices Do non-hermitian matrices obey any kind of interlacing theorem? Ritz values must be contained within the numerical range W (A) = {v A v : v = }, a closed, convex subset of C that contains σ(a).

59 ritz value localization for non-hermitian matrices Do non-hermitian matrices obey any kind of interlacing theorem? Ritz values must be contained within the numerical range W (A) = {v A v : v = }, a closed, convex subset of C that contains σ(a). Consider an extreme example: A =, W (A) = Repeat the following experiment many times: { z C : z } 2. 2 Generate random two dimensional subspaces, V = Ran V, where V V = I. Form V AV C 2 2 and compute Ritz values {θ, θ 2} = σ(v AV). Identify the leftmost and rightmost Ritz values. Since σ(a) = {}, interlacing is meaningless here...

60 ritz values of a jordan block leftmost Ritz value W (A) = { z C : z rightmost Ritz value } 2 2

61 ritz values of a jordan block leftmost Ritz value rightmost Ritz value, random (complex) two dimensional subspaces

62 three matrices with identical W (A) Compute k = 4 Ritz values for these 8 8 matrices. γ γ3 (γ and γ3 set to give same W (A) for all examples; % = /8.) % %2 %3 %4 %5 %6 %7.

63 three matrices with identical W (A) Compute k = 4 Ritz values for these 8 8 matrices. γ γ3 % %2 %3 %4 %5 %6 %7 (γ and γ3 set to give same W (A) for all examples; % = /8.) Smallest magnitude of k = 4 Ritz values,, random complex subspaces..

64 ritz value localization, sorted by real part Using Schur s eigenvalue majorization theorem for Hermitian matrices, we can establish an interlacing-type result. Theorem (Carden & E. 22) Let θ,..., θ k denote the Ritz values of A C n n drawn from a k < n dimensional subspace, labeled by decreasing real part: Re θ Re θ k. Then for j =,..., k, µ n k+j + + µ n k j + Re θ j µ + + µ j, j where µ µ n are the eigenvalues of 2 (A + A ).

65 ritz value localization, sorted by real part Using Schur s eigenvalue majorization theorem for Hermitian matrices, we can establish an interlacing-type result. Theorem (Carden & E. 22) Let θ,..., θ k denote the Ritz values of A C n n drawn from a k < n dimensional subspace, labeled by decreasing real part: Re θ Re θ k. Then for j =,..., k, µ n k+j + + µ n k j + Re θ j µ + + µ j, j where µ µ n are the eigenvalues of 2 (A + A ). The fact that θ j W (A) gives the well-known bound µ Re θ j µ n, j =,..., k. The theorem provides sharper bounds for interior Ritz values. The interior eigenvalues of 2 (A + A ) give additional insight; cf. eigenvalue inclusion regions of [Psarrakos & Tsatsomeros, 22]. Theorem applies to any subspace Range(V): Krylov, block Krylov, etc.

66 three matrices with identical W (A) Three matrices with the same W (A), different interior structure; 2 trials. For k = 4, numbers on right indicate max Ritz values in each region. A 4 A A

67 ritz value localization, sorted by magnitude The log-majorization of products of eigenvalues by products of singular values [Marshall, Olkin, Arnold 2] leads to a limit on Ritz value magnitudes. Theorem (Carden & E., 22) Let θ,..., θ k denote the Ritz values of A C n n drawn from a k < n dimensional subspace, labeled by decreasing magnitude: θ θ k. Then for j =,..., k, θ j (s s j ) /j, where s s n are the singular values of A.

68 ritz value localization, sorted by magnitude The log-majorization of products of eigenvalues by products of singular values [Marshall, Olkin, Arnold 2] leads to a limit on Ritz value magnitudes. Theorem (Carden & E., 22) Let θ,..., θ k denote the Ritz values of A C n n drawn from a k < n dimensional subspace, labeled by decreasing magnitude: θ θ k. Then for j =,..., k, θ j (s s j ) /j, where s s n are the singular values of A. Related results: Zvonimir Bujanovic [2] studies Ritz values of normal matrices from Krylov subspaces in his Ph.D. thesis (Zagreb). Jakeniah Christiansen [22] studies real Ritz values for n = 3 (SIURO).

69 some closing thoughts Krylov methods can further develop as a prominent tool for RandNLA. Polynomials are better than powers! Krylov methods have great advantages over power/subspace iteration. Block methods hold promise but additional subtleties. Large subspaces can be built rapidly; must maintain linear independence. Restarting is crucial in engineering computations, but analysis is tricky. Restarting controls the subspace dimension, refines the starting vector. Spectral transformations (shift-invert) can vastly accelerate convergence. You are not entirely constrained by the eigenvalue distribution of A. Non-Hermitian problems are solved everyday. The theory is incomplete and monsters are easy to construct, but the Krylov method (as implemented in eigs/arpack) works well. Can Random Matrix Theory shed light on Ritz value locations? What is the probability that A is stable if V AV is stable?

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Key words. Krylov subspace methods, Arnoldi algorithm, eigenvalue computations, containment gap, pseudospectra

Key words. Krylov subspace methods, Arnoldi algorithm, eigenvalue computations, containment gap, pseudospectra CONVERGENCE OF POLYNOMIAL RESTART KRYLOV METHODS FOR EIGENVALUE COMPUTATIONS CHRISTOPHER A. BEATTIE, MARK EMBREE, AND D. C. SORENSEN Abstract. Krylov subspace methods have proved effective for many non-hermitian

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

c 2009 Society for Industrial and Applied Mathematics

c 2009 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 3, No., pp. c 29 Society for Industrial and Applied Mathematics THE ARNOLDI EIGENVALUE ITERATION WITH EXACT SHIFTS CAN FAIL MARK EMBREE Abstract. The restarted Arnoldi algorithm,

More information

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard

More information

On the Ritz values of normal matrices

On the Ritz values of normal matrices On the Ritz values of normal matrices Zvonimir Bujanović Faculty of Science Department of Mathematics University of Zagreb June 13, 2011 ApplMath11 7th Conference on Applied Mathematics and Scientific

More information

Convergence of Polynomial Restart Krylov Methods for Eigenvalue Computations

Convergence of Polynomial Restart Krylov Methods for Eigenvalue Computations SIAM REVIEW Vol. 47,No. 3,pp. 492 515 c 2005 Society for Industrial and Applied Mathematics Convergence of Polynomial Restart Krylov Methods for Eigenvalue Computations Christopher A. Beattie Mark Embree

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems

More information

LARGE SPARSE EIGENVALUE PROBLEMS

LARGE SPARSE EIGENVALUE PROBLEMS LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems

More information

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A

More information

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna. Recent advances in approximation using Krylov subspaces V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The framework It is given an operator

More information

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012 On the Superlinear Convergence of MINRES Valeria Simoncini and Daniel B. Szyld Report 12-01-11 January 2012 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld 0 Chapter

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES 1 PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES MARK EMBREE, THOMAS H. GIBSON, KEVIN MENDOZA, AND RONALD B. MORGAN Abstract. fill in abstract Key words. eigenvalues, multiple eigenvalues, Arnoldi,

More information

Krylov subspace projection methods

Krylov subspace projection methods I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY RONALD B. MORGAN AND MIN ZENG Abstract. A restarted Arnoldi algorithm is given that computes eigenvalues

More information

The Eigenvalue Problem: Perturbation Theory

The Eigenvalue Problem: Perturbation Theory Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

c 2004 Society for Industrial and Applied Mathematics

c 2004 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 25, No. 4, pp. 1074 1109 c 2004 Society for Industrial and Applied Mathematics CONVERGENCE OF RESTARTED KRYLOV SUBSPACES TO INVARIANT SUBSPACES CHRISTOPHER BEATTIE, MARK

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

THE STABILITY OF GMRES CONVERGENCE, WITH APPLICATION TO APPROXIMATE DEFLATION PRECONDITIONING. 16 September 2011

THE STABILITY OF GMRES CONVERGENCE, WITH APPLICATION TO APPROXIMATE DEFLATION PRECONDITIONING. 16 September 2011 THE STABILITY OF GMRES CONVERGENCE, WITH APPLICATION TO APPROXIMATE DEFLATION PRECONDITIONING JOSEF A. SIFUENTES, MARK EMBREE, AND RONALD B. MORGAN 16 September 2011 Abstract. How does GMRES convergence

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini Iterative solvers for saddle point algebraic linear systems: tools of the trade V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The problem

More information

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR POLYNOMIAL PRECONDITIONED BICGSTAB AND IDR JENNIFER A. LOE AND RONALD B. MORGAN Abstract. Polynomial preconditioning is applied to the nonsymmetric Lanczos methods BiCGStab and IDR for solving large nonsymmetric

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Antti Koskela KTH Royal Institute of Technology, Lindstedtvägen 25, 10044 Stockholm,

More information

MULTIGRID ARNOLDI FOR EIGENVALUES

MULTIGRID ARNOLDI FOR EIGENVALUES 1 MULTIGRID ARNOLDI FOR EIGENVALUES RONALD B. MORGAN AND ZHAO YANG Abstract. A new approach is given for computing eigenvalues and eigenvectors of large matrices. Multigrid is combined with the Arnoldi

More information

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.

More information

ABSTRACT. Professor G.W. Stewart

ABSTRACT. Professor G.W. Stewart ABSTRACT Title of dissertation: Residual Arnoldi Methods : Theory, Package, and Experiments Che-Rung Lee, Doctor of Philosophy, 2007 Dissertation directed by: Professor G.W. Stewart Department of Computer

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter

More information

of dimension n 1 n 2, one defines the matrix determinants

of dimension n 1 n 2, one defines the matrix determinants HARMONIC RAYLEIGH RITZ FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter eigenvalue

More information

Department of Computer Science, University of Illinois at Urbana-Champaign

Department of Computer Science, University of Illinois at Urbana-Champaign Department of Computer Science, University of Illinois at Urbana-Champaign Probing for Schur Complements and Preconditioning Generalized Saddle-Point Problems Eric de Sturler, sturler@cs.uiuc.edu, http://www-faculty.cs.uiuc.edu/~sturler

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Index. for generalized eigenvalue problem, butterfly form, 211

Index. for generalized eigenvalue problem, butterfly form, 211 Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

ELA

ELA A METHOD FOR THE INVERSE NUMERICAL RANGE PROBLEM CHRISTOS CHORIANOPOULOS, PANAYIOTIS PSARRAKOS, AND FRANK UHLIG Abstract. For a given complex square matrix A, we develop, implement and test a fast geometric

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1

More information

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology The Davidson method is a popular technique to compute a few of the smallest (or largest)

More information

Inexact and Nonlinear Extensions of the FEAST Eigenvalue Algorithm

Inexact and Nonlinear Extensions of the FEAST Eigenvalue Algorithm University of Massachusetts Amherst ScholarWorks@UMass Amherst Doctoral Dissertations Dissertations and Theses 2018 Inexact and Nonlinear Extensions of the FEAST Eigenvalue Algorithm Brendan E. Gavin University

More information

Alternative correction equations in the Jacobi-Davidson method

Alternative correction equations in the Jacobi-Davidson method Chapter 2 Alternative correction equations in the Jacobi-Davidson method Menno Genseberger and Gerard Sleijpen Abstract The correction equation in the Jacobi-Davidson method is effective in a subspace

More information

The rational Krylov subspace for parameter dependent systems. V. Simoncini

The rational Krylov subspace for parameter dependent systems. V. Simoncini The rational Krylov subspace for parameter dependent systems V. Simoncini Dipartimento di Matematica, Università di Bologna valeria.simoncini@unibo.it 1 Given the continuous-time system Motivation. Model

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Rational Krylov methods for linear and nonlinear eigenvalue problems

Rational Krylov methods for linear and nonlinear eigenvalue problems Rational Krylov methods for linear and nonlinear eigenvalue problems Mele Giampaolo mele@mail.dm.unipi.it University of Pisa 7 March 2014 Outline Arnoldi (and its variants) for linear eigenproblems Rational

More information

Math 504 (Fall 2011) 1. (*) Consider the matrices

Math 504 (Fall 2011) 1. (*) Consider the matrices Math 504 (Fall 2011) Instructor: Emre Mengi Study Guide for Weeks 11-14 This homework concerns the following topics. Basic definitions and facts about eigenvalues and eigenvectors (Trefethen&Bau, Lecture

More information

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer

More information

Inexactness and flexibility in linear Krylov solvers

Inexactness and flexibility in linear Krylov solvers Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7) - IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy - October 15-19, 2007 in honor of Gérard Meurant for his 60 th

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Lecture notes: Applied linear algebra Part 2. Version 1

Lecture notes: Applied linear algebra Part 2. Version 1 Lecture notes: Applied linear algebra Part 2. Version 1 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 First, some exercises: xercise 0.1 (2 Points) Another least

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues

More information

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers MAX PLANCK INSTITUTE International Conference on Communications, Computing and Control Applications March 3-5, 2011, Hammamet, Tunisia. Model order reduction of large-scale dynamical systems with Jacobi-Davidson

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices

A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices Jean Christophe Tremblay and Tucker Carrington Chemistry Department Queen s University 23 août 2007 We want to

More information

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation

More information

MICHIEL E. HOCHSTENBACH

MICHIEL E. HOCHSTENBACH VARIATIONS ON HARMONIC RAYLEIGH RITZ FOR STANDARD AND GENERALIZED EIGENPROBLEMS MICHIEL E. HOCHSTENBACH Abstract. We present several variations on the harmonic Rayleigh Ritz method. First, we introduce

More information

Lecture 3: Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 2012 Overview of the Course These lectures

More information

Analysis Preliminary Exam Workshop: Hilbert Spaces

Analysis Preliminary Exam Workshop: Hilbert Spaces Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H

More information

The rate of convergence of the GMRES method

The rate of convergence of the GMRES method The rate of convergence of the GMRES method Report 90-77 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wiskunde en Informatica Faculty of Technical Mathematics

More information

Block Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems

Block Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems Mitglied der Helmholtz-Gemeinschaft Block Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems Birkbeck University, London, June the 29th 2012 Edoardo Di Napoli Motivation and Goals

More information

The EVSL package for symmetric eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota

The EVSL package for symmetric eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota The EVSL package for symmetric eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota 15th Copper Mountain Conference Mar. 28, 218 First: Joint work with

More information

Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods

Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods Lars Eldén Department of Mathematics Linköping University, Sweden Joint work with Valeria Simoncini February 21 Lars Eldén

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems Mario Arioli m.arioli@rl.ac.uk CCLRC-Rutherford Appleton Laboratory with Daniel Ruiz (E.N.S.E.E.I.H.T)

More information

randomized block krylov methods for stronger and faster approximate svd

randomized block krylov methods for stronger and faster approximate svd randomized block krylov methods for stronger and faster approximate svd Cameron Musco and Christopher Musco December 2, 25 Massachusetts Institute of Technology, EECS singular value decomposition n d left

More information

Matrix functions and their approximation. Krylov subspaces

Matrix functions and their approximation. Krylov subspaces [ 1 / 31 ] University of Cyprus Matrix functions and their approximation using Krylov subspaces Matrixfunktionen und ihre Approximation in Krylov-Unterräumen Stefan Güttel stefan@guettel.com Nicosia, 24th

More information

An Analysis of Linear Models, Linear Value-Function Approximation, and Feature Selection For Reinforcement Learning

An Analysis of Linear Models, Linear Value-Function Approximation, and Feature Selection For Reinforcement Learning An Analysis of Linear Models, Linear Value-Function Approximation, and Feature Selection For Reinforcement Learning Ronald Parr, Lihong Li, Gavin Taylor, Christopher Painter-Wakefield, Michael L. Littman

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

ILAS 2017 July 25, 2017

ILAS 2017 July 25, 2017 Polynomial and rational filtering for eigenvalue problems and the EVSL project Yousef Saad Department of Computer Science and Engineering University of Minnesota ILAS 217 July 25, 217 Large eigenvalue

More information

A hybrid reordered Arnoldi method to accelerate PageRank computations

A hybrid reordered Arnoldi method to accelerate PageRank computations A hybrid reordered Arnoldi method to accelerate PageRank computations Danielle Parker Final Presentation Background Modeling the Web The Web The Graph (A) Ranks of Web pages v = v 1... Dominant Eigenvector

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question. Math 304 Final Exam (May 8) Spring 206 No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question. Name: Section: Question Points

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

ON BEST APPROXIMATIONS OF POLYNOMIALS IN MATRICES IN THE MATRIX 2-NORM

ON BEST APPROXIMATIONS OF POLYNOMIALS IN MATRICES IN THE MATRIX 2-NORM ON BEST APPROXIMATIONS OF POLYNOMIALS IN MATRICES IN THE MATRIX 2-NORM JÖRG LIESEN AND PETR TICHÝ Abstract. We show that certain matrix approximation problems in the matrix 2-norm have uniquely defined

More information

PRECONDITIONED ITERATIVE METHODS FOR LINEAR SYSTEMS, EIGENVALUE AND SINGULAR VALUE PROBLEMS. Eugene Vecharynski. M.S., Belarus State University, 2006

PRECONDITIONED ITERATIVE METHODS FOR LINEAR SYSTEMS, EIGENVALUE AND SINGULAR VALUE PROBLEMS. Eugene Vecharynski. M.S., Belarus State University, 2006 PRECONDITIONED ITERATIVE METHODS FOR LINEAR SYSTEMS, EIGENVALUE AND SINGULAR VALUE PROBLEMS by Eugene Vecharynski M.S., Belarus State University, 2006 A thesis submitted to the University of Colorado Denver

More information

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction

More information

The Life Cycle of an Eigenvalue Problem

The Life Cycle of an Eigenvalue Problem The Life Cycle of an Eigenvalue Problem Mark Embree Department of Mathematics Virginia Tech Jeffrey Hokanson Charles Puelz Computational and Applied Mathematics Rice University Introduction We care about

More information