High-performance computation of pseudospectra

Size: px
Start display at page:

Download "High-performance computation of pseudospectra"

Transcription

1 1 / 80 High-performance computation of pseudospectra Jack Poulson Department of Mathematics Stanford University Innovative Computing Laboratory The University of Tennessee October 24, 2014

2 A note on collaborators 2 / 80 Multishift Hessenberg solves: collaboration with Greg Henry

3 3 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

4 4 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

5 5 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

6 6 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

7 7 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

8 8 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

9 9 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

10 Outline 10 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

11 11 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot

12 12 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot

13 13 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot

14 Van Loan s algorithm and the Demmel matrix 14 / 80 Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot D(n, β) = (β J β 1,n ) 1, n = 3, β = 100

15 Outline 15 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

16 Pseudospectra 16 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential

17 Pseudospectra 17 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential

18 Pseudospectra 18 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential

19 Pseudospectra 19 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential

20 Pseudospectra 20 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential

21 Outline 21 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

22 22 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

23 23 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

24 24 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

25 25 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

26 26 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

27 27 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]

28 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 28 / 80

29 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 29 / 80

30 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 30 / 80

31 (Extended) Van Loan algorithm for L 2 ɛ (A) 31 / 80 Algorithm: Two-norm pseudospectra via extended Van Loan algorithm Input: A F n n, shifts Ω C, restart size k Output: {φ(ξ)} ξ Ω { (ξi A) 1 2 } ξ Ω G := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (ξi G) 1 2 via Restarted Arnoldi Choose v 0 C n with v 0 2 = 1 while not converged do for j = 0,..., k 1 do // (ξi G) H (ξi G) 1 V j = V j H j + v j (β j e j ) H x j := (ξi G) H (ξi G) 1 v j Expand Arnoldi decomposition using x j [λ, v 0 ] := MaxEig(H k ) φ(ξ) := RealPart(λ)

32 [Hager-1984,Higham-1988] approach for L 1 ɛ (A) 32 / 80 Algorithm: One-norm pseudospectra via Hager-Higham algorithm Input: A F n n, Ω C Output: {φ(ξ)} ξ Ω { (A ξi) 1 1 } ξ Ω [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (A ξi) 1 1 via Hager-Higham algorithm Choose u C n with u 1 = 1 k := 0 repeat x := u y := Q(G ξi) 1 Q H x w := Q(G ξi) H Q H sign(y) u := e j where w(j) = w k := k + 1 until (k 2 and w w H x) or k 5 φ(ξ) := y 1 // Take the maximum between the current estimate and a heuristic b := 2 3n [( 1)j (n + j 1)] j=0:n 1 x := Q(G ξi) 1 Q H b φ(ξ) := max(φ(ξ), x 1 ) Block algorithm [Higham/Tisseur-2000] should be used in practice

33 [Hager-1984,Higham-1988] approach for L 1 ɛ (A) 33 / 80 Algorithm: One-norm pseudospectra via Hager-Higham algorithm Input: A F n n, Ω C Output: {φ(ξ)} ξ Ω { (A ξi) 1 1 } ξ Ω [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (A ξi) 1 1 via Hager-Higham algorithm Choose u C n with u 1 = 1 k := 0 repeat x := u y := Q(G ξi) 1 Q H x w := Q(G ξi) H Q H sign(y) u := e j where w(j) = w k := k + 1 until (k 2 and w w H x) or k 5 φ(ξ) := y 1 // Take the maximum between the current estimate and a heuristic b := 2 3n [( 1)j (n + j 1)] j=0:n 1 x := Q(G ξi) 1 Q H b φ(ξ) := max(φ(ξ), x 1 ) Block algorithm [Higham/Tisseur-2000] should be used in practice

34 Outline 34 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

35 35 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

36 36 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

37 37 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

38 38 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

39 39 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

40 40 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

41 41 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...

42 Interleaved Van Loan algorithm for L 2 ɛ (A) 42 / 80 Algorithm: Two-norm pseudospectra via interleaved extended Van Loan algorithm Input: A F n n, shift vector z C m, restart size k Output: f [ (z(s)i A) 1 2 ]s=0:m 1 G := Schur(A), RealSchur(A), or Hessenberg(A) Initialize each column of W 0 C n m to have unit two-norm I := (0,..., m 1) while I do for j = 0,..., k 1 do // (z(s)i G) H (z(s)i G) 1 W j (t) = W j (t)h j (t) + W j (:, t)(b j (t)e j ) H, s = I(t) X := MultishiftSolve(G, z(i), W j ) X := MultishiftSolve(G H, z(i), X) Expand Arnoldi decompositions foreach s = I(t) do [λ, W 0 (:, t)] := MaxEig(H k (t)) f (s) := RealPart(λ) if converged then Delete I(t) and W 0 (:, t)

43 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 43 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) Initialize each column of U C n m to have unit one-norm I := (0,..., m 1) k := 0 while I do X := U Y := Q H X Y := MultishiftSolve(G, z(i), Y ) Y := QY W := Q H sign(y ) W := MultishiftSolve(G H, z(i), W ) W := QW k := k + 1 foreach s = I(t) do x := X(:, t), y := Y (:, t), w := W (:, t) if (k 2 and w w H x) or k 5 then f (s) = y 1 Delete I(t) and U(:, t) else U(:, t) := e j where w(j) = w

44 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 44 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 // Take the maximum between the current estimates and a heuristic B := 2 3n [( 1)j (n + j 1)] j=0:n 1,s=0:m 1 // All columns are equal X := Q H B // All columns are equal X := MultishiftSolve(G, z, X) X := QX foreach s = 0,..., m 1 do f (s) := max(f (s), X(:, s) 1 ) Again: block algorithm from [Higham/Tisseur-2000] should be used in practice

45 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 45 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 // Take the maximum between the current estimates and a heuristic B := 2 3n [( 1)j (n + j 1)] j=0:n 1,s=0:m 1 // All columns are equal X := Q H B // All columns are equal X := MultishiftSolve(G, z, X) X := QX foreach s = 0,..., m 1 do f (s) := max(f (s), X(:, s) 1 ) Again: block algorithm from [Higham/Tisseur-2000] should be used in practice

46 46 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...

47 47 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...

48 48 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...

49 Outline 49 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

50 Investigating the Fox-Li matrix with python 50 / 80 import math, El n = 100 # m a t r i x s ize realres = imagres = 100 # g r i d r e s o l u t i o n # Display an instance of the Fox L i / Landau matrix A = El. D i s t M a t r i x ( El. ztag ) El. FoxLi (A, n ) El. Display (A, " Fox L i m a t r i x " ) # Display i t s s p e c t r a l p o r t r a i t p o r t r a i t = El. S p e c t r a l P o r t r a i t (A, realres, imagres ) El. EntrywiseMap ( p o r t r a i t, math. log10 ) El. Display ( p o r t r a i t, " S p e c t r a l p o r t r a i t of Fox L i matrix " ) # Display i t s s i n g u l a r values s = El.SVD(A) El. EntrywiseMap ( s, math. log10 ) El. Display ( s, " log10 ( svd (A ) ) " )

51 Investigating the Fox-Li matrix with python 51 / 80

52 Investigating the Fox-Li matrix with python 52 / 80

53 Investigating the Fox-Li matrix with python 53 / 80

54 Outline 54 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

55 55 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 10 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)

56 56 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 20 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)

57 57 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 30 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)

58 58 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 40 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)

59 59 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 50 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)

60 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 10 its 60 / 80

61 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 20 its 61 / 80

62 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 30 its 62 / 80

63 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 40 its 63 / 80

64 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 50 its 64 / 80

65 65 / 80 Uniform(20k), Ω = ( 103, 103) 2, pix pieces: 75 sec/iter on 256 nodes of Blue Gene/Q, (64 r.h.s./core, >11 TFlops), 1175 sec for Schur

66 66 / 80 Uniform(20k), Ω = ( 10.3, 10.3) 2, pix pieces: 75 sec/iter on 256 nodes of Blue Gene/Q, (64 r.h.s./core, >11 TFlops), 1175 sec for Schur

67 Outline 67 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work

68 68 / 80 Convergence issues Typically small number of iterations required for visual convergence. Given current batch iteration, can output image after each fixed number of iterations Problematic pixels typically uninteresting (resolvent gradient is small)

69 69 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

70 70 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

71 71 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

72 72 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

73 73 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

74 74 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

75 75 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra

76 Acknowledgments and Availability 76 / 80 Support Computational resources My hosts Jakub Kurzak and Jack Dongarra Availability Elemental is available under the New BSD License at libelemental.org Questions?

77 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 77 / 80

78 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 78 / 80

79 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 79 / 80

80 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 80 / 80

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Accelerating computation of eigenvectors in the nonsymmetric eigenvalue problem

Accelerating computation of eigenvectors in the nonsymmetric eigenvalue problem Accelerating computation of eigenvectors in the nonsymmetric eigenvalue problem Mark Gates 1, Azzam Haidar 1, and Jack Dongarra 1,2,3 1 University of Tennessee, Knoxville, TN, USA 2 Oak Ridge National

More information

Accelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem

Accelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem Accelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem Mark Gates 1, Azzam Haidar 1, and Jack Dongarra 1,2,3 1 University of Tennessee, Knoxville, TN, USA 2 Oak Ridge National

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

1 Sylvester equations

1 Sylvester equations 1 Sylvester equations Notes for 2016-11-02 The Sylvester equation (or the special case of the Lyapunov equation) is a matrix equation of the form AX + XB = C where A R m m, B R n n, B R m n, are known,

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

LARGE SPARSE EIGENVALUE PROBLEMS

LARGE SPARSE EIGENVALUE PROBLEMS LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Numerical linear algebra

Numerical linear algebra Numerical linear algebra Purdue University CS 51500 Fall 2017 David Gleich David F. Gleich Call me Prof Gleich Dr. Gleich Please not Hey matrix guy! Huda Nassar Call me Huda Ms. Huda Please not Matrix

More information

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems

More information

Index. for generalized eigenvalue problem, butterfly form, 211

Index. for generalized eigenvalue problem, butterfly form, 211 Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,

More information

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu

More information

ARPACK. Dick Kachuma & Alex Prideaux. November 3, Oxford University Computing Laboratory

ARPACK. Dick Kachuma & Alex Prideaux. November 3, Oxford University Computing Laboratory ARPACK Dick Kachuma & Alex Prideaux Oxford University Computing Laboratory November 3, 2006 What is ARPACK? ARnoldi PACKage Collection of routines to solve large scale eigenvalue problems Developed at

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Lecture 2: Numerical linear algebra

Lecture 2: Numerical linear algebra Lecture 2: Numerical linear algebra QR factorization Eigenvalue decomposition Singular value decomposition Conditioning of a problem Floating point arithmetic and stability of an algorithm Linear algebra

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 2012 Overview of the Course These lectures

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

Hessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods

Hessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods Hessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods Jens Peter M. Zemke Minisymposium on Numerical Linear Algebra Technical

More information

Computing Unstructured and Structured Polynomial Pseudospectrum Approximations

Computing Unstructured and Structured Polynomial Pseudospectrum Approximations Computing Unstructured and Structured Polynomial Pseudospectrum Approximations Silvia Noschese 1 and Lothar Reichel 2 1 Dipartimento di Matematica, SAPIENZA Università di Roma, P.le Aldo Moro 5, 185 Roma,

More information

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve

More information

Eigenvalue Problems and Singular Value Decomposition

Eigenvalue Problems and Singular Value Decomposition Eigenvalue Problems and Singular Value Decomposition Sanzheng Qiao Department of Computing and Software McMaster University August, 2012 Outline 1 Eigenvalue Problems 2 Singular Value Decomposition 3 Software

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

Least squares and Eigenvalues

Least squares and Eigenvalues Lab 1 Least squares and Eigenvalues Lab Objective: Use least squares to fit curves to data and use QR decomposition to find eigenvalues. Least Squares A linear system Ax = b is overdetermined if it has

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

ON MATRIX BALANCING AND EIGENVECTOR COMPUTATION

ON MATRIX BALANCING AND EIGENVECTOR COMPUTATION ON MATRIX BALANCING AND EIGENVECTOR COMPUTATION RODNEY JAMES, JULIEN LANGOU, AND BRADLEY R. LOWERY arxiv:40.5766v [math.na] Jan 04 Abstract. Balancing a matrix is a preprocessing step while solving the

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

Lecture on: Numerical sparse linear algebra and interpolation spaces. June 3, 2014

Lecture on: Numerical sparse linear algebra and interpolation spaces. June 3, 2014 Lecture on: Numerical sparse linear algebra and interpolation spaces June 3, 2014 Finite dimensional Hilbert spaces and IR N 2 / 38 (, ) : H H IR scalar product and u H = (u, u) u H norm. Finite dimensional

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Main matrix factorizations

Main matrix factorizations Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined

More information

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Krylov Subspaces. Lab 1. The Arnoldi Iteration Lab 1 Krylov Subspaces Lab Objective: Discuss simple Krylov Subspace Methods for finding eigenvalues and show some interesting applications. One of the biggest difficulties in computational linear algebra

More information

c 2009 Society for Industrial and Applied Mathematics

c 2009 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 3, No., pp. c 29 Society for Industrial and Applied Mathematics THE ARNOLDI EIGENVALUE ITERATION WITH EXACT SHIFTS CAN FAIL MARK EMBREE Abstract. The restarted Arnoldi algorithm,

More information

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY RONALD B. MORGAN AND MIN ZENG Abstract. A restarted Arnoldi algorithm is given that computes eigenvalues

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Diagonalizing Matrices

Diagonalizing Matrices Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B

More information

Numerical Mathematics

Numerical Mathematics Alfio Quarteroni Riccardo Sacco Fausto Saleri Numerical Mathematics Second Edition With 135 Figures and 45 Tables 421 Springer Contents Part I Getting Started 1 Foundations of Matrix Analysis 3 1.1 Vector

More information

Math 3191 Applied Linear Algebra

Math 3191 Applied Linear Algebra Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful

More information

PERFORMANCE ENHANCEMENT OF PARALLEL MULTIFRONTAL SOLVER ON BLOCK LANCZOS METHOD 1. INTRODUCTION

PERFORMANCE ENHANCEMENT OF PARALLEL MULTIFRONTAL SOLVER ON BLOCK LANCZOS METHOD 1. INTRODUCTION J. KSIAM Vol., No., -0, 009 PERFORMANCE ENHANCEMENT OF PARALLEL MULTIFRONTAL SOLVER ON BLOCK LANCZOS METHOD Wanil BYUN AND Seung Jo KIM, SCHOOL OF MECHANICAL AND AEROSPACE ENG, SEOUL NATIONAL UNIV, SOUTH

More information

Orthogonal iteration to QR

Orthogonal iteration to QR Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete

More information

Real Eigenvalue Extraction and the Distance to Uncontrollability

Real Eigenvalue Extraction and the Distance to Uncontrollability Real Eigenvalue Extraction and the Distance to Uncontrollability Emre Mengi Computer Science Department Courant Institute of Mathematical Sciences New York University mengi@cs.nyu.edu May 22nd, 2006 Emre

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Matrix Condition Estimators in 2-norm

Matrix Condition Estimators in 2-norm Matrix Condition Estimators in 2-norm Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Miroslav Tůma Institute of Computer Science

More information

Some minimization problems

Some minimization problems Week 13: Wednesday, Nov 14 Some minimization problems Last time, we sketched the following two-step strategy for approximating the solution to linear systems via Krylov subspaces: 1. Build a sequence of

More information

PROJECTED GMRES AND ITS VARIANTS

PROJECTED GMRES AND ITS VARIANTS PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft

More information

Eigenvalue problems. Eigenvalue problems

Eigenvalue problems. Eigenvalue problems Determination of eigenvalues and eigenvectors Ax x, where A is an N N matrix, eigenvector x 0, and eigenvalues are in general complex numbers In physics: - Energy eigenvalues in a quantum mechanical system

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Research Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics

Research Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics Research Matters February 25, 2009 The Nonlinear Eigenvalue Problem Nick Higham Part III Director of Research School of Mathematics Françoise Tisseur School of Mathematics The University of Manchester

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Accelerating linear algebra computations with hybrid GPU-multicore systems.

Accelerating linear algebra computations with hybrid GPU-multicore systems. Accelerating linear algebra computations with hybrid GPU-multicore systems. Marc Baboulin INRIA/Université Paris-Sud joint work with Jack Dongarra (University of Tennessee and Oak Ridge National Laboratory)

More information

Singular Value Decomposition of the frequency response operator

Singular Value Decomposition of the frequency response operator Measures of input-output amplification (across frequency) Largest singular value Hilbert-Schmidt norm (power spectral density) Systems with one spatial variable Two point boundary value problems EE 8235:

More information

Matrix stabilization using differential equations.

Matrix stabilization using differential equations. Matrix stabilization using differential equations. Nicola Guglielmi Universitá dell Aquila and Gran Sasso Science Institute, Italia NUMOC-2017 Roma, 19 23 June, 2017 Inspired by a joint work with Christian

More information

On Solving Large Algebraic. Riccati Matrix Equations

On Solving Large Algebraic. Riccati Matrix Equations International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University

More information

Problem Set Number 5, j/2.036j MIT (Fall 2014)

Problem Set Number 5, j/2.036j MIT (Fall 2014) Problem Set Number 5, 18.385j/2.036j MIT (Fall 2014) Rodolfo R. Rosales (MIT, Math. Dept.,Cambridge, MA 02139) Due Fri., October 24, 2014. October 17, 2014 1 Large µ limit for Liénard system #03 Statement:

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

VISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague

VISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague VSUALZNG PSEUDOSPECTRA FOR POLYNOMAL EGENVALUE PROBLEMS Adéla Klimentová *, Michael Šebek ** * Department of Control Engineering Czech Technical University in Prague ** nstitute of nformation Theory and

More information

A BLOCK ALGORITHM FOR MATRIX 1-NORM ESTIMATION, WITH AN APPLICATION TO 1-NORM PSEUDOSPECTRA

A BLOCK ALGORITHM FOR MATRIX 1-NORM ESTIMATION, WITH AN APPLICATION TO 1-NORM PSEUDOSPECTRA SIAM J. MATRIX ANAL. APPL. Vol. 21, No. 4, pp. 1185 1201 c 2000 Society for Industrial and Applied Mathematics A BLOCK ALGORITHM FOR MATRIX 1-NORM ESTIMATION, WITH AN APPLICATION TO 1-NORM PSEUDOSPECTRA

More information

Family Feud Review. Linear Algebra. October 22, 2013

Family Feud Review. Linear Algebra. October 22, 2013 Review Linear Algebra October 22, 2013 Question 1 Let A and B be matrices. If AB is a 4 7 matrix, then determine the dimensions of A and B if A has 19 columns. Answer 1 Answer A is a 4 19 matrix, while

More information

Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries

Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries Fakultät für Mathematik TU Chemnitz, Germany Peter Benner benner@mathematik.tu-chemnitz.de joint work with Heike Faßbender

More information

Math 411 Preliminaries

Math 411 Preliminaries Math 411 Preliminaries Provide a list of preliminary vocabulary and concepts Preliminary Basic Netwon s method, Taylor series expansion (for single and multiple variables), Eigenvalue, Eigenvector, Vector

More information

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A

More information

Finite-choice algorithm optimization in Conjugate Gradients

Finite-choice algorithm optimization in Conjugate Gradients Finite-choice algorithm optimization in Conjugate Gradients Jack Dongarra and Victor Eijkhout January 2003 Abstract We present computational aspects of mathematically equivalent implementations of the

More information

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012. Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems

More information

NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING

NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING C. Pozrikidis University of California, San Diego New York Oxford OXFORD UNIVERSITY PRESS 1998 CONTENTS Preface ix Pseudocode Language Commands xi 1 Numerical

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

8. Diagonalization.

8. Diagonalization. 8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

Master Thesis Literature Study Presentation

Master Thesis Literature Study Presentation Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

A fast and well-conditioned spectral method: The US method

A fast and well-conditioned spectral method: The US method A fast and well-conditioned spectral method: The US method Alex Townsend University of Oxford Leslie Fox Prize, 24th of June 23 Partially based on: S. Olver & T., A fast and well-conditioned spectral method,

More information

MATH 581D FINAL EXAM Autumn December 12, 2016

MATH 581D FINAL EXAM Autumn December 12, 2016 MATH 58D FINAL EXAM Autumn 206 December 2, 206 NAME: SIGNATURE: Instructions: there are 6 problems on the final. Aim for solving 4 problems, but do as much as you can. Partial credit will be given on all

More information

Efficient and Accurate Rectangular Window Subspace Tracking

Efficient and Accurate Rectangular Window Subspace Tracking Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu

More information

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 13, 2009 Today

More information

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information