High-performance computation of pseudospectra
|
|
- Lindsey Annabelle Davis
- 5 years ago
- Views:
Transcription
1 1 / 80 High-performance computation of pseudospectra Jack Poulson Department of Mathematics Stanford University Innovative Computing Laboratory The University of Tennessee October 24, 2014
2 A note on collaborators 2 / 80 Multishift Hessenberg solves: collaboration with Greg Henry
3 3 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
4 4 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
5 5 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
6 6 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
7 7 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
8 8 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
9 9 / 80 Overview of talk Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
10 Outline 10 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
11 11 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot
12 12 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot
13 13 / 80 Van Loan s algorithm and the Demmel matrix Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot
14 Van Loan s algorithm and the Demmel matrix 14 / 80 Efficient algorithm for evaluating resolvent over vertical line in complex plane [Van Loan-1985] Relationship to nearest stable matrix based on false conjecture, counter-example given in [Demmel-1987] Counter-example was first pseudospectral (computer) plot D(n, β) = (β J β 1,n ) 1, n = 3, β = 100
15 Outline 15 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
16 Pseudospectra 16 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential
17 Pseudospectra 17 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential
18 Pseudospectra 18 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential
19 Pseudospectra 19 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential
20 Pseudospectra 20 / 80 With convention (λi A) 1 p = for λ L(A), eigenvalues are singularities of resolvent norm (ξi A) 1 p. Natural generalization of spectrum for each p-norm and ɛ > 0: { L p ɛ (A) = ξ C : (ξi A) 1 p > 1 } ɛ Reinvented many times [Varah-1967,Landau-1975,Godunov et al.-1982,trefethen-1990,hinrichsen/pritchard-1992,...] Extensive review of field in Trefethen and Embree s book Spectra and pseudospectra... Most common tool is EigTool [Wright et al.-2001], but level 2 BLAS (trsv) and sequential
21 Outline 21 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
22 22 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
23 23 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
24 24 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
25 25 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
26 26 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
27 27 / 80 Relationship to previous work Embarrassingly parallel Lanczos [Braconnier-1996] Parallel sparse-direct shift-and-invert [Fraysse et al.-1996] Parallel path-following methods, e.g., [Mezher-2001,Bekas et al.-2000,2001,...] Our focus on level 3 (parallel) batch evaluation over point clouds after parallel Schur decomposition Key idea: simultaneously drive many Lanczos/Arnoldi methods via multishift extensions of standard TRSM algorithm Software for one and two-norm pseudospectra already released in Elemental [P. et al.-2013]
28 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 28 / 80
29 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 29 / 80
30 (Extended) Van Loan algorithm [Van Loan-1985,Lui-1997] Given a reduction to condensed form, A = QGQ H, { L p ɛ (A) = ξ C : Q(ξI G) 1 Q H p > 1 }, ɛ and, when p = 2, L 2 ɛ (A) = { ξ C : (ξi G) 1 2 > 1 }, ɛ First reduce to (quasi-)triangular/hessenberg form, then, (Restarted) Lanczos/Arnoldi for each (ξi G) 1 2, or Blocked Hager [Higham/Tisseur-2000] for each Q(ξI G) 1 Q H 1. Note that this algorithm makes less sense when (ξi A) 1 can already be applied in quadratic time (e.g., if A is 3D FEM discretization) 30 / 80
31 (Extended) Van Loan algorithm for L 2 ɛ (A) 31 / 80 Algorithm: Two-norm pseudospectra via extended Van Loan algorithm Input: A F n n, shifts Ω C, restart size k Output: {φ(ξ)} ξ Ω { (ξi A) 1 2 } ξ Ω G := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (ξi G) 1 2 via Restarted Arnoldi Choose v 0 C n with v 0 2 = 1 while not converged do for j = 0,..., k 1 do // (ξi G) H (ξi G) 1 V j = V j H j + v j (β j e j ) H x j := (ξi G) H (ξi G) 1 v j Expand Arnoldi decomposition using x j [λ, v 0 ] := MaxEig(H k ) φ(ξ) := RealPart(λ)
32 [Hager-1984,Higham-1988] approach for L 1 ɛ (A) 32 / 80 Algorithm: One-norm pseudospectra via Hager-Higham algorithm Input: A F n n, Ω C Output: {φ(ξ)} ξ Ω { (A ξi) 1 1 } ξ Ω [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (A ξi) 1 1 via Hager-Higham algorithm Choose u C n with u 1 = 1 k := 0 repeat x := u y := Q(G ξi) 1 Q H x w := Q(G ξi) H Q H sign(y) u := e j where w(j) = w k := k + 1 until (k 2 and w w H x) or k 5 φ(ξ) := y 1 // Take the maximum between the current estimate and a heuristic b := 2 3n [( 1)j (n + j 1)] j=0:n 1 x := Q(G ξi) 1 Q H b φ(ξ) := max(φ(ξ), x 1 ) Block algorithm [Higham/Tisseur-2000] should be used in practice
33 [Hager-1984,Higham-1988] approach for L 1 ɛ (A) 33 / 80 Algorithm: One-norm pseudospectra via Hager-Higham algorithm Input: A F n n, Ω C Output: {φ(ξ)} ξ Ω { (A ξi) 1 1 } ξ Ω [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) foreach ξ Ω do // Estimate (A ξi) 1 1 via Hager-Higham algorithm Choose u C n with u 1 = 1 k := 0 repeat x := u y := Q(G ξi) 1 Q H x w := Q(G ξi) H Q H sign(y) u := e j where w(j) = w k := k + 1 until (k 2 and w w H x) or k 5 φ(ξ) := y 1 // Take the maximum between the current estimate and a heuristic b := 2 3n [( 1)j (n + j 1)] j=0:n 1 x := Q(G ξi) 1 Q H b φ(ξ) := max(φ(ξ), x 1 ) Block algorithm [Higham/Tisseur-2000] should be used in practice
34 Outline 34 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
35 35 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
36 36 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
37 37 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
38 38 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
39 39 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
40 40 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
41 41 / 80 (Generalized) multishift solves Computing {(δ j F G) 1 y j } j equivalent to solving for X in where D = diag((δ 0, δ 1, )). FXD GX = Y, Multishift triangular solves (G triangular, F = I) require trivial changes to usual high-performance TRSM algorithms [Henry-1994] Quasi-triangular similar; adjust blocksizes to not split 2 2 Very recently used by [Gates/Haidar/Dongarra-2014] for eigenvectors of nonsymmetric matrices Generalized multishift (quasi-)triangular solves can similarly be made high-performance with a simple trick... Multishift Hessenberg solves heavily investigated for control theory [Datta et al-1994,henry-1994] Large fraction of work in Hessenberg case is level 1...
42 Interleaved Van Loan algorithm for L 2 ɛ (A) 42 / 80 Algorithm: Two-norm pseudospectra via interleaved extended Van Loan algorithm Input: A F n n, shift vector z C m, restart size k Output: f [ (z(s)i A) 1 2 ]s=0:m 1 G := Schur(A), RealSchur(A), or Hessenberg(A) Initialize each column of W 0 C n m to have unit two-norm I := (0,..., m 1) while I do for j = 0,..., k 1 do // (z(s)i G) H (z(s)i G) 1 W j (t) = W j (t)h j (t) + W j (:, t)(b j (t)e j ) H, s = I(t) X := MultishiftSolve(G, z(i), W j ) X := MultishiftSolve(G H, z(i), X) Expand Arnoldi decompositions foreach s = I(t) do [λ, W 0 (:, t)] := MaxEig(H k (t)) f (s) := RealPart(λ) if converged then Delete I(t) and W 0 (:, t)
43 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 43 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 [Q, G] := Schur(A), RealSchur(A), or Hessenberg(A) Initialize each column of U C n m to have unit one-norm I := (0,..., m 1) k := 0 while I do X := U Y := Q H X Y := MultishiftSolve(G, z(i), Y ) Y := QY W := Q H sign(y ) W := MultishiftSolve(G H, z(i), W ) W := QW k := k + 1 foreach s = I(t) do x := X(:, t), y := Y (:, t), w := W (:, t) if (k 2 and w w H x) or k 5 then f (s) = y 1 Delete I(t) and U(:, t) else U(:, t) := e j where w(j) = w
44 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 44 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 // Take the maximum between the current estimates and a heuristic B := 2 3n [( 1)j (n + j 1)] j=0:n 1,s=0:m 1 // All columns are equal X := Q H B // All columns are equal X := MultishiftSolve(G, z, X) X := QX foreach s = 0,..., m 1 do f (s) := max(f (s), X(:, s) 1 ) Again: block algorithm from [Higham/Tisseur-2000] should be used in practice
45 Interleaved Hager-Higham algorithm for L 1 ɛ (A) 45 / 80 Algorithm: One-norm pseudospectra via interleaved Hager-Higham algorithm Input: A F n n, shift vector z C m Output: f [ (z(j)i A) 1 1 ]j=0:m 1 // Take the maximum between the current estimates and a heuristic B := 2 3n [( 1)j (n + j 1)] j=0:n 1,s=0:m 1 // All columns are equal X := Q H B // All columns are equal X := MultishiftSolve(G, z, X) X := QX foreach s = 0,..., m 1 do f (s) := max(f (s), X(:, s) 1 ) Again: block algorithm from [Higham/Tisseur-2000] should be used in practice
46 46 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...
47 47 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...
48 48 / 80 Batching and deflation Maximum number of simultaneous shifts constrained by memory Could bring in new shift after each deflation, but easier to break into batches In practice, only small number of iterations needed per shift...
49 Outline 49 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
50 Investigating the Fox-Li matrix with python 50 / 80 import math, El n = 100 # m a t r i x s ize realres = imagres = 100 # g r i d r e s o l u t i o n # Display an instance of the Fox L i / Landau matrix A = El. D i s t M a t r i x ( El. ztag ) El. FoxLi (A, n ) El. Display (A, " Fox L i m a t r i x " ) # Display i t s s p e c t r a l p o r t r a i t p o r t r a i t = El. S p e c t r a l P o r t r a i t (A, realres, imagres ) El. EntrywiseMap ( p o r t r a i t, math. log10 ) El. Display ( p o r t r a i t, " S p e c t r a l p o r t r a i t of Fox L i matrix " ) # Display i t s s i n g u l a r values s = El.SVD(A) El. EntrywiseMap ( s, math. log10 ) El. Display ( s, " log10 ( svd (A ) ) " )
51 Investigating the Fox-Li matrix with python 51 / 80
52 Investigating the Fox-Li matrix with python 52 / 80
53 Investigating the Fox-Li matrix with python 53 / 80
54 Outline 54 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
55 55 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 10 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)
56 56 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 20 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)
57 57 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 30 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)
58 58 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 40 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)
59 59 / 80 FoxLi(15k), Ω = ( 1.2, 1.2) 2, pixels, 50 its (Au)(x) = if /π 1 1 e if (x y)2 u(y) dy, F = 16π 30 sec/iter on 256 cores of Stampede (256 r.h.s./core, >4 TFlops)
60 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 10 its 60 / 80
61 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 20 its 61 / 80
62 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 30 its 62 / 80
63 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 40 its 63 / 80
64 FoxLi(15k) work, Ω = ( 1.2, 1.2) 2, pix, 50 its 64 / 80
65 65 / 80 Uniform(20k), Ω = ( 103, 103) 2, pix pieces: 75 sec/iter on 256 nodes of Blue Gene/Q, (64 r.h.s./core, >11 TFlops), 1175 sec for Schur
66 66 / 80 Uniform(20k), Ω = ( 10.3, 10.3) 2, pix pieces: 75 sec/iter on 256 nodes of Blue Gene/Q, (64 r.h.s./core, >11 TFlops), 1175 sec for Schur
67 Outline 67 / 80 Van Loan s algorithm and the Demmel matrix Brief overview of pseudospectra Previous work High-performance batched analogues A brief example of the python interface Results Conclusions and future work
68 68 / 80 Convergence issues Typically small number of iterations required for visual convergence. Given current batch iteration, can output image after each fixed number of iterations Problematic pixels typically uninteresting (resolvent gradient is small)
69 69 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
70 70 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
71 71 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
72 72 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
73 73 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
74 74 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
75 75 / 80 Future directions Interleaved block one-norm pseudospectral algorithm More intelligent interpolation and stopping criteria Investigate (preconditioned) 2D/3D wave equations Better understanding of practical limits of SDC EVD Optional projection onto relevant eigenspaces Complex distributed Hessenberg QR/QZ with AED... Black-box high-performance generalized pseudospectra
76 Acknowledgments and Availability 76 / 80 Support Computational resources My hosts Jakub Kurzak and Jack Dongarra Availability Elemental is available under the New BSD License at libelemental.org Questions?
77 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 77 / 80
78 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 78 / 80
79 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 79 / 80
80 Generalized multishift (quasi-)trsm FXD GX = Y As long as a 2x2 block is not split: ( ) (F1,1 X 1 D G 1,1 X 1 ) + (F 1,2 X 2 D G 1,2 X 2 ) = F 2,2 X 2 D G 2,2 X 2 ( Y1 If F 2,2 and G 2,2 are small and square, have each process locally solve for a few right-hand sides of X 2. Then, form Ŷ 1 := Y 1 (F 1,2 X 2 D G 1,2 X 2 ) via a parallel GEMM with each column of F 1,2 X 2 appropriately scaled afterwards. All that is left is to recurse on F 1,1 X 1 D G 1,1 X 1 = Ŷ1. Y 2 ) When F = I, algorithm is much simpler. 80 / 80
APPLIED NUMERICAL LINEAR ALGEBRA
APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationAccelerating computation of eigenvectors in the nonsymmetric eigenvalue problem
Accelerating computation of eigenvectors in the nonsymmetric eigenvalue problem Mark Gates 1, Azzam Haidar 1, and Jack Dongarra 1,2,3 1 University of Tennessee, Knoxville, TN, USA 2 Oak Ridge National
More informationAccelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem
Accelerating computation of eigenvectors in the dense nonsymmetric eigenvalue problem Mark Gates 1, Azzam Haidar 1, and Jack Dongarra 1,2,3 1 University of Tennessee, Knoxville, TN, USA 2 Oak Ridge National
More informationArnoldi Methods in SLEPc
Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationMatrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland
Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More information1 Sylvester equations
1 Sylvester equations Notes for 2016-11-02 The Sylvester equation (or the special case of the Lyapunov equation) is a matrix equation of the form AX + XB = C where A R m m, B R n n, B R m n, are known,
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationContents. Preface... xi. Introduction...
Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationLARGE SPARSE EIGENVALUE PROBLEMS
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationNumerical linear algebra
Numerical linear algebra Purdue University CS 51500 Fall 2017 David Gleich David F. Gleich Call me Prof Gleich Dr. Gleich Please not Hey matrix guy! Huda Nassar Call me Huda Ms. Huda Please not Matrix
More informationLARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems
More informationIndex. for generalized eigenvalue problem, butterfly form, 211
Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,
More informationRemark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationSOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA
1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization
More informationIterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)
Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential
More informationOUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact
Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu
More informationARPACK. Dick Kachuma & Alex Prideaux. November 3, Oxford University Computing Laboratory
ARPACK Dick Kachuma & Alex Prideaux Oxford University Computing Laboratory November 3, 2006 What is ARPACK? ARnoldi PACKage Collection of routines to solve large scale eigenvalue problems Developed at
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationRemark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationLecture 2: Numerical linear algebra
Lecture 2: Numerical linear algebra QR factorization Eigenvalue decomposition Singular value decomposition Conditioning of a problem Floating point arithmetic and stability of an algorithm Linear algebra
More informationPseudospectra and Nonnormal Dynamical Systems
Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 2012 Overview of the Course These lectures
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationHessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods
Hessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods Jens Peter M. Zemke Minisymposium on Numerical Linear Algebra Technical
More informationComputing Unstructured and Structured Polynomial Pseudospectrum Approximations
Computing Unstructured and Structured Polynomial Pseudospectrum Approximations Silvia Noschese 1 and Lothar Reichel 2 1 Dipartimento di Matematica, SAPIENZA Università di Roma, P.le Aldo Moro 5, 185 Roma,
More informationQR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR
QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique
More informationDefinition (T -invariant subspace) Example. Example
Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin
More informationKrylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17
Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve
More informationEigenvalue Problems and Singular Value Decomposition
Eigenvalue Problems and Singular Value Decomposition Sanzheng Qiao Department of Computing and Software McMaster University August, 2012 Outline 1 Eigenvalue Problems 2 Singular Value Decomposition 3 Software
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationLeast squares and Eigenvalues
Lab 1 Least squares and Eigenvalues Lab Objective: Use least squares to fit curves to data and use QR decomposition to find eigenvalues. Least Squares A linear system Ax = b is overdetermined if it has
More informationAn Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84
An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationON MATRIX BALANCING AND EIGENVECTOR COMPUTATION
ON MATRIX BALANCING AND EIGENVECTOR COMPUTATION RODNEY JAMES, JULIEN LANGOU, AND BRADLEY R. LOWERY arxiv:40.5766v [math.na] Jan 04 Abstract. Balancing a matrix is a preprocessing step while solving the
More informationThe quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying
I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given
More informationLecture on: Numerical sparse linear algebra and interpolation spaces. June 3, 2014
Lecture on: Numerical sparse linear algebra and interpolation spaces June 3, 2014 Finite dimensional Hilbert spaces and IR N 2 / 38 (, ) : H H IR scalar product and u H = (u, u) u H norm. Finite dimensional
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationMain matrix factorizations
Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined
More informationKrylov Subspaces. Lab 1. The Arnoldi Iteration
Lab 1 Krylov Subspaces Lab Objective: Discuss simple Krylov Subspace Methods for finding eigenvalues and show some interesting applications. One of the biggest difficulties in computational linear algebra
More informationc 2009 Society for Industrial and Applied Mathematics
SIAM J. MATRIX ANAL. APPL. Vol. 3, No., pp. c 29 Society for Industrial and Applied Mathematics THE ARNOLDI EIGENVALUE ITERATION WITH EXACT SHIFTS CAN FAIL MARK EMBREE Abstract. The restarted Arnoldi algorithm,
More informationA HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY
A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY RONALD B. MORGAN AND MIN ZENG Abstract. A restarted Arnoldi algorithm is given that computes eigenvalues
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More informationNumerical Mathematics
Alfio Quarteroni Riccardo Sacco Fausto Saleri Numerical Mathematics Second Edition With 135 Figures and 45 Tables 421 Springer Contents Part I Getting Started 1 Foundations of Matrix Analysis 3 1.1 Vector
More informationMath 3191 Applied Linear Algebra
Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful
More informationPERFORMANCE ENHANCEMENT OF PARALLEL MULTIFRONTAL SOLVER ON BLOCK LANCZOS METHOD 1. INTRODUCTION
J. KSIAM Vol., No., -0, 009 PERFORMANCE ENHANCEMENT OF PARALLEL MULTIFRONTAL SOLVER ON BLOCK LANCZOS METHOD Wanil BYUN AND Seung Jo KIM, SCHOOL OF MECHANICAL AND AEROSPACE ENG, SEOUL NATIONAL UNIV, SOUTH
More informationOrthogonal iteration to QR
Notes for 2016-03-09 Orthogonal iteration to QR The QR iteration is the workhorse for solving the nonsymmetric eigenvalue problem. Unfortunately, while the iteration itself is simple to write, the derivation
More informationAlgorithms that use the Arnoldi Basis
AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to
More informationSection 4.5 Eigenvalues of Symmetric Tridiagonal Matrices
Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete
More informationReal Eigenvalue Extraction and the Distance to Uncontrollability
Real Eigenvalue Extraction and the Distance to Uncontrollability Emre Mengi Computer Science Department Courant Institute of Mathematical Sciences New York University mengi@cs.nyu.edu May 22nd, 2006 Emre
More information7. Symmetric Matrices and Quadratic Forms
Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value
More informationHere is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J
Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationMatrix Condition Estimators in 2-norm
Matrix Condition Estimators in 2-norm Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Miroslav Tůma Institute of Computer Science
More informationSome minimization problems
Week 13: Wednesday, Nov 14 Some minimization problems Last time, we sketched the following two-step strategy for approximating the solution to linear systems via Krylov subspaces: 1. Build a sequence of
More informationPROJECTED GMRES AND ITS VARIANTS
PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft
More informationEigenvalue problems. Eigenvalue problems
Determination of eigenvalues and eigenvectors Ax x, where A is an N N matrix, eigenvector x 0, and eigenvalues are in general complex numbers In physics: - Energy eigenvalues in a quantum mechanical system
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationResearch Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics
Research Matters February 25, 2009 The Nonlinear Eigenvalue Problem Nick Higham Part III Director of Research School of Mathematics Françoise Tisseur School of Mathematics The University of Manchester
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationAccelerating linear algebra computations with hybrid GPU-multicore systems.
Accelerating linear algebra computations with hybrid GPU-multicore systems. Marc Baboulin INRIA/Université Paris-Sud joint work with Jack Dongarra (University of Tennessee and Oak Ridge National Laboratory)
More informationSingular Value Decomposition of the frequency response operator
Measures of input-output amplification (across frequency) Largest singular value Hilbert-Schmidt norm (power spectral density) Systems with one spatial variable Two point boundary value problems EE 8235:
More informationMatrix stabilization using differential equations.
Matrix stabilization using differential equations. Nicola Guglielmi Universitá dell Aquila and Gran Sasso Science Institute, Italia NUMOC-2017 Roma, 19 23 June, 2017 Inspired by a joint work with Christian
More informationOn Solving Large Algebraic. Riccati Matrix Equations
International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University
More informationProblem Set Number 5, j/2.036j MIT (Fall 2014)
Problem Set Number 5, 18.385j/2.036j MIT (Fall 2014) Rodolfo R. Rosales (MIT, Math. Dept.,Cambridge, MA 02139) Due Fri., October 24, 2014. October 17, 2014 1 Large µ limit for Liénard system #03 Statement:
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationVISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague
VSUALZNG PSEUDOSPECTRA FOR POLYNOMAL EGENVALUE PROBLEMS Adéla Klimentová *, Michael Šebek ** * Department of Control Engineering Czech Technical University in Prague ** nstitute of nformation Theory and
More informationA BLOCK ALGORITHM FOR MATRIX 1-NORM ESTIMATION, WITH AN APPLICATION TO 1-NORM PSEUDOSPECTRA
SIAM J. MATRIX ANAL. APPL. Vol. 21, No. 4, pp. 1185 1201 c 2000 Society for Industrial and Applied Mathematics A BLOCK ALGORITHM FOR MATRIX 1-NORM ESTIMATION, WITH AN APPLICATION TO 1-NORM PSEUDOSPECTRA
More informationFamily Feud Review. Linear Algebra. October 22, 2013
Review Linear Algebra October 22, 2013 Question 1 Let A and B be matrices. If AB is a 4 7 matrix, then determine the dimensions of A and B if A has 19 columns. Answer 1 Answer A is a 4 19 matrix, while
More informationStructured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries
Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries Fakultät für Mathematik TU Chemnitz, Germany Peter Benner benner@mathematik.tu-chemnitz.de joint work with Heike Faßbender
More informationMath 411 Preliminaries
Math 411 Preliminaries Provide a list of preliminary vocabulary and concepts Preliminary Basic Netwon s method, Taylor series expansion (for single and multiple variables), Eigenvalue, Eigenvector, Vector
More informationECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems
ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A
More informationFinite-choice algorithm optimization in Conjugate Gradients
Finite-choice algorithm optimization in Conjugate Gradients Jack Dongarra and Victor Eijkhout January 2003 Abstract We present computational aspects of mathematically equivalent implementations of the
More informationMath Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.
Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems
More informationNUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING
NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING C. Pozrikidis University of California, San Diego New York Oxford OXFORD UNIVERSITY PRESS 1998 CONTENTS Preface ix Pseudocode Language Commands xi 1 Numerical
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationNumerical optimization
Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal
More informationNumerical Methods for Solving Large Scale Eigenvalue Problems
Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems
More informationMaster Thesis Literature Study Presentation
Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview
More informationMATH 221, Spring Homework 10 Solutions
MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the
More informationA fast and well-conditioned spectral method: The US method
A fast and well-conditioned spectral method: The US method Alex Townsend University of Oxford Leslie Fox Prize, 24th of June 23 Partially based on: S. Olver & T., A fast and well-conditioned spectral method,
More informationMATH 581D FINAL EXAM Autumn December 12, 2016
MATH 58D FINAL EXAM Autumn 206 December 2, 206 NAME: SIGNATURE: Instructions: there are 6 problems on the final. Aim for solving 4 problems, but do as much as you can. Partial credit will be given on all
More informationEfficient and Accurate Rectangular Window Subspace Tracking
Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu
More informationComputing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm
Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 13, 2009 Today
More informationEIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems
EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues
More information8 Numerical methods for unconstrained problems
8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields
More information