Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries
|
|
- Carmella Ferguson
- 5 years ago
- Views:
Transcription
1 Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries Fakultät für Mathematik TU Chemnitz, Germany Peter Benner joint work with Heike Faßbender (TU Braunschweig) and Hongguo Xu (University of Kansas) Theoretical and Computational Aspects of Matrix Algorithms Dagstuhl, October 12 17, 2003
2 Overview Overview The Hamiltonian eigenproblem Applications The symplectic Lanczos method Choosing the free parameters Numerical examples Conclusions Peter Benner TU Chemnitz 2
3 Hamiltonian eigenproblem The Hamiltonian Eigenproblem Hx = λx H R 2n 2n is a Hamiltonian matrix, i.e., (HJ) T = HJ, where J = [ 0 Hamiltonian matrices have explicit block structure implied by symmetry of HJ, H = [ A B C A T ] I n I n 0, B = B T, C = C T ; are skew-selfadjoint w.r.t. indefinite inner product x, y J := x T Jy. form Lie algebra with corresponding Lie group: S n := {S R 2n 2n SJS T = J} = symplectic matrices, hence S 1 HS is Hamiltonian for all S S 2n. ]. Peter Benner TU Chemnitz 3
4 Hamiltonian eigenproblem Spectral Properties Spectral symmetry with respect to real and imaginary axes: λ Λ (H) λ, ±λ Λ (H) Hence, λ R ±λ Λ (H), λ = jω jr ±jω Λ (H), λ C \ jr ±λ, ±λ Λ (H), Peter Benner TU Chemnitz 4
5 applications Systems and control: Applications Solution methods for algebraic and differential Riccati equations. Design of LQR/LQG/H 2 /H controllers and filters for continuous-time linear control systems. Stability radii and system norm computations; optimization of system norms. Passivity-preserving model reduction based on balancing. Reduced-order control for infinite-dim. systems based on inertial manifolds. Computational physics: exponential inegrators for Hamiltonian dynamics. Quantum chemistry: computing excitation energies in many-particle systems using random phase approximation (RPA). Quadratic eigenvalue problems... Peter Benner TU Chemnitz 5
6 applications Quadratic Eigenvalue Problems Consider λ 2 Mx + λgx + Kx = 0, where M = M T spd, K = K T, G = G T. Motivation: linear elasticity: computation of corner singularities in 3D anisotropic elastic structures [Apel/Mehrmann/Watkins 01] FE discretization in structural analysis [H.R. Schwarz 84] acoustic simulation of poro-elastic materials [Meerbergen 99] gyroscopic systems, rotor systems [Lancaster 99] Peter Benner TU Chemnitz 6
7 applications Spectral Symmetry of QEP Spectrum of matrix polynomial λ 2 Mx + λgx + Kx = 0, G = G T, looks like Hamiltonian symmetry! Peter Benner TU Chemnitz 7
8 applications Linearization of QEP Setting y = λx yields linear (generalized) eigenvalue problem ([ 0 K M 0 ] λ [ M G 0 M ]) [ y x ] = 0. H = N = [ ] 0 K M 0 [ ] M G 0 M is Hamiltonian: (HJ) T = HJ is skew-hamiltonian: (NJ) T = NJ = H λn is a Hamiltonian/skew-Hamiltonian pencil [Mehrmann/Watkins 00] Peter Benner TU Chemnitz 8
9 applications Re-formulation of H λn Note: N = Z 1 Z 2 = [ I 1 2 G 0 M ] [ M 1 2 G 0 I ]. Hence, as M is spd where Z1 1 (H λn)z 1 2 = W λi, W = [ I 1 2 G 0 I ] [ 0 K M 1 0 ] [ I 1 2 G 0 I ] W is Hamiltonian! Peter Benner TU Chemnitz 9
10 symplectic Lanczos method Observations: Structured Lanczos Method for Hamiltonian Matrices Preservation of spectral symmetry requires structured eigensolver. Hamiltonian structure is preserved under symplectic similarity transformations. For large, sparse problems want Krylov subspace method. = Need symplectic basis for Krylov subspace K(H, 2k, v) := span{v, Hv, H 2 v,..., H 2k 1 }. H nonsymmetric apply Arnoldi method or nonsymmetric Lanczos method. Arnoldi method can only produce symplectic Krylov basis in very particular instances! [Ammar/Mehrmann 91] = Need variant of nonsymmetric Lanczos! Peter Benner TU Chemnitz 10
11 symplectic Lanczos method Symplectic Lanczos Transformation For every Hamiltonian matrix H there exists symplectic similarity transformation H = S 1 HS = δ 1 β 1 ζ 2 δ 2 ζ 2 β 2 ζ 3 δ 3 ζ ζ n δ n ζ n β n ν 1 δ 1 ν 2 δ 2 ν 3 δ ν n δ n Hamiltonian J-tridiagonal (J-Hessenberg) form = Need 4n 1 parameters to represent H. Peter Benner TU Chemnitz 11
12 symplectic Lanczos method Relation of Reduced Form and Krylov Subspace Theorem: If H = S 1 HS is in Hamiltonian J-tridiagonal form, then K(H, 2n 1, v) = SR with s 1 = v is an SR decomposition of the Krylov matrix K(H, 2n 1, v) := [ v, Hv,..., H 2n 1 v ]. Peter Benner TU Chemnitz 12
13 symplectic Lanczos method Some Notation Permutation using P := [e 1, e 3,..., e 2n 1, e 2, e 4,..., e 2n ] (perfect shuffle) yields H P := P HP T, S P := P SP T, J P := P JP T = diag ([ ],..., [ ]) H P := S 1 P H P S P = δ 1 β 1 0 ζ 2 ν 1 δ ζ 2 δ 2 β 2 0 ζ ν 2 δ ζ ζ n ζ n δ n β n 0 0 ν n δ n Peter Benner TU Chemnitz 13
14 symplectic Lanczos method Derivation of Symplectic Lanczos Algorithm via J-Tridiagonalization Compute (permuted) J-tridiagonal form of a (permuted) Hamiltonian matrix H (H P ) by column-wise evaluation of HS = S H (H P S P = S P HP, where S P = [v 1, w 1, v 2, w 2,..., v n, w n ] is a permuted symplectic matrix). = H P v m = δ m v m + ν m w m ν m w m = H P v m δ m v m =: w m H P w m = ζ m v m 1 + β m v m δ m w m + ζ m+1 v m+1 ζ m+1 v m+1 = H P w m ζ m v m 1 β m v m + δ m w m =: ṽ m+1. = Choose parameters δ m, β m, ν m, ζ m such that resulting algorithm computes J P -orthogonal basis of Krylov subspace K(H P, v 1, l) = span{v 1, H P v 1,..., H 2l 1 P v 1 }. Peter Benner TU Chemnitz 14
15 symplectic Lanczos method Want S T P J P S P = J P = Choice of Parameters ν m+1 = vm+1j T P H P v m+1, β m = wmj T P H P w m,. ζ m 0, δ m are free! ζ m = ṽ m 2 = v m 2 = 1 Possible choices for δ m : δ m = 1 [B./Faßbender 97] δ m = v T mh P v m = v m w m [Lin/Ferng/Wang 97] δ m = 0 = algorithm is equivalent to HR/HZ tridiagonalization [B./Faßbender/Watkins 97, Watkins 02] Peter Benner TU Chemnitz 15
16 symplectic Lanczos method Generic Symplectic Lanczos Algorithm Choose initial vector ṽ 1 R 2n, ṽ 1 0. m = 1; v 0 = 0; ζ 1 = ṽ 1 2 ; ν 0 = 1 while ζ m 0 and ν m 1 0 v m = ṽ m /ζ m w m = H P v m δ m v m ν m = v T mj P H P v m if ν m 0 then w m = w m /ν m β m = w T mj P H P w m ṽ m+1 = H P w m ζ m v m 1 β m v m + δ m w m ζ m+1 = ṽ m+1 2 end if m = m + 1 end while Peter Benner TU Chemnitz 16
17 symplectic Lanczos method Lanczos Recursion and Breakdown H 2k P Define SP 2k = [v 1, w 1, v 2, w 2,..., v k, w k ] R 2n 2k, let principal submatrix of H P, then the symplectic Lanczos recursion is be the leading 2k 2k H P S 2k P = S 2k P 2k M P + ζ k+1 v k+1 e T 2k. Breakdown is possible: ṽ m = 0 ζ m = 0, lucky breakdown! w m = 0 ν m = 0, lucky breakdown! symplectic invariant subspace of H is detected, invariant subspace of H P of dimension 2m 1 is detected, ν m = 0 but v m 0 and w m 0 = serious breakdown! Peter Benner TU Chemnitz 17
18 choosing the free parameters Theorem: Uniqueness of Lanczos Vectors If there are two symplectic Lanczos recurrences with v 1 = v 1 HV k = V k H k + ζ k+1 ṽ k+1 e T 2k, HV k = V k H k + ζ k+1 ṽ k+1 e T 2k, then there exists a trivial symplectic matrix D k =, ˆDk, ˆF k diagonal, such that V k = V k D k. [ ˆDk ˆFk ˆD 1 k ] Proof: Consequence of uniqueness of SR decomposition up to trivial symplectic factors and relation of SR decomposition to Krylov subspace. Peter Benner TU Chemnitz 18
19 choosing the free parameters Condition of Lanczos Basis Accuracy of computed eigenvalues is affected by condition of Lanczos basis. keep cond 2 (V n ) small (i.e., choose best trivial symplectic factor D k ) V n symplectic = V 1 n = JV T n J = cond 2 (V n ) = V n 2 2. V n 2 can be bounded by max { v k 2, w k 2 } V n 2 2n max { v k 2, w k 2 }, 1 k n 1 k n = Minimize τ := max 1 k n { v k 2, w k 2 }. Peter Benner TU Chemnitz 19
20 choosing the free parameters Optimizing One Free Parameter Fix ζ k such that v k 2 = 1 for all k = τ = max { w k 2 } = max { 1 1 k n 1 k n ν k Hv k δ k v k 2 }. Requiring symplectic Lanczos basis ν k is fixed. Choice of δ k independent of δ 1,..., δ k 1 = δ k = argmin{ Hv k δv k 2 }. δ R Peter Benner TU Chemnitz 20
21 choosing the free parameters Consider the quadratic form Minimizing τ f k (δ) = Hv k δv k 2 2 = (Hv k δv k ) T (Hv k δv k ) = vk T H T Hv k 2δvk T Hv k + δ 2 vk T v }{{ k. } =1 First-order necessary condition for a minimum is Hence, f k(δ) = 2v T k Hv k + 2δ = 0. δ k = v T k Hv k, i.e., v k w k (local orthogonality condition) Same choice as in [Ferng/Lin/Wang 97]. Peter Benner TU Chemnitz 21
22 numerical examples Numerical Examples I Control of string of high-speed vehicles: n = 398, v = e, different choices for δ j 3500 Condition of Lanczos basis δ = 1 δ = 0 orthogonality condition random δ number of Lanczos steps Peter Benner TU Chemnitz 22
23 numerical examples Numerical Examples II Random stable gyroscopic system: n = 200, different choices for δ j 15 Structured eigensolver QR algorithm Computed Eigenvalues 8 Condition of Lanczos basis Im(λ) δ = 1 δ = 0 orthogonality condition random δ Re(λ) x number of Lanczos steps Peter Benner TU Chemnitz 23
24 conclusions Conclusions Optimal choice of free parameters in symplectic Lanczos method can improve condition of Lanczos basis and may improve Ritz value accuracy. Optimization of two free parameters possible but... Thorough comparison necessary. For practical applications, use implicitly restarted variants [B./Faßbender 97, Watkins 02], but still more details necessary: shift-and-invert, deflation, locking, purging,... Variants for symplectic eigenproblem available [B./Faßbender 00]. Two-sided symplectic (implicitly restarted) Arnoldi based on symplectic URV decomposition under development [B./Kreßner/Mehrmann/Xu]. Peter Benner TU Chemnitz 24
Index. for generalized eigenvalue problem, butterfly form, 211
Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,
More informationThe Newton-ADI Method for Large-Scale Algebraic Riccati Equations. Peter Benner.
The Newton-ADI Method for Large-Scale Algebraic Riccati Equations Mathematik in Industrie und Technik Fakultät für Mathematik Peter Benner benner@mathematik.tu-chemnitz.de Sonderforschungsbereich 393 S
More informationSolving large Hamiltonian eigenvalue problems
Solving large Hamiltonian eigenvalue problems David S. Watkins watkins@math.wsu.edu Department of Mathematics Washington State University Adaptivity and Beyond, Vancouver, August 2005 p.1 Some Collaborators
More informationA Structure-Preserving Method for Large Scale Eigenproblems. of Skew-Hamiltonian/Hamiltonian (SHH) Pencils
A Structure-Preserving Method for Large Scale Eigenproblems of Skew-Hamiltonian/Hamiltonian (SHH) Pencils Yangfeng Su Department of Mathematics, Fudan University Zhaojun Bai Department of Computer Science,
More informationReport Number 09/32. A Hamiltonian Krylov-Schur-type method based on the symplectic Lanczos process. Peter Benner, Heike Faßbender, Martin Stoll
Report Number 09/32 A Hamiltonian Krylov-Schur-type method based on the symplectic Lanczos process by Peter Benner, Heike Faßbender, Martin Stoll Oxford Centre for Collaborative Applied Mathematics Mathematical
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationArnoldi Methods in SLEPc
Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,
More informationThe quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying
I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationDefinite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra
Definite versus Indefinite Linear Algebra Christian Mehl Institut für Mathematik TU Berlin Germany 10th SIAM Conference on Applied Linear Algebra Monterey Bay Seaside, October 26-29, 2009 Indefinite Linear
More informationPerturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop
Perturbation theory for eigenvalues of Hermitian pencils Christian Mehl Institut für Mathematik TU Berlin, Germany 9th Elgersburg Workshop Elgersburg, March 3, 2014 joint work with Shreemayee Bora, Michael
More informationSolution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method
Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts
More informationMatrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland
Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationAPPLIED NUMERICAL LINEAR ALGEBRA
APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation
More informationNonlinear palindromic eigenvalue problems and their numerical solution
Nonlinear palindromic eigenvalue problems and their numerical solution TU Berlin DFG Research Center Institut für Mathematik MATHEON IN MEMORIAM RALPH BYERS Polynomial eigenvalue problems k P(λ) x = (
More informationCanonical forms of structured matrices and pencils
Canonical forms of structured matrices and pencils Christian Mehl and Hongguo Xu Abstract This chapter provides a survey on the development of canonical forms for matrices and matrix pencils with symmetry
More informationEfficient Wavefield Simulators Based on Krylov Model-Order Reduction Techniques
1 Efficient Wavefield Simulators Based on Krylov Model-Order Reduction Techniques From Resonators to Open Domains Rob Remis Delft University of Technology November 3, 2017 ICERM Brown University 2 Thanks
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationAn SVD-Like Matrix Decomposition and Its Applications
An SVD-Like Matrix Decomposition and Its Applications Hongguo Xu Abstract [ A matrix S C m m is symplectic if SJS = J, where J = I m I m. Symplectic matrices play an important role in the analysis and
More informationContents 1 Introduction and Preliminaries 1 Embedding of Extended Matrix Pencils 3 Hamiltonian Triangular Forms 1 4 Skew-Hamiltonian/Hamiltonian Matri
Technische Universitat Chemnitz Sonderforschungsbereich 393 Numerische Simulation auf massiv parallelen Rechnern Peter Benner Ralph Byers Volker Mehrmann Hongguo Xu Numerical Computation of Deating Subspaces
More informationKrylov subspace projection methods
I.1.(a) Krylov subspace projection methods Orthogonal projection technique : framework Let A be an n n complex matrix and K be an m-dimensional subspace of C n. An orthogonal projection technique seeks
More informationAn Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems
An Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems H. Voss 1 Introduction In this paper we consider the nonlinear eigenvalue problem T (λ)x = 0 (1) where T (λ) R n n is a family of symmetric
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationSKEW-HAMILTONIAN AND HAMILTONIAN EIGENVALUE PROBLEMS: THEORY, ALGORITHMS AND APPLICATIONS
SKEW-HAMILTONIAN AND HAMILTONIAN EIGENVALUE PROBLEMS: THEORY, ALGORITHMS AND APPLICATIONS Peter Benner Technische Universität Chemnitz Fakultät für Mathematik benner@mathematik.tu-chemnitz.de Daniel Kressner
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationLARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems
More informationSummary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method
Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,
More informationLARGE SPARSE EIGENVALUE PROBLEMS
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems
More informationCyclic reduction and index reduction/shifting for a second-order probabilistic problem
Cyclic reduction and index reduction/shifting for a second-order probabilistic problem Federico Poloni 1 Joint work with Giang T. Nguyen 2 1 U Pisa, Computer Science Dept. 2 U Adelaide, School of Math.
More informationOn Solving Large Algebraic. Riccati Matrix Equations
International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University
More informationNumerical Methods I: Eigenvalues and eigenvectors
1/25 Numerical Methods I: Eigenvalues and eigenvectors Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu November 2, 2017 Overview 2/25 Conditioning Eigenvalues and eigenvectors How hard are they
More informationKrylov Subspace Methods that Are Based on the Minimization of the Residual
Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean
More informationIterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)
Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential
More informationp q m p q p q m p q p Σ q 0 I p Σ 0 0, n 2p q
A NUMERICAL METHOD FOR COMPUTING AN SVD-LIKE DECOMPOSITION HONGGUO XU Abstract We present a numerical method to compute the SVD-like decomposition B = QDS 1 where Q is orthogonal S is symplectic and D
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More informationFinite dimensional indefinite inner product spaces and applications in Numerical Analysis
Finite dimensional indefinite inner product spaces and applications in Numerical Analysis Christian Mehl Technische Universität Berlin, Institut für Mathematik, MA 4-5, 10623 Berlin, Germany, Email: mehl@math.tu-berlin.de
More informationOn prescribing Ritz values and GMRES residual norms generated by Arnoldi processes
On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard
More informationOn the Ritz values of normal matrices
On the Ritz values of normal matrices Zvonimir Bujanović Faculty of Science Department of Mathematics University of Zagreb June 13, 2011 ApplMath11 7th Conference on Applied Mathematics and Scientific
More informationKrylov Subspaces. The order-n Krylov subspace of A generated by x is
Lab 1 Krylov Subspaces Lab Objective: matrices. Use Krylov subspaces to find eigenvalues of extremely large One of the biggest difficulties in computational linear algebra is the amount of memory needed
More informationA robust numerical method for the γ-iteration in H control
A robust numerical method for the γ-iteration in H control Peter Benner Ralph Byers Volker Mehrmann Hongguo Xu December 6, 26 Abstract We present a numerical method for the solution of the optimal H control
More informationRitz Value Bounds That Exploit Quasi-Sparsity
Banff p.1 Ritz Value Bounds That Exploit Quasi-Sparsity Ilse Ipsen North Carolina State University Banff p.2 Motivation Quantum Physics Small eigenvalues of large Hamiltonians Quasi-Sparse Eigenvector
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationApplied Linear Algebra
Applied Linear Algebra Peter J. Olver School of Mathematics University of Minnesota Minneapolis, MN 55455 olver@math.umn.edu http://www.math.umn.edu/ olver Chehrzad Shakiban Department of Mathematics University
More informationKarhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques
Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationKrylov Subspaces. Lab 1. The Arnoldi Iteration
Lab 1 Krylov Subspaces Lab Objective: Discuss simple Krylov Subspace Methods for finding eigenvalues and show some interesting applications. One of the biggest difficulties in computational linear algebra
More informationModel reduction of large-scale dynamical systems
Model reduction of large-scale dynamical systems Lecture III: Krylov approximation and rational interpolation Thanos Antoulas Rice University and Jacobs University email: aca@rice.edu URL: www.ece.rice.edu/
More informationKrylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17
Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve
More informationFinding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems
Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer
More informationThe skew-symmetric orthogonal solutions of the matrix equation AX = B
Linear Algebra and its Applications 402 (2005) 303 318 www.elsevier.com/locate/laa The skew-symmetric orthogonal solutions of the matrix equation AX = B Chunjun Meng, Xiyan Hu, Lei Zhang College of Mathematics
More informationMath 504 (Fall 2011) 1. (*) Consider the matrices
Math 504 (Fall 2011) Instructor: Emre Mengi Study Guide for Weeks 11-14 This homework concerns the following topics. Basic definitions and facts about eigenvalues and eigenvectors (Trefethen&Bau, Lecture
More informationRecent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.
Recent advances in approximation using Krylov subspaces V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The framework It is given an operator
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationEigenvalue Problems CHAPTER 1 : PRELIMINARIES
Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14
More informationEigenvalue and Eigenvector Problems
Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems
More informationLecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm
CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationFORTRAN 77 Subroutines for the Solution of Skew-Hamiltonian/Hamiltonian Eigenproblems Part I: Algorithms and Applications 1
SLICOT Working Note 2013-1 FORTRAN 77 Subroutines for the Solution of Skew-Hamiltonian/Hamiltonian Eigenproblems Part I: Algorithms and Applications 1 Peter Benner 2 Vasile Sima 3 Matthias Voigt 4 August
More informationMatrix Equations and and Bivariate Function Approximation
Matrix Equations and and Bivariate Function Approximation D. Kressner Joint work with L. Grasedyck, C. Tobler, N. Truhar. ETH Zurich, Seminar for Applied Mathematics Manchester, 17.06.2009 Sylvester matrix
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9 Minimizing Residual CG
More informationECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems
ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A
More informationA Jacobi Davidson Method for Nonlinear Eigenproblems
A Jacobi Davidson Method for Nonlinear Eigenproblems Heinrich Voss Section of Mathematics, Hamburg University of Technology, D 21071 Hamburg voss @ tu-harburg.de http://www.tu-harburg.de/mat/hp/voss Abstract.
More informationRecycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB
Recycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB Kapil Ahuja Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements
More informationMatrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein
Matrix Mathematics Theory, Facts, and Formulas with Application to Linear Systems Theory Dennis S. Bernstein PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD Contents Special Symbols xv Conventions, Notation,
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview
More informationKeeping σ fixed for several steps, iterating on µ and neglecting the remainder in the Lagrange interpolation one obtains. θ = λ j λ j 1 λ j σ, (2.
RATIONAL KRYLOV FOR NONLINEAR EIGENPROBLEMS, AN ITERATIVE PROJECTION METHOD ELIAS JARLEBRING AND HEINRICH VOSS Key words. nonlinear eigenvalue problem, rational Krylov, Arnoldi, projection method AMS subject
More informationKrylov Subspace Methods for Large/Sparse Eigenvalue Problems
Krylov Subspace Methods for Large/Sparse Eigenvalue Problems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan April 17, 2012 T.-M. Huang (Taiwan Normal University) Krylov
More informationComputation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.
Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationResearch Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics
Research Matters February 25, 2009 The Nonlinear Eigenvalue Problem Nick Higham Part III Director of Research School of Mathematics Françoise Tisseur School of Mathematics The University of Manchester
More informationAA242B: MECHANICAL VIBRATIONS
AA242B: MECHANICAL VIBRATIONS 1 / 17 AA242B: MECHANICAL VIBRATIONS Solution Methods for the Generalized Eigenvalue Problem These slides are based on the recommended textbook: M. Géradin and D. Rixen, Mechanical
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture
More informationNonlinear Eigenvalue Problems: A Challenge for Modern Eigenvalue Methods
Nonlinear Eigenvalue Problems: A Challenge for Modern Eigenvalue Methods Volker Mehrmann Heinrich Voss November 29, 2004 Abstract We discuss the state of the art in numerical solution methods for large
More informationAlgorithms that use the Arnoldi Basis
AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to
More informationA note on the product of two skew-hamiltonian matrices
A note on the product of two skew-hamiltonian matrices H. Faßbender and Kh. D. Ikramov October 13, 2007 Abstract: We show that the product C of two skew-hamiltonian matrices obeys the Stenzel conditions.
More informationMATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL
MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left
More informationSolving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners
Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Eugene Vecharynski 1 Andrew Knyazev 2 1 Department of Computer Science and Engineering University of Minnesota 2 Department
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationON THE REDUCTION OF A HAMILTONIAN MATRIX TO HAMILTONIAN SCHUR FORM
ON THE REDUCTION OF A HAMILTONIAN MATRIX TO HAMILTONIAN SCHUR FORM DAVID S WATKINS Abstract Recently Chu Liu and Mehrmann developed an O(n 3 ) structure preserving method for computing the Hamiltonian
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationContents 1 Introduction 1 Preliminaries Singly structured matrices Doubly structured matrices 9.1 Matrices that are H-selfadjoint and G-selfadjoint...
Technische Universitat Chemnitz Sonderforschungsbereich 9 Numerische Simulation auf massiv parallelen Rechnern Christian Mehl, Volker Mehrmann, Hongguo Xu Canonical forms for doubly structured matrices
More informationABSTRACT OF DISSERTATION. Ping Zhang
ABSTRACT OF DISSERTATION Ping Zhang The Graduate School University of Kentucky 2009 Iterative Methods for Computing Eigenvalues and Exponentials of Large Matrices ABSTRACT OF DISSERTATION A dissertation
More informationGMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems
GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems Tsung-Ming Huang Department of Mathematics National Taiwan Normal University December 4, 2011 T.-M. Huang (NTNU) GMRES
More informationA refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices
A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices Jean Christophe Tremblay and Tucker Carrington Chemistry Department Queen s University 23 août 2007 We want to
More informationA Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations
A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations Peter Benner Max-Planck-Institute for Dynamics of Complex Technical Systems Computational Methods in Systems and Control Theory
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationPassivity Preserving Model Reduction for Large-Scale Systems. Peter Benner.
Passivity Preserving Model Reduction for Large-Scale Systems Peter Benner Mathematik in Industrie und Technik Fakultät für Mathematik Sonderforschungsbereich 393 S N benner@mathematik.tu-chemnitz.de SIMULATION
More information1 Conjugate gradients
Notes for 2016-11-18 1 Conjugate gradients We now turn to the method of conjugate gradients (CG), perhaps the best known of the Krylov subspace solvers. The CG iteration can be characterized as the iteration
More informationThe German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.
Chapter 4 EIGENVALUE PROBLEM The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English. 4.1 Mathematics 4.2 Reduction to Upper Hessenberg
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationLinear Algebra 2 Spectral Notes
Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More information