Structure preserving Krylov-subspace methods for Lyapunov equations

Similar documents
ADI-preconditioned FGMRES for solving large generalized Lyapunov equations - A case study

SyreNe System Reduction for Nanoscale IC Design

System Reduction for Nanoscale IC Design (SyreNe)

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations

Model reduction of coupled systems

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Krylov subspace methods for projected Lyapunov equations

Model Order Reduction of Electrical Circuits with Nonlinear Elements

BALANCING-RELATED MODEL REDUCTION FOR DATA-SPARSE SYSTEMS

MODEL REDUCTION BY A CROSS-GRAMIAN APPROACH FOR DATA-SPARSE SYSTEMS

Solving large Hamiltonian eigenvalue problems

Model order reduction of electrical circuits with nonlinear elements

Model reduction of nonlinear circuit equations

Algorithms that use the Arnoldi Basis

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL

Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries

Parametrische Modellreduktion mit dünnen Gittern

The Lanczos and conjugate gradient algorithms

Numerical Methods I Non-Square and Sparse Linear Systems

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

KRYLOV SUBSPACE METHODS FOR PROJECTED LYAPUNOV EQUATIONS. 1. Introduction. Consider the projected continuous-time algebraic Lyapunov equation (PCALE)

Balanced Truncation Model Reduction of Large and Sparse Generalized Linear Systems

Model Order Reduction of Continuous LTI Large Descriptor System Using LRCF-ADI and Square Root Balanced Truncation

PROJECTED GMRES AND ITS VARIANTS

The Newton-ADI Method for Large-Scale Algebraic Riccati Equations. Peter Benner.

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Efficient Implementation of Large Scale Lyapunov and Riccati Equation Solvers

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

S N. hochdimensionaler Lyapunov- und Sylvestergleichungen. Peter Benner. Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz

Scientific Computing: An Introductory Survey

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

M.A. Botchev. September 5, 2014

A POD Projection Method for Large-Scale Algebraic Riccati Equations

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University

Krylov Subspace Type Methods for Solving Projected Generalized Continuous-Time Lyapunov Equations

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

KRYLOV SUBSPACE ITERATION

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Computational Linear Algebra

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

AMS526: Numerical Analysis I (Numerical Linear Algebra)

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Course Notes: Week 1

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers

Passivity Preserving Model Reduction for Large-Scale Systems. Peter Benner.

Iterative methods for Linear System

Model Reduction for Dynamical Systems

A Continuation Approach to a Quadratic Matrix Equation

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD

FEM and sparse linear system solving

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

STA141C: Big Data & High Performance Statistical Computing

Matrix Theory, Math6304 Lecture Notes from September 27, 2012 taken by Tasadduk Chowdhury

Simple iteration procedure

Order reduction numerical methods for the algebraic Riccati equation. V. Simoncini

Matrix Equations and and Bivariate Function Approximation

arxiv: v3 [math.na] 6 Jul 2018

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Moving Frontiers in Model Reduction Using Numerical Linear Algebra

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Numerical Methods in Matrix Computations

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

STA141C: Big Data & High Performance Statistical Computing

SYSTEM-THEORETIC METHODS FOR MODEL REDUCTION OF LARGE-SCALE SYSTEMS: SIMULATION, CONTROL, AND INVERSE PROBLEMS. Peter Benner

Lecture 8: Fast Linear Solvers (Part 7)

COMP6237 Data Mining Covariance, EVD, PCA & SVD. Jonathon Hare

University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012

9.1 Preconditioned Krylov Subspace Methods

Towards high performance IRKA on hybrid CPU-GPU systems

Accelerating Model Reduction of Large Linear Systems with Graphics Processors

A Comparison of Adaptive Algebraic Multigrid and Lüscher s Inexact Deflation

6.4 Krylov Subspaces and Conjugate Gradients

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

On the Ritz values of normal matrices

On solving linear systems arising from Shishkin mesh discretizations

Numerical Methods for Solving Large Scale Eigenvalue Problems

The Conjugate Gradient Method

Block Krylov Space Solvers: a Survey

Linear and Nonlinear Matrix Equations Arising in Model Reduction

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Model Reduction for Unstable Systems

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

The Singular Value Decomposition

Max Planck Institute Magdeburg Preprints

Principal Component Analysis

Towards parametric model order reduction for nonlinear PDE systems in networks

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

Chemnitz Scientific Computing Preprints

Model reduction of large-scale dynamical systems

Arnoldi Methods in SLEPc

Numerical Solution of Matrix Equations Arising in Control of Bilinear and Stochastic Systems

Adaptive rational Krylov subspaces for large-scale dynamical systems. V. Simoncini

On Solving Large Algebraic. Riccati Matrix Equations

The rational Krylov subspace for parameter dependent systems. V. Simoncini

Transcription:

Structure preserving Krylov-subspace methods for Lyapunov equations Matthias Bollhöfer, André Eppler Institute Computational Mathematics TU Braunschweig MoRePas Workshop, Münster September 17, 2009 System Reduction for Nanoscale IC Design

2 / 26 Overview 1 Introduction 2 3 4

3 / 26 Overview Introduction 1 Introduction 2 3 4

SyreNe-Project Goals Develop and compare methods for system reduction in the design of high dimensional nanoelectronic ICs. (Integrated Circuits) Test these methods in the practice of semiconductor development. Two complementary approaches: reduction of the whole system by a global method creation of reduced order models for single devices and large linear sub-circuits 4 / 26

5 / 26 Generalized projected Lyapunov-equations EXA T + AXE T = P l BB T P T l, X = P T r XP r (1) E T YA + A T YE = P T r C T CP r, Y = P l YP T l (2) E, A R n n B, C T R n ns equations arising from the work group of T. Stykel E singular, n s n existence and uniqueness of solution proved

5 / 26 Generalized Lyapunov-equations EXA T + AXE T = B B T (1) E T YA + A T YE = C T (2) E, A R n n B, CT R n ns equations arising from the work group of T. Stykel E singular, n s n existence and uniqueness of solution proved

5 / 26 Generalized Lyapunov-equations EXA T + AXE T = B B T (1) E T YA + A T YE = C T (2) E, A R n n B, CT R n ns equations arising from the work group of T. Stykel E singular, n s n existence and uniqueness of solution proved

6 / 26 Definitions Let A := E A + A E be the Lyapunov operator and vec be the operator R n n R n n which puts the columns of a matrix as column vector X = vec(x). Rewrite the Lyapunov equations (1), (2) as linear systems AX = B with B = vec( B B T ) and C = vec( C T C). Problem: Dimension n 2 AY = C Good news: E, A are usually sparse when dealing with circuit equations.

6 / 26 Definitions Let A := E A + A E be the Lyapunov operator and vec be the operator R n n R n n which puts the columns of a matrix as column vector X = vec(x). Rewrite the Lyapunov equations (1), (2) as linear systems AX = B with B = vec( B B T ) and C = vec( C T C). Problem: Dimension n 2 AY = C Good news: E, A are usually sparse when dealing with circuit equations.

Structure preservation principle 7 / 26 right hand side is low rank and symmetric these properties transfer to the solution X of (1) iterative solver has to keep that structure in each step possible: Krylov-subspace methods, only need linear combination of vectors and applying the Lyapunov-operator use factorization X = VZV T =

Structure preservation - linear combination 8 / 26 Let X 1 = V 1 Z 1 V T 1, X 2 = V 2 Z 2 V T 2 be low rank matrices, then X 3 = α 1 X 1 + α 2 X 2 = α 1 V 1 Z 1 V1 T + α 2V 2 Z 2 V2 T ( α1 Z = (V 1 V 2 ) 1 0 }{{} 0 α 2 Z 2 V 3 has again a low rank factorization. Rank estimation rank(x 3 ) rank(x 1 ) + rank(x 2 ) } {{ } Z 3 ) (V 1 V 2 ) T }{{} V3 T

Structure preservation - apply Lyapunov operator 9 / 26 Let X = VZV T be a low rank matrix, then again X a = AX = E VZV T A T + A VZV T E T ( ) 0 Z = (EV AV ) (EV AV ) T }{{} Z 0 }{{} V a }{{} Va T Z a a low rank factorization exists. Rank estimation rank(x a ) 2 rank(x)

10 / 26 Overview Introduction 1 Introduction 2 3 4

11 / 26 ADI general Introduction properties iterative method for solving AX = B need shift parameters τ j, essential for convergence behavior, difficult to compute can be applied to solve Lyapunov equations X 0 = 0

11 / 26 ADI general Introduction properties iterative method for solving AX = B (E + τ j A)X j 1 = B B T X j 1 (E τ j A) T 2 (E + τ j A)X j = B B T X T (E τ j 1 j A) T 2 need shift parameters τ j, essential for convergence behavior, difficult to compute can be applied to solve Lyapunov equations X 0 = 0

ADI general Introduction properties iterative method for solving AX = B need shift parameters τ j, essential for convergence behavior, difficult to compute can be applied to solve Lyapunov equations E, A R n n B, CT R n ns X 0 = 0 EXA T + AXE T = B B T E T YA + A T YE = C T C 11 / 26

ADI for gen. Lyapunov 12 / 26 Standard ADI recursion (E + τ j A)X j 1 = B B T X j 1 (E τ j A) T (3) 2 (E + τ j A)X j = B B T X T (E τ j 1 j A) T (4) 2 Use (3) and (4) to obtain X j = 2τ j (E + τ j A) 1 B BT (E + τ j A) T (5) + (E + τ j A) 1 (E τ j A)X j 1 (E τ j A) T (E + τ j A) T. This symmetric sum can be factored X j = Z j Z T j.

ADI for gen. Lyapunov 12 / 26 Standard ADI recursion (E + τ j A)X j 1 = B B T X j 1 (E τ j A) T (3) 2 (E + τ j A)X j = B B T X T (E τ j 1 j A) T (4) 2 Use (3) and (4) to obtain X j = 2τ j (E + τ j A) 1 B BT (E + τ j A) T (5) + (E + τ j A) 1 (E τ j A)X j 1 (E τ j A) T (E + τ j A) T. This symmetric sum can be factored X j = Z j Z T j.

CF-ADI [Li,White 04] 13 / 26 Cholesky Factor Alternating Direct Implicit Iteration computes the Cholesky-Factor Z of the solution X = ZZ T Algorithm 1 compute shift parameters τ 1,...τ j 2 z 1 = 2τ 1 (E + τ 1 A) 1 B Z = [z 1 ] 3 For i=2..j z i = P i 1 z i 1, with 2τi+1 P i = [I (τ i+1 + τ i )(E + τ i+1 A) 1 ] 2τi Z = [Z z i ]

14 / 26 Overview Introduction 1 Introduction 2 3 4

15 / 26 Introduction iterative method for solving AX = B allows flexible preconditioning in each step calculates an orthonormal basis of m-dimensional Krylov space K m (R 0, A) = span(r 0, AR 0,..., A m 1 R 0 )

algorithm [Saad 92] Initialize: choose X 0 and dimension m Arnoldi process: 1 compute R 0 = B AX 0, h 1,0 = R 0, V 1 = R 0 h 1,0 2 for j = 1... m a) compute W j = M 1 V j b) compute V j+1 = AW j c) compute i = 1... j MGS h i,j = (V j+1, V i ), V j+1 = V j+1 h i,j V i h j+1,j = V j+1, V j+1 = V j+1 h j+1,j 3 Let W m := [W 1... W m ] Compute solution X m = X 0 + W m Y m with Y m minimizes h 1,0 e 1 H m Y Restart : If not converged X 0 = X m 16 / 26

Rank truncation strategy 1 obtain starting factorization VZV T 2 compute QR factorization of V = QR 3 compute EVD of RZR T = UΣU T 4 truncate rank to given relative tolerance (tol p, tol r ), keep only first r columns Û = U(r), ˆΣ = Σ(r) V Z V T X = V Z V T 17 / 26

Rank truncation strategy 1 obtain starting factorization VZV T 2 compute QR factorization of V = QR 3 compute EVD of RZR T = UΣU T 4 truncate rank to given relative tolerance (tol p, tol r ), keep only first r columns Û = U(r), ˆΣ = Σ(r) V Z V T X = QR Z R T Q T 17 / 26

Rank truncation strategy 1 obtain starting factorization VZV T 2 compute QR factorization of V = QR 3 compute EVD of RZR T = UΣU T 4 truncate rank to given relative tolerance (tol p, tol r ), keep only first r columns Û = U(r), ˆΣ = Σ(r) V Z V T X = Q UΣU T Q T 17 / 26

Rank truncation strategy 1 obtain starting factorization VZV T 2 compute QR factorization of V = QR 3 compute EVD of RZR T = UΣU T 4 truncate rank to given relative tolerance (tol p, tol r ), keep only first r columns Û = U(r), ˆΣ = Σ(r) V Z V T X = Q Û ˆΣ ÛT Q T 17 / 26

Rank truncation strategy 17 / 26 1 obtain starting factorization VZV T 2 compute QR factorization of V = QR 3 compute EVD of RZR T = UΣU T 4 truncate rank to given relative tolerance (tol p, tol r ), keep only first r columns Û = U(r), ˆΣ = Σ(r) ˆV ˆΣ ˆV T X = ˆV ˆΣ ˆV T Result X = ˆV ˆΣ ˆV T with ˆV = QÛ ˆV unitary matrix ˆΣ diagonal matrix

Preconditioning 18 / 26 Properties must preserve structure (important fact) may vary in each step should increase convergence rate possible candidate: CF-ADI Apply 1 cycle of CF-ADI to EW j A T + AW j E T = V j.

Preconditioning 18 / 26 Properties must preserve structure (important fact) may vary in each step should increase convergence rate possible candidate: CF-ADI Apply 1 cycle of CF-ADI to EW j A T + AW j E T = V j.

Preconditioning 18 / 26 Properties must preserve structure (important fact) may vary in each step should increase convergence rate possible candidate: CF-ADI Apply 1 cycle of CF-ADI to EW j A T + AW j E T = V j.

19 / 26 Overview Introduction Number of shifts Perturbed shifts Rank truncation 1 Introduction 2 3 4 Number of shifts Perturbed shifts Rank truncation

20 / 26 Number of shifts Introduction Number of shifts Perturbed shifts Rank truncation rank X = 18 CF-ADI more sensitive to number of shift parameters

21 / 26 Number of shifts Introduction Number of shifts Perturbed shifts Rank truncation minimum number of shifts necessary increasing the number does not result in further profit convenient range from 10 to 30 optimal number? (problem dependend) is less sensitive to specific number Conjecture A lower number of shifts is sufficient compared to pure CF-ADI.

21 / 26 Number of shifts Introduction Number of shifts Perturbed shifts Rank truncation minimum number of shifts necessary increasing the number does not result in further profit convenient range from 10 to 30 optimal number? (problem dependend) is less sensitive to specific number Conjecture A lower number of shifts is sufficient compared to pure CF-ADI.

22 / 26 Introduction Number of shifts Perturbed shifts Rank truncation Perturbed shifts τ i δ [0.99, 1.01] randomly chosen,12 shifts CF-ADI much more sensitive to perturbation of shift parameters

Rank truncation Introduction Number of shifts Perturbed shifts Rank truncation rank X = 18, tolr r = 1e 12, 12 shifts certain accuracy of tol p needed for convergence problem depended similar results for tol r 23 / 26

Summary Introduction Benefits structure preservation (low rank, symmetry) is fulfilled algorithm is less sensitive to perturbations and number of ADI shift parameters τ j Problems convergence strongly depends on CF-ADI number of columns of CF-ADI iterates increases in each step but rank only increases slightly expensive rank truncation Remedy: need rank compression or stopping criteria within CF-ADI 24 / 26

25 / 26 Outlook Introduction Possible improvements compress ranks within ADI compute faster low rank factorization improve rank truncation strategy (parameters tol p and tol r ) use different Krylov-subspace methods

25 / 26 Outlook Introduction Possible improvements compress ranks within ADI compute faster low rank factorization improve rank truncation strategy (parameters tol p and tol r ) use different Krylov-subspace methods

} { z Dipl.-Math. techn. Thomas Mach Prof. Dr. Michael Hinze Dipl.Technomath. Martin Kunkel Prof. Dr. Heike Faßbender Juan Amorocho M.Sc. Dr. Patrick Lang Dipl.-Math. Oliver Schmidt Dr. Tatjana Stykel Dr. Andreas Steinbrecher } } Dipl.-Math. techn. André Eppler TU Berlin Prof. Dr. Matthias Bollhöfer { z { z } } Dipl.-Math. techn. André Schneider { z { z TU Chemnitz } ITWM Kaiserslautern TU Braunschweig z Prof. Dr. Peter Benner Universität Hamburg TU Braunschweig { BMBF Verbundprojekt SyreNe