Model reduction of large-scale systems by least squares

Similar documents
An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

Model reduction of large-scale dynamical systems

Model Reduction for Unstable Systems

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

Rational Krylov Methods for Model Reduction of Large-scale Dynamical Systems

Iterative Rational Krylov Algorithm for Unstable Dynamical Systems and Generalized Coprime Factorizations

Krylov Techniques for Model Reduction of Second-Order Systems

MODEL REDUCTION BY A CROSS-GRAMIAN APPROACH FOR DATA-SPARSE SYSTEMS

CME 345: MODEL REDUCTION

Model reduction for linear systems by balancing

Fluid flow dynamical model approximation and control

Balanced Truncation 1

BALANCING-RELATED MODEL REDUCTION FOR DATA-SPARSE SYSTEMS

Comparison of Model Reduction Methods with Applications to Circuit Simulation

Model reduction of interconnected systems

Model Reduction of Inhomogeneous Initial Conditions

Model reduction via tangential interpolation

Chris Beattie and Serkan Gugercin

Optimal Approximation of Dynamical Systems with Rational Krylov Methods. Department of Mathematics, Virginia Tech, USA

AN OVERVIEW OF MODEL REDUCTION TECHNIQUES APPLIED TO LARGE-SCALE STRUCTURAL DYNAMICS AND CONTROL MOTIVATING EXAMPLE INVERTED PENDULUM

S. Gugercin and A.C. Antoulas Department of Electrical and Computer Engineering Rice University

An Interpolation-Based Approach to the Optimal H Model Reduction

Institut für Mathematik

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Inexact Solves in Krylov-based Model Reduction

Krylov-Subspace Based Model Reduction of Nonlinear Circuit Models Using Bilinear and Quadratic-Linear Approximations

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

H 2 optimal model reduction - Wilson s conditions for the cross-gramian

3 Gramians and Balanced Realizations

FEL3210 Multivariable Feedback Control

Passivity Preserving Model Reduction for Large-Scale Systems. Peter Benner.

Chapter 6 Balanced Realization 6. Introduction One popular approach for obtaining a minimal realization is known as Balanced Realization. In this appr

Approximation of the Linearized Boussinesq Equations

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Passive Interconnect Macromodeling Via Balanced Truncation of Linear Systems in Descriptor Form

Model reduction of coupled systems

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

H 2 -optimal model reduction of MIMO systems

Numerical Methods I Non-Square and Sparse Linear Systems

Lecture notes: Applied linear algebra Part 1. Version 2

Model Order Reduction of Continuous LTI Large Descriptor System Using LRCF-ADI and Square Root Balanced Truncation

MODEL-order reduction is emerging as an effective

Algebra C Numerical Linear Algebra Sample Exam Problems

Canonical lossless state-space systems: staircase forms and the Schur algorithm

A modified low-rank Smith method for large-scale Lyapunov equations

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Order Reduction for Large Scale Finite Element Models: a Systems Perspective

Krylov subspace projection methods

Model Reduction for Linear Dynamical Systems

Hankel Optimal Model Reduction 1

Introduction to Iterative Solvers of Linear Systems

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Numerical Methods - Numerical Linear Algebra

Iterative Methods for Solving A x = b

MIT Final Exam Solutions, Spring 2017

Zeros and zero dynamics

BALANCING AS A MOMENT MATCHING PROBLEM

Robust Multivariable Control

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77

Subspace-based Identification

Gramians of structured systems and an error bound for structure-preserving model reduction

Control Systems. Laplace domain analysis

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Lecture Note #6 (Chap.10)

A comparison of model reduction techniques from structural dynamics, numerical mathematics and systems and control

Gramians based model reduction for hybrid switched systems

Chapter 6 Balanced Realization 6. Introduction One popular approach for obtaining a minimal realization is known as Balanced Realization. In this appr

Solving Large Scale Lyapunov Equations

Numerical Methods in Matrix Computations

Lecture Note #6 (Chap.10)

Total least squares. Gérard MEURANT. October, 2008

Stability preserving post-processing methods applied in the Loewner framework

There are six more problems on the next two pages

APPLIED LINEAR ALGEBRA

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Adaptive rational Krylov subspaces for large-scale dynamical systems. V. Simoncini

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Technische Universität Berlin

arxiv: v5 [math.na] 16 Nov 2017

Control Systems Design

Properties of Matrices and Operations on Matrices

State will have dimension 5. One possible choice is given by y and its derivatives up to y (4)

Computational Methods. Eigenvalues and Singular Values

Introduction. Chapter One

Empirical Gramians and Balanced Truncation for Model Reduction of Nonlinear Systems

Index. for generalized eigenvalue problem, butterfly form, 211

arxiv: v3 [math.na] 6 Jul 2018

5.3 The Upper Half Plane

DESIGN OF OBSERVERS FOR SYSTEMS WITH SLOW AND FAST MODES

Analysis of Discrete-Time Systems

Arnoldi Methods in SLEPc

LinGloss. A glossary of linear algebra

6.241 Dynamic Systems and Control

Model reduction of large-scale systems

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

DELFT UNIVERSITY OF TECHNOLOGY

Transcription:

Model reduction of large-scale systems by least squares Serkan Gugercin Department of Mathematics, Virginia Tech, Blacksburg, VA, USA gugercin@mathvtedu Athanasios C Antoulas Department of Electrical and Computer Engineering Rice University, Houston, TX, USA aca@ecericeedu Abstract In this paper we introduce an approximation method for model reduction of large-scale dynamical systems This is a projection which combines aspects of the SVD and Krylov based reduction methods This projection can be efficiently computed using tools from numerical analysis, namely the rational Krylov method for the Krylov side of the projection and a low-rank Smith type iteration to solve a Lyapunov equation for the SVD side of the projection For discrete time systems, the proposed approach is based on the least squares fit of the (r + 1) st column of a Hankel matrix to the preceding r columns, where r is the order of the reduced system The reduced system is asymptotically stable, matches the first r Markov parameters of the full order model and minimizes a weighted H 2 error The method is also generalized for moment matching at arbitrary interpolation points Application to continuous time systems is achieved via the bilinear transformation Numerical examples prove the effectiveness of the approach The proposed method is significant because it combines guaranteed stability and moment matching, together with an optimization criterion 1 Introduction A large number of physical phenomena, such as electric circuits, heat transfer through various media, wave propagation, vibration suppression of bridges, and flexible beams, are modeled by linear dynamical systems Direct numerical simulation of these mathematical models has been a successful means for studying the underlying complex physical phenomena In the sequel we will be interested in linear time invariant (LTI) systems which are described in state space form: [ A b Σ : { σx(t) = Ax(t) + bu(t) y(t) = cx(t) + du(t) Σ := c d ], (11) where A R n n, b R n, c T R n, d R, and σ denotes the shift or the derivative operator depending on whether we are dealing with discrete or continuous time systems In (11), x(t) R n is the state, u(t) R is the input, and y(t) R is the output of Σ The transfer function G(s) of Σ is given by G(s) = c(si A) 1 b + d Although the results apply equally to multi-input-multi-output (MIMO) systems, for simplicity we will consider only single-input-single-output (SISO) systems In many applications, such as circuit simulation, or time dependent PDE control problems (see [5] for more examples), n is quite large In these large-scale settings, the system dimension makes the computations infeasible due to memory, time limitations and ill-conditioning, which is the basic motivation for model reduction The goal of model reduction is, then, to produce a lower dimensional system that has approximately the same response characteristics as the original system, with reduced storage requirements and evaluation time The resulting reduced model might be used to replace the original system in a simulation or to develop a low dimensional controller suitable for real time applications This work was supported in part by the NSF through Grants CCR-9988393 and DMS-9972591 1

This paper focuses on the following model reduction problem Given the linear dynamical system Σ in (11), find a reduced order system Σ r { [ ] ξ(t) = Ar ξ(t) + b Σ r : r u(t) Ar b Σ y r (t) = c r ξ(t) + d r u(t) r := r, (12) where A r R r r, b r R r, c T r Rr, d r R, with r n such that the following properties are satisfied: 1 The approximation error y y r is small, and there exists a global error bound 2 System properties, like stability, are preserved 3 The procedure is computationally efficient Moreover, the reduced order models Σ r will be constructed by projection In other words, we will construct matrices V R n r and Z R n r with Z T V = I r such that Σ r : A r = Z T AV, b r = Z T b, c r = cv, d r = d (13) Notice that V Z T is an oblique projector Model reduction methods can be classified into three categories: (a) Singular value decomposition (SVD) based methods, (b) Krylov (moment matching) based methods, and (c) SVD-Krylov based methods The key ingredients of the SVD based methods are the Hankel singular values, which are the singular values of the so-called Hankel operator associated with the system Σ In SVD based model reduction, Hankel singular values play the same role as that of the singular values in the approximation of constant matrices One of the most common approaches in this category is Balanced Model Reduction first introduced by Mullis and Roberts [27] and later by Moore [26] in the systems and control literature To apply balanced reduction, by simultaneously diagonalizing certain system gramians, first the system is transformed to a basis where the states which are difficult to reach are simultaneously difficult to observe Then, the reduced model is obtained by truncating the states which have this property Two other related approaches in this category are Hankel Norm Approximation [16] and Balanced Singular Perturbation Approximation [23, 25] When applied to stable systems, these three approaches preserve stability and provide bounds on the approximation error, ie, meet goals 1 and 2 above Despite these nice system theoretic properties, for large-scale settings, SVD based methods are expensive to implement because they require dense matrix factorizations; the resulting computational complexity is O(n 3 ) and the storage requirements are O(n 2 ) For efficient implementations of balancing related model reduction algorithms, see [2, 37, 38, 39, 4, 41, 42] Krylov based model reduction methods are based on matching the so-called moments of the original transfer function around selected frequencies The k th, k, moment of a system at σ C is given by the k th derivative of the transfer function at σ Then the moment matching problem consists in finding a reduced order model that matches a certain number of moments of the original model at the selected frequencies The Lanczos [24] and Arnoldi [8] procedures, and the rational Krylov method [18] belong to this category of model reduction Krylov based methods are numerically reliable, can be implemented iteratively, and therefore can be applied to systems of high order The number of computations is reduced to O(n 2 r) ( to O(nr) if the matrix A is sparse) and the storage requirement to O(nr); hence goal 3 of the model reduction problem is met But unlike the SVD based methods, stability is not guaranteed and there exists no a priori computable global error bounds for the approximation error Stability can be obtained via restarting [33], however in this case exact moment matching is lost Moreover, recently in [19, 21], a global error expression has been obtained for the Lanczos procedure 2 c r d r

The current research trend in the area of model reduction is to connect SVD and Krylov based approximation methods; see, for example, [22, 34, 3, 36, 2] These methods aim at combining the theoretical features of the SVD based methods such as stability and global error bounds, with the efficient numerical implementation of the moment matching based methods In this paper we present a model reduction method which achieves these goals It is a projection method where the projection depends both on Krylov subspaces and on the system gramians Hence the approach is a combination of the SVD and Krylov based methods, and aims to retain the best features of each For a SISO discrete time system, the approach is based on the least squares fit of the (r+1) st column of Hankel matrix to the preceding r columns The method guarantees stability Moreover, the reduced model matches the first r Markov parameters of the full order model and minimizes a weighted H 2 error It turns out that the approach is related to Prony s method [31, 28], used in signal processing for infinite impulse response (IIR) filter design We also generalize this approach to moment matching at arbitrary interpolation points Application to continuous time systems is achieved by means of the bilinear transformation The proposed method proves efficient, retains stability and matches some moments The rest of the paper is organized as follows First, we review the balanced model reduction method, an SVD based method, and the rational Krylov method, a Krylov based method, in Sections 21 and 22, respectively Section 3 introduces the proposed method for discrete-time systems followed by the continuous-time case in Section 4 Large-scale implementation issues are discussed in Section 5 Section 6 presents a comparison with the SVD and Krylov based methods Three numerical examples are presented in Section 7 followed by the conclusions in Section 8 In the sequel, we will assume that the full order model Σ in (11) is asymptotically stable 1 and minimal; ie, the pairs (A,b) and (c,a) are, respectively, reachable and observable 2 Overview of basic model reduction methods 21 Balanced reduction: An SVD-based method In this section, we review the balanced model reduction method for SISO continuous-time dynamical systems Let Σ be the SISO continuous time model in (11) Closely related to this system are two continuoustime Lyapunov equations: AP + PA T + bb T =, A T Q + QA + c T c = (21) Under the assumptions that Σ is asymptotically stable and minimal, it is well known that the above equations have unique symmetric positive definite solutions P, Q R n n, called the reachability and observability gramians, respectively The square roots of the eigenvalues of the product PQ are the singular values of the Hankel operator associated with Σ and are called Hankel singular values σ i (Σ) of the system Σ: σ i (Σ) = λ i (PQ) In many cases, the eigenvalues of P, Q as well as the Hankel singular values σ i (Σ) decay rapidly This phenomenon is explained to a large extent in [6] This observation leads to model reduction by truncation The minimal and asymptotically stable system Σ is called balanced if P = Q = Σ = diag(σ 1 I m1,σ 2 I m2,,σ q I mq ), 1 A continuous-time system Σ is called asymptotically stable if R(λ i(a)) < where R(λ) denotes the real part of λ Similarly, a discrete-time system Σ is called asymptotically stable if ρ(a) < 1 where ρ(a) denotes the spectral radius of A 3

where q is the number of distinct Hankel singular values, σ 1 > σ 2 > > σ q >, m i, i = 1,,q are the multiplicities of σ i and m 1 + m 2 + m q = n It follows from the above definition that balancing amounts to the simultaneous diagonalization of the two positive definite matrices P and Q The balanced system has the property that the states which are difficult to reach, ie require a large input energy to reach, are simultaneously difficult to observe, ie yield small observation energy The states which have this property correspond to small Hankel singular values Hence a reduced model is simply obtained by truncating these states from the balanced system However in large-scale settings balancing the whole system Σ and then truncating is numerically inefficient and can be ill-conditioned Instead below we will refer to an implementation of balanced reduction which directly obtains a reduced balanced system without balancing the whole system Let P = UU T and Q = LL T This is always possible since both P and Q are symmetric positive definite matrices The matrices U and L are called square roots of the gramians P and Q respectively Let U T L = ZSY T be the singular value decomposition (SVD) It is easy to show that the singular values of U T L are indeed the Hankel singular values, hence we have U T L = ZΣY T where Σ = diag(σ 1 I m1,σ 2 I m2,,σ q I mq ) Let Σ 1 = diag(σ 1 I m1,σ 2 I m2,,σ k I mk ), k < q, r := m 1 + + m k, and define W 1 := LY 1 Σ 1/2 1 and V 1 := UZ 1 Σ 1/2 1, where Z 1 and Y 1 are composed of the leading r columns of Z and Y, respectively It is easy to check that W T 1 V 1 = I r and hence that V 1 W T 1 is an oblique projector We obtain a reduced model of order r by projection as follows: Σ r : A r = W T 1 AV 1, b r = W T 1 b, c r = cv 1, d r = d (22) Noting that PW 1 = V 1 Σ 1 and QV 1 = W 1 Σ 1 gives W T 1 (AP + PA T + bb T )W 1 = A r Σ 1 + Σ 1 A T r + b r b T r, V T 1 (AT Q + QA + c T c)v 1 = A T r Σ 1 + Σ 1 A r + c T r c r Thus, the reduced model is balanced and asymptotically stable (due to the Lyapunov inertia theorem) for any k q As mentioned earlier, the formulas above provide a numerically stable scheme for computing the reduced order model based on a numerically stable scheme for computing the square roots U and L directly in upper triangular and lower triangular form, respectively It is important to truncate Z, Σ, Y to Z 1,Σ 1,Y 1 prior to forming W 1 or V 1 It is also important to avoid formulas involving inversion of L or U as these matrices are typically ill-conditioned due to the decay of the eigenvalues of the gramians The reduced system Σ r defined by (22) is asymptotically stable, and the error system satisfies the following H error bound: Σ Σ r H 2(σ k+1 + + σ q ) (23) For details, see [5, 45] The above result states that if the neglected Hankel singular values are small, Σ and Σ r are guaranteed to be close Note that (23) is an a priori error bound Hence given an error tolerance, one can decide how many states to truncate before computing the reduced model However, the computational complexity of this method is O(n 3 ) and the storage requirements are O(n 2 ) 22 Moment matching and the rational Krylov method In this section, we review the second class of model reduction methods, ie Krylov based (moment matching) methods 4

Given the full-order model Σ as in (11), we expand the transfer function G(s) = c(si A) 1 b + d about a point σ C in the complex plane which is not a pole of Σ: G(s) = c(σi A (σ s)i) 1 b + d = c(i (σi A) 1 (σ s)) 1 (σi A) 1 b + d = η σ () + η(1) σ (σ s) + η(2) σ (σ s)2 + η σ (3) (σ s)3 + = η σ (j) (σ s) j j= η (j) σ is called the j th moment of Σ about σ for j, and is given by η σ () = c(σi A) 1 b + d, and η σ (j) = c(σi A) (j+1) b, for j = 1,2, if σ η () = d, and η (j) = ca j 1 b, for j = 1,2, if σ = (24) The moment matching approximation problem consists in finding a reduced order model Σ r with transfer function G r (s) = ˆη σ () + ˆη σ (1) (σ s) + ˆη σ (2) (σ s) 2 + ˆη σ (3) (σ s) 3 + such that for an appropriate k there holds η (j) σ = ˆη (j) σ j =,1,2,,k (25) For the special case where σ =, the moments are called Markov parameters and the corresponding moment matching problem is the problem of partial realization; for the solution of this problem, see [1], [3] Importantly, this problem can be solved in a recursive and numerically reliable way, by means of the Lanczos and Arnoldi procedures In general for an arbitrary σ C, this becomes the rational interpolation problem, see for example [2] In this case rational Lanczos/Arnoldi procedures give a numerically efficient solution The Padé via Lanczos method of [11], which exploits the connection between the Lanczos procedure and the Padé approximation [9], is the leading effort for this case For the multi-point rational interpolation problem, the solution is given by the rational Krylov method of Ruhe [32] In this case one matches the moments of the transfer function at selected frequencies and therefore a better approximation of the transfer function over a broader frequency range is obtained We would like to note that the Krylov based methods match moments without ever computing them This is important since the computations of moments is usually ill-conditioned 221 Multi-point rational interpolation by Krylov projection methods In the multi-point rational interpolation problem by projection, the goal is to find matrices V R n r and Z R n r with Z T V = I r so that the reduced system Σ r is obtained by projection as in (13) and matches the moments of Σ at the selected interpolation points, ie (25) holds Multi-point rational interpolation by projection was first proposed by Skelton et al in [1, 44, 43] Grimme [18] showed how one can obtain the required projection by Krylov methods Next we will show how one can achieve this goal by Krylov projection methods where the matrices V and Z span unions of certain Krylov spaces First, for a matrix F C n n and a vector g C n, we define the Krylov space of index j: K j (F,g) := Im([ g Fg F 2 g F j 1 g ]) (26) The following theorem is presented in [18] It shows how to construct the matrices V and Z as the span of a union of certain Krylov subspaces so that the multi-point rational interpolation problem by Krylov projection is solved: 5

Theorem 21 If K K bk ((σ k I A) 1,(σ k I A) 1 b) V = Im(V ) (27) k=1 and K K ck ((σ k I A) T,(σ k I A) T c T ) Z = Im(Z) (28) k=1 where Z T V = I and σ k are chosen such that the matrices σ k I A are invertible, k {1,,K}, then the moments of Σ and Σ r satisfy η (j) σ = ˆη(j) σ, for j k =,1,,b k + c k 1 and k = 1,2,,K (29) Theorem 21 states that for moment matching based model reduction all one has to do is to construct full-rank matrices V and Z with Im(V ) and Im(Z) satisfying (27) and (28) respectively Efficient implementations of the rational Krylov method can be found in [18] If one uses only one sided projection, the following corollary results Corollary 21 If K K bk ((σ k I A) 1,(σ k I A) 1 b) V = Im(V ), (21) k=1 Z is any left-inverse of V, ie Z T V = I, and σ k are chosen such that the matrices σ k I A are invertible, k {1,,K}, then the moments of Σ and Σ r satisfy η (j) σ = ˆη(j) σ, for j k =,1,2,,b k 1 and k = 1,2,,K (211) Unlike the SVD based methods, such as balanced reduction, for the Krylov based methods, stability of the reduced model is not guaranteed and no a priori error bound exists However, the methods are numerically reliable and can be implemented iteratively; see, for example, [18, 24, 8, 12, 13] for efficient implementations of the Krylov based methods They reduce the computational cost to O(n 2 r) ( to O(nr) if the matrix A is sparse) and the storage requirements to O(nr) The rational Krylov method has proven to be very efficient for model reduction, see [18], [12], [13] and references therein But it has the drawback that the selection of interpolation points is a difficult task since it is an ad-hoc process For a discussion on the choice of the interpolation points, see [19] This issue will be further analyzed in Example 73 in a comparison to the proposed method 3 Model reduction through least-squares: An SVD-Krylov based approach Recently much research has been done to obtain model reduction algorithms which connect SVD and Krylov based methods; see, for example, [2, 22, 34, 3, 36] These methods aim at combining the theoretical features of the SVD based methods such as stability and global error bound, with the efficient numerical implementation of moment matching based methods This section introduces a model reduction method which achieves these goals It is a projection method where the projection depends both on Krylov subspaces and on the system gramians Hence the approach is a combination of the SVD and Krylov 6

based methods, and aims to retain the best features of each For discrete time systems,it is based on the least squares fit of the (r + 1) st column of Hankel matrix to the preceding r columns The reduced system preserves stability and matches the first r Markov parameters of the full order model (FOM) The application to continuous time is achieved by making use of the bilinear transformation The approach has connections with Prony s method [31, 28], used for infinite impulse response (IIR) filter design in signal processing In 179, Prony derived a special formulation to analyze elastic properties of gases which resulted in linear equations Later, a more general form of Prony s method was applied to digital filter design In the sequel, we will review Prony s method based on the discussion by Parks and Burrus [28] Given a reduced order r, the motivation behind the algorithm is to approximately match K > 2r moments in a somehow optimal way rather than exactly matching 2r moments as done in Krylov based methods We would like to note that when applied to model reduction of dynamical systems, Prony s method in general does not guarantee stability and requires explicit computation of the moments However by letting K, we obtain a model reduction algorithm by projection which avoids the explicit computation of the moments, matches the first r Markov parameters, guarantees asymptotic stability, and minimizes a weighted H 2 error We also generalize the approach to moment matching at arbitrary interpolation points 31 Least-squares model reduction in discrete-time [ ] A b Given is the SISO discrete time system Σ =, where A R n n, b R n, c T R n, and d R c d We assume that Σ is asymptotically stable and minimal It is well known that an r th order Padé approximant [9] matches the first 2r Markov parameters exactly However, no general statement can be made about how well the remaining Markov parameters are matched The method we propose below, matches the first r Markov parameters exactly, and all of the remaining ones approximately in a (weighted) least-squares sense Hence we do expect a good match even for the higher order Markov parameters First some definitions are in order The Hankel matrix H, the infinite reachability matrix R and the infinite observability matrix O corresponding to the system Σ are: H := η 1 η 2 η 3 η 2 η 3 η 4 η 3 η 4 η 5 [, R := b Ab A r 1 b ], O := [ c T A T c T (A T ) r 1 c T ]T, where η i := ca i 1 b is the i th Markov parameter It readily follows that H = OR (31) Let H r and h r+1 be, respectively, the first r columns and the (r + 1) st column of the Hankel matrix H: η 1 η 2 η r η r+1 η 2 η 3 η r+1 η r+2 H r =, h r+1 = (32) η r η r+1 η 2r 1 η 2r 7

Also, denote by R r, the reachability matrix of index r, ie, R r := [ b Ab A r 1 b ] From the above definitions follows [ Ir H r = H and ] [ Ir = OR ] = H r = OR r, h r+1 = He r+1 = ORe r+1 = h r+1 = OA r b Then, the proposed approximation is obtained by computing the least squares fit of the (r + 1) st column h r+1 of H to the preceding r columns of H r ; in other words H r x LS = h r+1 + e LS where e LS is the least squares error Recall that in discrete-time, the observability gramian is the solution to the Stein equation and satisfies Hence it follows that the solution x LS is given by A T QA + c T c = Q, (33) Q = O T O x LS = (H T r H r ) 1 H T r h r+1 = (R T r QR r) 1 R T r QAr b The characteristic polynomial of the reduced system is χ LS (z) = [ 1 z z r ] [ xls 1 ] (34) The reduced model Σ LS is, then, obtained by projection as follows: [ Z T AV Z T ] b Σ LS =, where Z T = (R T r cv d QR r) 1 R T r Q and V = R r (35) Remark 31 It is clear from (35) that the least squares solution reduces to obtaining the two matrices Z and V so that Im(Z) = Im(QR r ) and Im(V ) = Im(R r ) Im(V ) is the r th Krylov subspace of A and b, ie, Im(V ) = K r (A,B) Hence V reflects the moment matching side of the algorithm On the other hand, Z contains the observability gramian Q So, Z reflects the SVD-based side of the algorithm Proposition 31 The least squares solution Σ LS is independent of the original realization of Σ; in other words, Σ LS is unique Proof: Let T, det(t) be a similarity transformation of Σ, and let Ã, b, c, d be the realization of Σ in the transformed basis, ie, Ã = TAT 1, b = Tb, c = ct 1, and d = d 8

For this realization, the reachability matrix R r of index r and the observability gramian Q are easily computed as R r = T R r, Q = T T QT 1 Plugging these relationships into the formula for Z and Ṽ from (35) yields Z T = (R T r QR r) 1 R T r QT 1 = Z T T 1 and Ṽ = T R r = TV Therefore, the reduced order model Σ LS of the transformed realization is given by [ Z Σ T ÃṼ Z T b ] [ Z T T 1 TAT 1 TV Z T T 1 ] Tb LS = = cv d ct 1 = Σ TV d LS Hence, a state-space transformation of Σ yields the same reduced system This concludes the proof Theorem 31 [7] Least-squares model reduction: The least squares solution Σ LS matches the first r Markov parameters of Σ and is asymptotically stable Proof: The first part of the result immediately follows by construction of the solution as in (35) with V = R r and the one sided projection result as discussed in Corollary 21 To prove the second part, ie, asymptotic stability of Σ LS, we first show that Σ LS is stable: It follows from Proposition 31 that, without loss of generality, we can take Q = I Thus from (35), we obtain Z = V where V spans the image of R r, V T V = I r and A LS = V T AV Recall that the observability gramian Q satisfies A T QA+c T c = Q Since Q = I, it follows that A T A I n Multiplying this by V T and V, respectively, from left and right, one gets V T A T AV I r, which in turn yields AV V T A T I n Multiplying the final inequality, once more, by V T and V, respectively, from left and right, we get } V T {{ AV} V} T {{ A T V} A LS A T LS = A LS A T LS I This shows stability, ie, λ i (A LS ) 1 This means that A LS may have poles on the unit circle To prove that this is not the case, we first complete the columns of V to a unitary matrix: U = [ V S ], where UU T = U T U = I n Then A T A+c T c = I n is equivalent to U T A T UU T AU +U T c T cu = I n Written explicitly, the last equation yields [ ] [ Ir V T A T V V T A T ][ S V T AV V T ] [ AS V S T A T V S T A T S S T AV S T = T c T cv V T c T ] cs AS S T c T cv S T c T cs I n r The (1,1) block of the this equation yields I r V T A T V V T AV V T A T SS T AV = V T c T cv (36) Now assume that A LS has a pole on the unit circle, ie, there exits an x R r such that A LS x = V T AV x = λx where λ = 1 Multiplying (36) with x T on the left and with x on the right yields x T V T A T SS T AV x + x T V T c T cv x = S T AV x 2 + cv x 2 = 9

This implies cv x = and S T AV x = Now let y := AV x Then [ ] [ ] [ U T V T λx y = S T y = y = UU T λx y = [ V S ] ] y = λv x Hence V x is an eigenvector of A with λ = 1 and is in the null space of c This is a contradiction to the observability of the pair (c,a) Therefore, the reduced system Σ LS is asymptotically stable Let x LS = [ α α r 1 ] T Since the characteristic polynomial of Σ LS is given by χ LS as in (34) and Σ LS matches the first r Markov parameters of Σ as shown in Theorem 31 above, the reduced system Σ LS has the following form: α 1 [ ] 1 α 1 ALS b Σ LS = LS = (37) c LS d α r 2 1 α r 1 η 1 η 2 η r 1 η r d Remark 32 As stated in the above proof, one can assume Q = I In this case the least-squares approximation becomes the Arnoldi procedure [8, 4] for (A, b) However, transforming Σ to the state-space realization where Q = I requires taking the complete Cholesky factorization of the gramian Q, and inverting In the large-scale setting, this is a hard task to accomplish and one will not pursue it unless Q has special structure, and the computations are well-conditioned 311 Error analysis This section presents an analysis of the error resulting from model reduction by the least-squares method It will be shown that Σ LS as in (35) is obtained by solving a weighted H 2 optimization problem Before stating the main result of this section, some definitions are in order: Let the least-squares solution Σ LS in (37) have the impulse response h LS (n) and the transfer function Define H LS (z) = B LS(z) A LS (z) = b + b 1 z 1 + + b r z r 1 + a 1 z 1 + + a r z r a LS (n) := a n for n =,,r with a = 1 (38) Similarly, let ˆΣ be any r th order approximation to Σ with the impulse response ĥ(n) satisfying ĥ(n) = h(n) for n =,,r, and the transfer function Define Ĥ(z) = ˆB(z) Â(z) = ˆb + ˆb 1 z 1 + + ˆb r z r 1 + â 1 z 1 + + â r z r â(n) := â n for n =,,r with â = 1 The next result shows that the least-squares model reduction method minimizes the H 2 norm of the weighted error system H we := Â(z)(H(z) Ĥ(z)) 1

Lemma 31 Given the above set-up, there holds a LS (n) = arg min â(n) â(n) ( h(n) ĥ(n) ) 2, (39) or equivalently, ) A LS (z) = arg min (H(z) Â(z) Ĥ(z) H2, (31) Â(z) where ( ) denotes the convolution operator Proof: A proof can be found in Section 33 on page 15 Lemma 31 shows that the reduced model Σ LS is the optimal solution of a weighted H 2 minimization problem This also means that besides exactly matching the first r Markov parameters, Σ LS approximately matches the higher order moments in a weighted least-squares sense In view of Theorem 31 and Lemma 31, the proposed method combines the following aspects of SVD and Krylov methods: 1 Σ LS matches the the first r Markov parameters; 2 Σ LS is guaranteed to be stable; 3 Σ LS minimizes a weighted H 2 error norm 32 Moment matching at arbitrary interpolation points The least-squares solution Σ LS matches the first r Markov parameters of Σ, hence yields a better approximation around high frequencies However one would like to interpolate the original model at various points in oder to achieve a better approximation over a broader frequency range This goal can be easily achieved by a modification of the matrix V in (35) Instead of letting V = R r, given L interpolation points σ 1,,σ L with multiplicities β k, k = 1,,L, let V span the union of Krylov subspaces corresponding to these points, that is, Im(V ) = L K βk ((σ k I A) 1,(σ k I A) 1 b), (311) k=1 where K j (F,g) := Im([ g Fg F 2 g F j 1 g ]) is the j th Krylov subspace of F and g We know how to construct V in (311) in a numerically efficient way using the rational Krylov method To reflect the fact that V spans a rational Krylov subspace, we will call the resulting method the Rational Least-Squares Method and the corresponding reduced model, denoted by Σ RL, the Rational Least-Squares Approximant The following lemma holds: [ ] A b Lemma 32 Given the discrete-time system Σ =, the L interpolation points σ c d 1,,σ L with multiplicities β k, k = 1,,L, let V satisfy (311) Let the rational least-squares approximant Σ RL be constructed as follows: [ ] [ ARL b Σ RL = RL Z T AV Z T ] b =, (312) cv d c RL d RL 11

where Z T = (V T QV ) 1 V T Q and Im(V ) = L K βk ((σ k I A) 1,(σ k I A) 1 b) (313) Then Σ RL is asymptotically stable and matches β k moments of Σ at σ k for k = 1,,L, ie, for k = 1,,L k=1 c(σ k I A) 1 b + d = c RL (σ k I A RL ) 1 b RL + d RL and (314) c(σ k I A) (j k+1) b = c RL (σ k I A RL ) (j k+1) b RL for j k = 1,,β k 1, (315) Proof: The asymptotic stability of Σ RL follows from the asymptotic stability of Σ LS Notice that the proof of Theorem 31 does not depend on whether V spans the image of the reachability matrix R r or the rational Krylov subspace (311), and consequently Σ RL is asymptotically stable The moment matching property is a direct consequence of the one sided projection result as discussed in Corollary 21 33 Prony s Method In this section we briefly review Prony s method, and present its connection to the least-squares model reduction method introduced in Section 31 The goal in Prony s Method is to design a system (digital filter) with a prescribed time-domain response h(n) for n = 1,, K + 1 Let the transfer function of the to-be-modeled filter be H(z) = B(z) A(z) = b + b 1 z 1 + + b M z M 1 + a 1 z 1, (316) + + a N z N where M N It is well known that the impulse response h(n) is related to H(z) by the z-transform H(z) = h(n)z n n= One can rewrite (316) as B(z) = H(z)A(z), which is simply the z-transform of the convolution Using the given first K + 1 terms of the impulse response, we can write this convolution in matrix form as b b 1 b 2 b M = }{{} =: b h h 1 h h 2 h 1 h h M h h K h K N } {{ } =:H 1 a 1 a 2 a N }{{} =:ā (317) One can uncouple the computations of a n, for n = 1,,N and b n, for n =,,M, by partitioning b, H and ā as follows: [ ] [ ][ ] b H = 1 1, (318) h 21 H 2 a 12

where b = b b 1 b M RM+1, a = a 1 a 2 a N RN, h 21 = h M+1 h M+2 h K RK M, H 1 R (M+1) (N+1), and H 2 R (K M) (N) Then one obtains two sets of equations: h 21 = H 2 a (319) b = H 1 ā (32) First (319) must be solved for a and then b is obtained from (32) Hence, the computations of the coefficients a and b are decoupled If K = M + N, H 2 is square If H 2 is not singular, a unique solution a to (319) is computed and a unique b is obtained from (32) Note that there are M + N + 1 unknowns and the same number of impulse response coefficients are matched It is obvious that this is the Padé approximation [9] of a dynamical system, where one simply matches the first M + N + 1 Markov parameters of the underlying dynamical system However, we would like to note that for large-scale settings, Padé approximation is never computed this way The connection between moment matching and Krylov subspace projection allows us to match the moments without computing them as discussed in Section 22 If H 2 is singular, then (319) has many solutions and h(n) can be generated by a lower order system This amounts to computing a minimal realization of a non-minimal system As stated earlier in Section 31, the problem with Padé approximation, ie, exactly matching the first K + 1 terms of h(n), is that it says nothing about interpolation for n > K Hence, to control h(n) over a larger range, Prony s method tries to approximately match h(n) for n > K in a somehow optimal way Towards this goal, let K > M + N It is obvious [ ] that in this case (319) and (32) cannot be solved e1 exactly Hence, introduce an error term e = in (318) as [ b ] [ e1 + e 2 (321) leads to the two sets of equations: ] [ = e 2 H 1 h 21 H 2 ][ 1 a ] = Hā (321) b + e 1 = H 1 ā (322) e 2 = h 21 + H 2 a (323) A least-squares solution of (323) gives a By plugging this a value into (322), an exact solution for b is obtained from (322), ie, e 1 = Let ĥ(n) denote the impulse response of the resulting approximate system obtained by Prony s method Then e 1 = implies that the approximate impulse response ĥ(n) matches the desired impulse response h(n) for n =,, M However, for n > M, there exists no exact matching Let ε(n) denote the length-(k + 1) solution error between the desired impulse response h(n) and the approximate impulse response ĥ(n), ie, ε(n) = h(n) ĥ(n) for n =,,K Then it can be easily shown that the equation error e of Prony s method is a weighted version of the absolute error ε, namely [ ] [ ] e1 ε1 = e = A e c ε = A c (324) 2 ε 2 13

where A c = 1 a 1 1 a N a 1 1 a N a 1 1 a N a 1 1 R (K+1) (K+1) (325) In other words, A c is the (K + 1) (K + 1) convolution matrix composed of the coefficients of a(n) As discussed above, by construction, we know that e 1 = ε 1 =, ie h(n) = ĥ(n) for n =,,M Based on this observation (324) can be written as e 2 = Āc ε 2, where Āc is the leading (K M) (K M) submatrix of A c Hence Prony s method minimizes a weighted version of the error norm 331 Connection to least-squares model reduction Let the least-squares approximant Σ LS as defined in (35) have order r Then, there holds: N = M = r Also, because of the fact that the D-term of the reduced model is the same as the D-term of Σ, one has b = d where b is given in (316) Since, for a discrete time system Σ, the i th Markov parameter is the i th element of the impulse response of h(n), by applying Prony s method to the impulse response h(n) of Σ, one matches the first r Markov parameters of Σ exactly, and the following K r ones approximately in the least-square sense, where K > 2r However, the main disadvantage is that unlike Padé approximation via Krylov projection, where the moments are matched without being computed, Prony s method requires the explicit computation and usage of the moments To get an r th order model with K > 2r in Prony s method, one will need to compute the first K Markov parameters and use them in the solution of the resulting least-squares problem However, such an approach will fail in large-scale settings Indeed avoiding the computation of moments [11] is the main motivation of Krylov based methods, such as Padé via Lanczos The leastsquares model reduction method avoids these ill-conditioned computations of the Markov parameters by letting K This information is contained in the observability gramian Q Hence as already mentioned, one obtains an r th order reduced model which matches the first r Markov parameters exactly, and the remaining higher order ones approximately in a (weighted) least-squares sense without explicitly computing them Furthermore, the stability result does not hold for Prony s method in general However, by letting K =, in addition to avoiding explicit computation of the moments, we guarantee asymptotic stability in the least-squares model reduction Finally, as Lemma 31 presents, K = results in a minimization of a weighted H 2 error norm In summary, by letting K, we have obtained a generalization of Prony s method, with main attributes: 1 no explicit computation of the Markov parameters; 2 guaranteed stability; 3 minimization of a weighted H 2 error norm 14

Having presented the connection between Prony s method and the least-squares model reduction method, we are now ready to give a proof of Lemma (31) Proof of Lemma 31: Since we have K = in our least-squares method, it is clear from (324) and the definition of A c in (325) that the error e of Prony s method is obtained as the convolution of the filter a LS (n) with the absolute error ε(n) = h(n) h LS (n) In the frequency domain, one has E(z) = A LS (z)(h(z) H LS (z)) Then, from the definition of the H 2 norm and Parseval s theorem it follows that e 2 2 = E(z) 2 H 2 Recall that Prony s method minimizes e 2 2 over all possible solutions Ĥ(z) with e 1 = This observation leads to the desired result 4 Least-squares model reduction in continuous time In this section, we will examine how to apply the least-squares method of the previous section to model reduction of continuous-time systems As expected, the bilinear transformation is the key towards this goal 41 Bilinear Transformation between continuous and discrete time systems The bilinear transformation is a spectral transformation w : C C of the type w : s z = 1 + s 1 s, (41) which maps the open left-half plane onto the open unit disc, ie if R(s) < = z = 1 + s 1 s < 1 (42) The transfer function H d (z) of a discrete-time system obtained via bilinear transformation of a continuous time system H c (s) is given by ( ) z 1 H d (z) = H c (43) z + 1 Due to (42), if H c (s) is asymptotically stable, then so is H d (z), ie, the bilinear transformation preserves stability The inverse bilinear transformation is defined by w 1 : z s = z 1 z + 1, (44) and maps the open unit disc onto the left half plane H c (s) is obtained from H d (z) via inverse bilinear transformation as ( ) 1 + s H c (s) = H d (45) 1 s Like the bilinear transformation, the inverse bilinear transformation preserves stability as well 15

[ Ad B Let the discrete-time system Σ d = d C d z= 1+s 1 s A c,b c,c c,d c D d ] [ ] Ac B and the continuous time system Σ c = c be obtained from each other via the (inverse) bilinear transformation The following relationships hold true: A c = (A d + I) 1 (A d I) B c = 2(A d + I) 1 B d C c = 2C d (A d + I) 1 D c = D d C d (A d + I) 1 B d 42 Least-squares reduction in the continuous time case A d = (I + A c )(I A c ) 1 B d = 2 (I A c ) 1 B c C d = 2 C c (I A c ) 1 (46) D d = D c + C c (I A c ) 1 B c C c D c s= z 1 z+1 A d,b d,c d,d d (47) Given the continuous time system Σ, the goal is to find a reduced order model Σ r, via (rational) leastsquares reduction, which interpolates Σ at σ k, with multiplicities β k for k = 1,,L This problem can be solved using Rational Least Squares Reduction via the Bilinear Transformation as explained below It is based on the following observation Let H d (z) be obtained from H c (s) via the bilinear transformation Then ( ) 1 + σ H c (σ) = H d 1 σ Algorithm 41 Continuous time least-squares model reduction 1 First apply a bilinear transformation to Σ to obtain the discrete time model Σ d 2 Apply the rational least squares method of Section 32 to Σ d at the interpolation points ς k = 1+σ k 1 σ k with the multiplicities β k for k = 1,,L Let Σ RL denote the reduced model 3 Apply an inverse bilinear transformation to Σ RL to obtain the continuous time reduced model Σ r The following lemma summarizes the important properties of the above algorithm [ ] A b Lemma 41 Given an asymptotically stable continuous time system Σ = and the interpolation c d points σ k with multiplicities β k and σ k 1 for k = 1,,L, [ let Σ r be ] obtained by continuous time rational Ar b least-squares reduction as in Algorithm 41 Then Σ r = r is asymptotically stable and matches the first β k moments of Σ at σ k, for k = 1,,L, ie, for k = 1,,L c r c(σ k I A) 1 b + d = c r (σ k I A r ) 1 b r + d r, and (48) c(σ k I A) (j k+1) b = c r (σ k I A r ) (j k+1) b r, for j k = 1,,β k 1, (49) Proof: The asymptotic stability directly follows from the fact that both the discrete time rational least squares reduction and the bilinear transformation preserve asymptotic stability To prove the moment d r 16

matching property, we proceed as follows: Let H c (s), H r (s), H d (z) and H RL (z), denote the transfer functions of, respectively, Σ, Σ r, Σ d and Σ RL From the rational least-squares method in discrete-time, we have d j H d dz j = dj H RL dz j 1 σ k z= 1+σ k z= 1+σ k 1 σ k for j =,,b k 1 (41) It follows from the equalities ( ) 1 + s H c (s) = H d 1 s ( ) 1 + s and H r (s) = H RL 1 s (411) and from (41) for j = that H c (σ k ) = H r (σ k ) Hence, Σ r matches the first moment at σ k To prove the result for higher order moments, we differentiate the equalities in (411) The first-order derivatives yield dh c ds = dh d dα dα ds and dh r ds = dh RL dα dα ds, (412) where α = 1+s 1 s Note that since σ k 1, the derivatives of α are well defined Plugging s = σ k into these relationships together with (41) for j = 1 leads to dh c ds = dh RL s=σk dz (413) s=σk Therefore, Σ r matches the second moment of Σ at σ k The result for higher order moments follows similarly Some remarks are in order: Remark 41 (1) Because of the bilinear transformation, the D-term of the continuous time least squares approximation does not need to be the same as that of the original model However, by choosing one of the interpolation point σ j =, one can guarantee that Σ and Σ r has the same D-term (2) Lemma 41 excludes the case σ k = 1 However, if required, one can easily match the moments at σ k = 1 by making use of a generalized bilinear transformation z = µ+s µ s where µ > with µ 1 In this case the interpolation points in discrete time are given by ς k = µ+σ k µ σ k (3) Both the discrete-time and continuous-time rational least-square methods perform well as many examples show in Section 7 We believe that the method is important for model reduction as it is a combination of SVD based and Krylov based methods, guarantees stability and matches moments at the same time One remaining issue is to establish an error bound 5 Implementation issues in large-scale settings In this section, we discuss the difficulties that might be encountered in large-scale settings and propose remedies to overcome them We will first discuss the discrete-time problem in Section 51 followed by the continuous time case in Section 52 17

51 Large-scale discrete time implementations It is clear from (35) that the least-squares reduction problem amounts to the computation of two full-rank matrices V and Z such that Im(V ) = Im(R r ) and Im(Z) = Im(QR r ) with Z T V = I r (51) The image of V spans a Krylov subspace and we know how to construct it for large-scale settings, see Section 22 The real concern is to obtain the observability gramian Q It is well known that computing a full-rank Q exactly is an ill-conditioned problem in large-scale settings; see, for example, [2, 29, 5] and the references therein Therefore, we will replace the exact gramian Q with a low-rank approximation Q To achieve this goal, we propose to use low-rank Smith (LR-Smith) type methods, which have proven to be effective for this purpose, such as the LR-Smith Method of Penzl [29] or the Modified LR-Smith Method of Gugercin et al [2] Even though Q is approximated by a low-rank approximation Q, the matrix V in (35) is not affected because it does not depend on Q Hence only the matrix Z is modified Let us denote this modified matrix by Z, ie, Z T := (R T r QR r ) 1 R T r Q (52) 511 Low-rank version and the update of (52) based on the low-rank Smith iterate Although the formula (52) uses Q in the dense form and consequently needs O(n 2 ) storage, by using a low-rank Smith method, we will avoid storing an n n matrix In the LR-Smith s method [29, 2], we never compute Q explicitly, instead we compute and store only a low-rank Cholesky factor Y R n k such that Q = Y Y T approximates the full-rank gramian Q, where k is the number of steps taken in the LR-Smith iteration Based on this observation, Z in (52) can be written as Z T = (R T r QR r ) 1 R T r Q = (R T r Y Y T R }{{} r ) 1 R T r Y Y T = (ΩΩ T ) 1 ΩY T :=Ω where Ω R r k This way, we never compute and store a dense n n matrix explicitly Only a low-rank n k matrix Y is stored Now assume that Q is approximated by a k step LR-Smith iteration, and due to a slow convergence in the Smith s method, k steps were not enough to approximate the gramian Q Hence, the goal is to go one step further to the (k + 1) st step in the LR-Smith iteration, and update Z as well as the reduced order model using the information from the k th step Let Q k and Q k+1 denote the approximate gramians obtained by k and k + 1 step LR-Smith iterations, respectively Recall that Q solves the Stein equation A T QA + c T c = Q It follows from the implementation of the LR-Smith method that (see [29, 2]) Q k and Q k+1 satisfy Q k+1 = Q k + (A T ) k c T ca k Hence, one obtains R T r Q k+1 R r = R T r Q k R r + R T r (AT ) k c T ca k R r (53) This final equality reveals that the multiplication R T r Q k+1 R r, and consequently Z can be iteratively obtained Therefore, in order to update Z for the (k + 1) st step Smith iterate Q k+1, one needs to compute only the latter part of the last equation The iteratively computed R T r Q k+1 R r in (53) and Z can be used to obtain the updated reduced model We finally note that even though (53) is formulated in the dense case, a low-rank implementation directly follows from (53) above 18

52 Implementation issues for the continuous time problem For continuous time systems, in addition to the implementation issues discussed in Section 51 above, there is the additional difficulty arising from the bilinear transformation The question is how to apply the bilinear transformation for large-scale problems We note that the inverse bilinear transformation (47) to obtain the final reduced order continuous time system is not a problem since the order is already reduced Hence it is enough to discuss the bilinear transformation (46) from the full-order continuous time model to the full-order discrete-time model The best way to overcome the numerical problems associated with this transformation is to never apply it explicitly, in other words A d, B d, C d and D d in (46) are never explicitly computed Instead, triangular solvers are employed to compute the operations, such as multiplications, associated with these four matrices Note that only one shift is applied and consequently only one LU decomposition needs to be computed Since A will be sparse in most cases, one can deal with the problem easily using a sparse direct factorization Another important observation is that in order to obtain Q by means of a Smith iteration, application of the bilinear transformation is necessary (see [29, 2, 3]) We conclude that the bilinear transformation does not introduce significant additional numerical complexity Especially for sparse systems, the transformation can be efficiently applied without explicitly computing the transformed state-space matrices Remark 51 We would like to note that as in the case of approximate balanced truncation [2, 29], stability of the reduced system is not always guaranteed when one uses approximate low-rank gramians However, this does not seem to be a difficulty in practice; we obtained a stable reduced system for each of our computational examples where we used the Modified LR-Smith Method of [2] 6 Comparison with SVD and Krylov based methods SVD-based methods, such as balanced reduction, are likely to yield reduced order models which yield globally good approximations, ie they match the original model over the whole frequency range This can be seen from the fact that there exist upper bounds for the H norm of the error system This property of the SVD methods is due to explicit usage of the system gramians P and Q which carry the global information On the contrary, Krylov based methods do not use the system gramians, hence do not include any global information in the underlying projection They contain local information through moment matching Therefore, even though Krylov methods result in better local approximants than SVD based methods, their global behavior is usually poorer Consequently in most cases, the H norm of the error for Krylov based methods is much higher than that of the SVD based methods However, when the interpolation points are chosen appropriately, this difference can get smaller, and furthermore the H 2 performance of both approaches can be comparable (even better in some cases), see [19, 21] where the interpolation points are chosen as the mirror images of the poles of Σ The least-squares method is a combination of these two approaches At the cost of reducing the number of moments matched by half, global information via the usage of Q is added (or equivalently, at the cost of removing the global information contained in P, some moments are matched) Hence, the proposed method obviously lies between the other two approaches We expect it to behave better locally than SVD-based methods, and to behave better globally than Krylov based methods We also expect the least-squares method to yield lower H 2 and H error norms than Krylov based methods These heuristics are examined through several examples in Section 7 19