H 2 -optimal model reduction of MIMO systems

Size: px
Start display at page:

Download "H 2 -optimal model reduction of MIMO systems"

Transcription

1 H 2 -optimal model reduction of MIMO systems P. Van Dooren K. A. Gallivan P.-A. Absil Abstract We consider the problem of approximating a p m rational transfer function Hs of high degree by another p m rational transfer function b Hs of much smaller degree. We derive the gradients of the H 2-norm of the approximation error and show how stationary points can be described via tangential interpolation. Keyword Multivariable systems, model reduction, optimal H 2 approximation, tangential interpolation. Introduction In this paper we will consider the problem of approximating a real p m rational transfer function Hs of McMillan degree N by a real p m rational transfer function Ĥs of lower McMillan degree n using the H 2 -norm as approximation criterion. Since a transfer function has an unbounded H 2 -norm if it is not proper a rational transfer function is proper if it is zero at s =, we will constrain both Hs and Ĥs to be proper. Such transfer functions have state-space realizations A,B,C R N2 R Nm R pn and Â, B,Ĉ Rn2 R nm R pn satisfying Hs := CsI N A B and Ĥs := ĈsI n  B. The realization {Â, B,Ĉ} is not unique in the sense that the triple {ÂT, B T,ĈT } := {T ÂT,T B, ĈT } for any matrix T GLn, R defines the same transfer function : Ĥs = ĈsI n  B = Ĉ T si n ÂT BT. It is known see e.g. Theorem 4.7 in Byrnes and Falb [3] that the geometric quotient of R n2 R nm R pn under GLn, R is a smooth, irreducible variety of dimension nm + p. This implies that the set Rat n p,m of p m proper rational transfer functions of degree n can be parameterized with only nm+p real parameters in a locally smooth manner. A possible approach for building a reduced order model {Â, B,Ĉ} from a full order model {A,B,C} is tangential interpolation, which can always be achieved see [4] by solving two Sylvester equations for the unknowns W,V R N n AV V Σσ + BR =, 2 and constructing the reduced order model of degree n as follows W T A Σ T µw T + L T C =, 3 {Â, B,Ĉ} := {W T V W T AV,W T V W T B,CV }, 4 Corresponding author. address: paul.vandooren@uclouvain.be CESAME, Université catholique de Louvain, B-348 Louvain-la-Neuve, Belgium School of Computational Science, Florida State University, Tallahassee FL 3236, USA This paper presents research supported by the Belgian Network DYSCO Dynamical Systems, Control, and Optimization, funded by the Interuniversity Attraction Poles Programme, initiated by the Belgian State, Science Policy Office and by the National Science Foundation under contract OCI The scientific responsibility rests with its authors.

2 provided the matrix W T V is invertible which also implies that V and W must have full rank n. The interpolation conditions {Σ σ,r} and {Σ µ,l} where Σ µ,σ σ R n n, R R m n and L R p n are known to uniquely determine the projected system {Â, B,Ĉ} [4]. The equations above can be expressed in another coordinate system by applying invertible transformations of the type { Q Σ σ Q,RQ } and { P Σ µ P,LP } to the interpolation conditions, which yields transformed matrices V P and WQ but does not affect the transfer function of the reduced order model {Â, B,Ĉ} see [4]. Therefore, the interpolation conditions essentially impose nm + p real conditions, since Σ σ and Σ µ can be transformed to their Jordan canonical form. In the case that both matrices are simple no Jordan blocks of size larger than we can assume Σ σ and Σ µ to be block diagonal with a diagonal block σ i or µ i for each real condition and a 2 2 diagonal block [ σ i σ i+ ] [ µ σ i+ σ i or i µ i+ ] µ i+ µ i for each pair of complex conjugate conditions. We refer to [] for a more elaborate discussion on this and for a discrete-time version of the results of this paper. In this paper we first compute the gradients of the H 2 error of the approximation problem and then show that its stationary points satisfy special tangential interpolation conditions that generalize earlier results for SISO systems and help understand numerical algorithms to solve this model reduction problem. 2 The H 2 approximation problem Let Es be an arbitrary proper transfer function, with realization triple {A e,b e,c e }. If Es is unstable, its H 2 -norm is defined to be. Otherwise, its squared H 2 -norm is defined as the trace of a matrix integral [2] : J := Es 2 H 2 := tr Ejω H Ejω dω 2π = tr EjωEjω H dω 2π. By Parseval s identity, this can also be expressed using the state space realization as see [2] J = tr [C e exp Aet B e ][C e exp Aet B e ] T dt = tr [C e exp Aet B e ] T [C e exp Aet B e ]dt. This can also be related to an expression involving the Gramians P e and Q e defined as P e := [exp Aet B e ][exp Aet B e ] T dt, Q e := which are also known to be the solutions of the Lyapunov equations [exp Aet B e ] T [C e exp Aet ]dt, A e P e + P e A T e + B e B T e =, Q e A e + A T e Q e + C T e C e =. 5 Using these, it easily follows that the squared H 2 -norm of Es can also be expressed as J = tr B T e Q e B e = tr C e P e C T e. 6 We now apply this to the error function Es := Hs Ĥs. A realization of Es in partitioned form is given by {A e,b e,c e } := {[ [ A B B] [,, C Ĉ] Â] }, and the Lyapunov equations 5 become [ ] P X P e := X T, P [ ] [ ] [ ][ A P X P X A T Â X T + P X T P Â T ] [ ] [ + B BT] B B T =, 7 and [ ] Q Y Q e := Y T, Q [ A T Â T ] [ Q ] Y Y T Q [ ] [ ] Q Y A C T [ + Y Q][ T + C Ĉ] =. 8 Â ĈT 2

3 To minimize the H 2 -norm, J, of the error function Es we must minimize [ ] [ [ ] J = tr B T Q Y BT Y Q] B B T = tr B T QB + 2B T Y B + B T Q B, 9 where Q, Y and Q depend on A, Â, C and Ĉ through the Lyapunov equation 8, or equivalently [ ] [ ][ ] P X C T J = tr C Ĉ = tr CPC P T ĈT 2CXĈT + Ĉ PĈT, X T where P, X and P depend on A, Â, B and B through the Lyapunov equation 7. Note that the terms B T QB and CPC T in the above expressions are constant, and hence can be discarded in the optimization. 3 Optimality conditions The expansions above can be used to express first order optimality conditions for the squared H 2 -norm in terms of the gradients of J versus Â, B and Ĉ. We define a gradient as follows. Definition 3. The gradient of a real scalar function fx of a real matrix variable X R n p is the real matrix X fx R n p defined by It yields the expansion [ X fx] i,j = d dx i,j fx, i =,...,n, j =,...,p. fx + = fx + X fx, + O 2, where M,N := tr M T N. The following lemma is useful in the derivation of our results see [7]. Lemma 3.2 If AM + MB + C = and NA + BN + D =, then trcn = trdm. Starting from the characterizations 7, and 8,9 of the H 2 norm and using Lemma 3.2 we easily derive succinct forms of the gradients. This theorem is originally due to Wilson [8]. Theorem 3.3 The gradients ba J, bb J and bc J of J := Es 2 H 2 are given by ba J = 2 Q P + Y T X, bb J = 2 Q B + Y T B, bc J = 2Ĉ P CX, where A T Y + Y  CT Ĉ =,  T Q + Q + ĈT Ĉ =, 2 X T A T + ÂXT + BB T =, P T +  P + B B T =. 3 Proof. For finding an expression for ba J we consider the characterization J = tr B T QB + 2B T Y B + B T Q B, A T Y + Y  CT Ĉ =,  T Q + Q + ĈT Ĉ =. Then the first order perturbation J corresponding to ba is given by J = tr 2 BB T Y + B B T bq where Y and bq depend on ba via the equations A T Y + Y  + Y ba =,  T bq + bq  + T b A Q + Q ba =. 4 3

4 It follows from applying Lemma 3.2 to the Sylvester equations 3,4 that tr BB T Y = tr X T Y ba and tr B BT bq = tr P T Q ba + Q ba and therefore J = tr 2X T Y ba + P T A b Q + Q ba = tr 2X T Y ba + 2 P Q ba = 2 Q P + Y T X, ba. Since J also equals ba J, ba, it follows that ba J = 2 Q P + Y T X. To find an expression for bb J we perturb B in the characterization J = tr B T QB + 2B T Y B + B T Q B. which yields the first order perturbation J = tr 2B T Y bb + T Q B bb + BT Q bb = 2Y T B + Q B, bb. Since J also equals bb J, bb, it follows that bb J = 2 Q B + Y T B. In a similar fashion we can write the first order perturbation of J = tr CPC T 2CXĈT + Ĉ PĈT to obtain bc J = 2Ĉ P CX. The gradient forms of Theorem 3.3 allow us to derive our fundamental theoretical result. Theorem 3.4 At every stationary point of J where P and Q are invertible, we have the following identities  = W T AV, B = W T B, Ĉ = CV, W T V = I n with W := Y Q, V := X P 5 where X, Y, P and Q satisfy the Sylvester equations 2,3. Proof. Since we are at a stationary point of J, the gradients versus Â, B and Ĉ must be zero : Q P + Y T X =, Q B + Y T B =, Ĉ P CX =. Since P and Q are invertible, we can define W := Y Q and V := X P. It then follows that W T V = I n, B = W T B, Ĉ = CV. Multiplying the first equation of 3 with W and using X T = PV T, yields PV T A T W +  PV T W + BB T W =. Using V T W = I, B T W = B T and the second equation of 3 it then follows that  = W T AV. If we rewrite the above theorem as a projection problem, then we are constructing a projector Π := V W T implying W T V = I n where V and W are given by the following transposed Sylvester equations QW T A + ÂT QW T + ĈT C =, AV P + V PÂT + B B T =. 6 Notice that P and Q can be interpreted as normalizations to ensure that W T V = I n. 4

5 It was shown in [4] that projecting a system via Sylvester equations always amounts to satisfying tangential interpolation conditions. The Sylvester equations 6 show that the parameters of reduced order models corresponding to stationary points must have specific relationships with the parameters of the tangential interpolation conditions 2,3,4. First note that  = Σ σ = Σ µ requires that the left and right interpolation points are identical and equal to the negatives of the poles of the reduced order model. For SISO systems, choosing identical left and right interpolation point sets implies that Ĥs and Hs and, at least, their first derivatives match at the interpolation points. Theorem 3.4 therefore generalizes to MIMO systems the conditions of [6] on the H 2 -norm stationary points for SISO systems. The simple additive result for the orders of rational interpolation for SISO systems, however, is replaced by more complicated tangential conditions for MIMO systems that require the definition of tangential direction vectors that can be vector polynomials of s. The Sylvester equations 6 show that these direction vectors are also related to parameters of realizations of Ĥs. If the Sylvester equations are expressed in the coordinate system with  in Jordan form then the transformed B and Ĉ contain the parameters that define the tangential interpolation directions. 4 Tangential interpolation revisited Theorem 3.4 provides the fundamental characterization of the stationary points of J via tangential interpolation conditions and their relationship to the realizations of Ĥs. It is instructive to illustrate those relationships in a particular coordinate system and derive an explicit form of the tangential interpolation conditions. We assume here that all poles of Ĥs are distinct but possibly complex the so-called generic case. Hence the transfer functions Hs and Ĥs have real realizations {A,B,C} and {Â, B,Ĉ} with  diagonalizable. The interpretation of these conditions for multiple poles or higher order poles becomes more involved and can be found in an extended version of this paper []. Given our assumptions, we have for Ĥs the partial fraction expansion Ĥs = n i= ĉ i bh i s λ i, 7 where b i C m and ĉ i C p and where λ i, b i,ĉ i,i =,...,n is a self-conjugate set. We must keep in mind that the number of parameters in {Â, B, Ĉ} is not minimal and hence that the gradient conditions of Theorem 3.3 must be redundant. We make this more explicit in the theorem below. For this we will need s i, t H i, the complex left and right eigenvectors of the real matrix  corresponding to the complex eigenvalue λ i. Because of the expansion 7, we then have : Âs i = λ i s i, Ĉs i = ĉ i, t H i  = λ i t H i, t H i B = b H i. Theorem 4. Let Ĥs = n i= ĉi b H i /s λ i have distinct first order poles where λ i, b i,ĉ i, i =,...,n is self-conjugate. Then 2 B bj T s i = [H T λ i ĤT λ i ]ĉ i 8 2 th i bc J T = b H i [H T λ i ĤT λ i ] 9 2 th i ba J T s i = b H d i ds [HT s ĤT s] ĉ i s= λi b 2 2 th i ba J T s j = 2 λ i λ j [ b H i bb J T s j t H i bc J T ĉ j ] 2 5

6 Proof. Define y i := Y s i, q i := Qs i, x i := Xt i and p i := Pt i. Then from 2,3 we have It follows that from which we obtain A T + λ i Iy i = C T ĉ i, ÂT + λ i I q i = ĈT ĉ i, x H i A T + λ i I = b H i B T, p H i ÂT + λ i I = b H i y i = A T + λ i I C T ĉ i, q i = ÂT + λ i I Ĉ T ĉ i, 22 x H i = b H i B T A T + λ i I, p H i = b H i B T. 2 b B J T s i = B T Q + B T Y s i = [H T λ i ĤT λ i ]ĉ i, B T ÂT + λ i I, 23 2 th i bc J T = t H i PĈT X T C T = b H i [H T λ i ĤT λ i ]. From the 22,23 it also follows that 2 th i ba J T s j = t H i P Q+X T Y s j = b H i [ B T ÂT + λ i I ÂT + λ j I Ĉ T B T A T + λ i I A T + λ j I C T ]ĉ i. If we use d ds Hs = CsI A 2 B and dsĥs d = ĈsI Â 2 B, then for i = j we obtain 2 th i ba J T s i = b H d i ds [HT s ĤT s] ĉ i. s= λi b For i j we use the identity M + λ j I M + λ i I = λ i λ j [M + λ j I M + λ i I ] to obtain 2 th i ba J T s j = λ i λ j t H i [H T λ i ĤT λ i ] [H T λ j ĤT λ j ] s j and finally 2 th i ba J T s j = 2 λ i λ j [ b H i bb J T s j t H i bc J T ĉ j ]. Let S := [ s... s n ], then the above theorem shows that the off-diagonal elements of S ba J T S vanish when bb J T and bc J T vanish. Therefore we need to impose only conditions on diags ba J T S, on bb J T and on bc J T to characterize stationary points of J. These are exactly nm + p conditions since the vectors b H i or ĉ i can be scaled as indicated in Section. Moreover one can view them as nm + p real conditions since the poles λ i come in complex conjugate pairs. The following corollary easily follows. Corollary 4.2 If bb J T =, bc J T = and diag S ba J T S = then ba J = and the following tangential interpolation conditions are satisfied for all λ i,i =,...,n : [H T λ i ĤT λ i ]ĉ i =, bh i [H T d λ i ĤT λ i ] =, bh i ds [HT s ĤT s] ĉ i =. 24 s= λi b Notice that we retrieve the conditions of [6] for the SISO case since then b H i and ĉ i are just nonzero scalars that can be divided out. The conditions above then become the familiar 2n interpolation conditions H λ i = Ĥ λ i, d ds Hs s= b λ i = d ds Ĥs s= b λi, i =,...,n. 6

7 5 Concluding remarks The H 2 norm of a stable proper transfer function Es is a smooth function of the parameters {A e,b e,c e } of its state-space realization because the squared norm of Es is differentiable versus the parameters {A e,b e,c e } as long as A e is stable the Lyapunov equations are then invertible linear maps and the trace is a smooth function of its parameters. If Ĥs is an isolated local minimum of the error function Hs Ĥs 2 H 2, then the continuity of the norm implies that a small perturbation of Hs will induce also only a small perturbation of that local minimum. This explains why we can construct a characterization of the optimality conditions without assuming anything about the structure of the poles of the transfer functions Hs and Ĥs. Those ideas also lead to algorithms. One can view 2,3 and 5 as two coupled equations X,Y, P, Q = FÂ, B,Ĉ and Â, B,Ĉ = GX,Y, P, Q for which we have a fixed point Â, B,Ĉ = GFÂ, B,Ĉ at every stationary point of J Â, B,Ĉ. This automatically suggests an iterative procedure X,Y, P, Q i+ = FÂ, B,Ĉ i+, Â, B,Ĉ i+ = GX,Y, P, Q i, which is expected to converge to a nearby fixed point. This is essentially the idea behind existing algorithms using Sylvester equations in their iterations see [2]. Another approach would be to use the gradients or the interpolation conditions of Theorem 4. to develop descent methods or even Newton-like methods, as was done for the SISO case in [5]. The two fundamental contributions of this paper are, first, the characterization of the stationary points of J via tangential interpolation conditions and their relationship to the realizations of Ĥs given by Theorem 3.4, and, second, the fact that this can be done using Sylvester equations without assuming anything about the structure of either Hs or Ĥs thereby providing a framework to relate existing algorithms and to develop and understand new ones. References [] P.-A. Absil, K. A. Gallivan and P. Van Dooren. Multivariable H 2 -optimal approximation and tangential interpolation. Internal CESAME Report, Catholic University of Louvain, September 27. [2] A. Antoulas. Approximation of Large-Scale Dynamical Systems. Siam Publications, Philadelphia 25 [3] C. Byrnes and P. Falb. Applications of algebraic geometry in systems theory. American Journal of Mathematics, 2: , April 979. [4] K. A. Gallivan, A. Vandendorpe, and P. Van Dooren. Model reduction of MIMO systems via tangential interpolation. SIAM J. Matrix Anal. Appl., 262: , 24 [5] S. Gugercin, A. Antoulas and C. Beattie. Rational Krylov methods for optimal H2 model reduction, submitted for publication, 26. [6] L. Meier and D. Luenberger. Approximation of linear constant systems. IEEE Trans. Aut. Contr., 2: , 967. [7] W.-Y. Yan and J. Lam. An approximate approach to H 2 optimal model reduction. IEEE Trans. Aut. Contr. 447:34-358, 999. [8] D. A. Wilson, Optimum solution of model reduction problem, Proc. Inst. Elec. Eng., 7:665, 97. 7

H 2 optimal model reduction - Wilson s conditions for the cross-gramian

H 2 optimal model reduction - Wilson s conditions for the cross-gramian H 2 optimal model reduction - Wilson s conditions for the cross-gramian Ha Binh Minh a, Carles Batlle b a School of Applied Mathematics and Informatics, Hanoi University of Science and Technology, Dai

More information

Model reduction of interconnected systems

Model reduction of interconnected systems Model reduction of interconnected systems A Vandendorpe and P Van Dooren 1 Introduction Large scale linear systems are often composed of subsystems that interconnect to each other Instead of reducing the

More information

Projection of state space realizations

Projection of state space realizations Chapter 1 Projection of state space realizations Antoine Vandendorpe and Paul Van Dooren Department of Mathematical Engineering Université catholique de Louvain B-1348 Louvain-la-Neuve Belgium 1.0.1 Description

More information

Krylov Techniques for Model Reduction of Second-Order Systems

Krylov Techniques for Model Reduction of Second-Order Systems Krylov Techniques for Model Reduction of Second-Order Systems A Vandendorpe and P Van Dooren February 4, 2004 Abstract The purpose of this paper is to present a Krylov technique for model reduction of

More information

Model reduction via tangential interpolation

Model reduction via tangential interpolation Model reduction via tangential interpolation K. Gallivan, A. Vandendorpe and P. Van Dooren May 14, 2002 1 Introduction Although most of the theory presented in this paper holds for both continuous-time

More information

Model Reduction for Unstable Systems

Model Reduction for Unstable Systems Model Reduction for Unstable Systems Klajdi Sinani Virginia Tech klajdi@vt.edu Advisor: Serkan Gugercin October 22, 2015 (VT) SIAM October 22, 2015 1 / 26 Overview 1 Introduction 2 Interpolatory Model

More information

Fluid flow dynamical model approximation and control

Fluid flow dynamical model approximation and control Fluid flow dynamical model approximation and control... a case-study on an open cavity flow C. Poussot-Vassal & D. Sipp Journée conjointe GT Contrôle de Décollement & GT MOSAR Frequency response of an

More information

On some interpolation problems

On some interpolation problems On some interpolation problems A. Gombani Gy. Michaletzky LADSEB-CNR Eötvös Loránd University Corso Stati Uniti 4 H-1111 Pázmány Péter sétány 1/C, 35127 Padova, Italy Computer and Automation Institute

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

Affine iterations on nonnegative vectors

Affine iterations on nonnegative vectors Affine iterations on nonnegative vectors V. Blondel L. Ninove P. Van Dooren CESAME Université catholique de Louvain Av. G. Lemaître 4 B-348 Louvain-la-Neuve Belgium Introduction In this paper we consider

More information

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems Proceedings of the 44th IEEE Conference on Decision and Control, and the European Control Conference 2005 Seville, Spain, December 12-15, 2005 WeC10.4 An iterative SVD-Krylov based method for model reduction

More information

Implicit Volterra Series Interpolation for Model Reduction of Bilinear Systems

Implicit Volterra Series Interpolation for Model Reduction of Bilinear Systems Max Planck Institute Magdeburg Preprints Mian Ilyas Ahmad Ulrike Baur Peter Benner Implicit Volterra Series Interpolation for Model Reduction of Bilinear Systems MAX PLANCK INSTITUT FÜR DYNAMIK KOMPLEXER

More information

Control Systems. Laplace domain analysis

Control Systems. Laplace domain analysis Control Systems Laplace domain analysis L. Lanari outline introduce the Laplace unilateral transform define its properties show its advantages in turning ODEs to algebraic equations define an Input/Output

More information

Balanced Truncation 1

Balanced Truncation 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI

More information

Generalized Shifted Inverse Iterations on Grassmann Manifolds 1

Generalized Shifted Inverse Iterations on Grassmann Manifolds 1 Proceedings of the Sixteenth International Symposium on Mathematical Networks and Systems (MTNS 2004), Leuven, Belgium Generalized Shifted Inverse Iterations on Grassmann Manifolds 1 J. Jordan α, P.-A.

More information

BALANCING AS A MOMENT MATCHING PROBLEM

BALANCING AS A MOMENT MATCHING PROBLEM BALANCING AS A MOMENT MATCHING PROBLEM T.C. IONESCU, J.M.A. SCHERPEN, O.V. IFTIME, AND A. ASTOLFI Abstract. In this paper, we treat a time-domain moment matching problem associated to balanced truncation.

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

REVISITING THE STABILITY OF COMPUTING THE ROOTS OF A QUADRATIC POLYNOMIAL

REVISITING THE STABILITY OF COMPUTING THE ROOTS OF A QUADRATIC POLYNOMIAL Electronic Transactions on Numerical Analysis. Volume 44, pp. 73 82, 2015. Copyright c 2015,. ISSN 1068 9613. ETNA REVISITING THE STABILITY OF COMPUTING THE ROOTS OF A QUADRATIC POLYNOMIAL NICOLA MASTRONARDI

More information

Similarity matrices for colored graphs

Similarity matrices for colored graphs Similarity matrices for colored graphs Paul Van Dooren Catherine Fraikin Abstract In this paper we extend the notion of similarity matrix which has been used to define similarity between nodes of two graphs

More information

Zeros and zero dynamics

Zeros and zero dynamics CHAPTER 4 Zeros and zero dynamics 41 Zero dynamics for SISO systems Consider a linear system defined by a strictly proper scalar transfer function that does not have any common zero and pole: g(s) =α p(s)

More information

Inexact Solves in Krylov-based Model Reduction

Inexact Solves in Krylov-based Model Reduction Inexact Solves in Krylov-based Model Reduction Christopher A. Beattie and Serkan Gugercin Abstract We investigate the use of inexact solves in a Krylov-based model reduction setting and present the resulting

More information

Realization-independent H 2 -approximation

Realization-independent H 2 -approximation Realization-independent H -approximation Christopher Beattie and Serkan Gugercin Abstract Iterative Rational Krylov Algorithm () of [] is an effective tool for tackling the H -optimal model reduction problem.

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2018 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

Eigenvalues and Eigenvectors: An Introduction

Eigenvalues and Eigenvectors: An Introduction Eigenvalues and Eigenvectors: An Introduction The eigenvalue problem is a problem of considerable theoretical interest and wide-ranging application. For example, this problem is crucial in solving systems

More information

Jordan Canonical Form

Jordan Canonical Form Jordan Canonical Form Massoud Malek Jordan normal form or Jordan canonical form (named in honor of Camille Jordan) shows that by changing the basis, a given square matrix M can be transformed into a certain

More information

Gramians of structured systems and an error bound for structure-preserving model reduction

Gramians of structured systems and an error bound for structure-preserving model reduction Gramians of structured systems and an error bound for structure-preserving model reduction D.C. Sorensen and A.C. Antoulas e-mail:{sorensen@rice.edu, aca@rice.edu} September 3, 24 Abstract In this paper

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

Math Ordinary Differential Equations

Math Ordinary Differential Equations Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

More information

Model reduction of large-scale dynamical systems

Model reduction of large-scale dynamical systems Model reduction of large-scale dynamical systems Lecture III: Krylov approximation and rational interpolation Thanos Antoulas Rice University and Jacobs University email: aca@rice.edu URL: www.ece.rice.edu/

More information

Non-negative matrix factorization with fixed row and column sums

Non-negative matrix factorization with fixed row and column sums Available online at www.sciencedirect.com Linear Algebra and its Applications 9 (8) 5 www.elsevier.com/locate/laa Non-negative matrix factorization with fixed row and column sums Ngoc-Diep Ho, Paul Van

More information

Krylov-Subspace Based Model Reduction of Nonlinear Circuit Models Using Bilinear and Quadratic-Linear Approximations

Krylov-Subspace Based Model Reduction of Nonlinear Circuit Models Using Bilinear and Quadratic-Linear Approximations Krylov-Subspace Based Model Reduction of Nonlinear Circuit Models Using Bilinear and Quadratic-Linear Approximations Peter Benner and Tobias Breiten Abstract We discuss Krylov-subspace based model reduction

More information

MIT Final Exam Solutions, Spring 2017

MIT Final Exam Solutions, Spring 2017 MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Note on the convex hull of the Stiefel manifold

Note on the convex hull of the Stiefel manifold Note on the convex hull of the Stiefel manifold Kyle A. Gallivan P.-A. Absil July 9, 00 Abstract In this note, we characterize the convex hull of the Stiefel manifold and we find its strong convexity parameter.

More information

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM Ralf L.M. Peeters Bernard Hanzon Martine Olivi Dept. Mathematics, Universiteit Maastricht, P.O. Box 616, 6200 MD Maastricht,

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

EIGENVALUES AND EIGENVECTORS 3

EIGENVALUES AND EIGENVECTORS 3 EIGENVALUES AND EIGENVECTORS 3 1. Motivation 1.1. Diagonal matrices. Perhaps the simplest type of linear transformations are those whose matrix is diagonal (in some basis). Consider for example the matrices

More information

CSL361 Problem set 4: Basic linear algebra

CSL361 Problem set 4: Basic linear algebra CSL361 Problem set 4: Basic linear algebra February 21, 2017 [Note:] If the numerical matrix computations turn out to be tedious, you may use the function rref in Matlab. 1 Row-reduced echelon matrices

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information

CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren

CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren CESAME, Université catholique de Louvain Bâtiment Euler, Avenue G. Lemaître 4-6 B-1348 Louvain-la-Neuve,

More information

NOTES ON LINEAR ODES

NOTES ON LINEAR ODES NOTES ON LINEAR ODES JONATHAN LUK We can now use all the discussions we had on linear algebra to study linear ODEs Most of this material appears in the textbook in 21, 22, 23, 26 As always, this is a preliminary

More information

Module 09 From s-domain to time-domain From ODEs, TFs to State-Space Modern Control

Module 09 From s-domain to time-domain From ODEs, TFs to State-Space Modern Control Module 09 From s-domain to time-domain From ODEs, TFs to State-Space Modern Control Ahmad F. Taha EE 3413: Analysis and Desgin of Control Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/

More information

The Bock iteration for the ODE estimation problem

The Bock iteration for the ODE estimation problem he Bock iteration for the ODE estimation problem M.R.Osborne Contents 1 Introduction 2 2 Introducing the Bock iteration 5 3 he ODE estimation problem 7 4 he Bock iteration for the smoothing problem 12

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 1964 1986 www.elsevier.com/locate/laa An iterative SVD-Krylov based method for model reduction of large-scale dynamical

More information

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems

An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems An iterative SVD-Krylov based method for model reduction of large-scale dynamical systems Serkan Gugercin Department of Mathematics, Virginia Tech., Blacksburg, VA, USA, 24061-0123 gugercin@math.vt.edu

More information

Iterative Rational Krylov Algorithm for Unstable Dynamical Systems and Generalized Coprime Factorizations

Iterative Rational Krylov Algorithm for Unstable Dynamical Systems and Generalized Coprime Factorizations Iterative Rational Krylov Algorithm for Unstable Dynamical Systems and Generalized Coprime Factorizations Klajdi Sinani Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University

More information

On Solving Large Algebraic. Riccati Matrix Equations

On Solving Large Algebraic. Riccati Matrix Equations International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University

More information

Model Reduction of Inhomogeneous Initial Conditions

Model Reduction of Inhomogeneous Initial Conditions Model Reduction of Inhomogeneous Initial Conditions Caleb Magruder Abstract Our goal is to develop model reduction processes for linear dynamical systems with non-zero initial conditions. Standard model

More information

OPTIMAL H 2 MODEL REDUCTION IN STATE-SPACE: A CASE STUDY

OPTIMAL H 2 MODEL REDUCTION IN STATE-SPACE: A CASE STUDY OPTIMAL H 2 MODEL REDUCTION IN STATE-SPACE: A CASE STUDY RLM Peeters, B Hanzon, D Jibetean Department of Mathematics, Universiteit Maastricht, PO Box 66, 6200 MD Maastricht, The Netherlands, E-mail: ralfpeeters@mathunimaasnl

More information

Jordan Canonical Form Homework Solutions

Jordan Canonical Form Homework Solutions Jordan Canonical Form Homework Solutions For each of the following, put the matrix in Jordan canonical form and find the matrix S such that S AS = J. [ ]. A = A λi = λ λ = ( λ) = λ λ = λ =, Since we have

More information

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix. Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric

More information

Math Linear Algebra Final Exam Review Sheet

Math Linear Algebra Final Exam Review Sheet Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of

More information

Subdiagonal pivot structures and associated canonical forms under state isometries

Subdiagonal pivot structures and associated canonical forms under state isometries Preprints of the 15th IFAC Symposium on System Identification Saint-Malo, France, July 6-8, 29 Subdiagonal pivot structures and associated canonical forms under state isometries Bernard Hanzon Martine

More information

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY D.D. Olesky 1 Department of Computer Science University of Victoria Victoria, B.C. V8W 3P6 Michael Tsatsomeros Department of Mathematics

More information

Canonical lossless state-space systems: staircase forms and the Schur algorithm

Canonical lossless state-space systems: staircase forms and the Schur algorithm Canonical lossless state-space systems: staircase forms and the Schur algorithm Ralf L.M. Peeters Bernard Hanzon Martine Olivi Dept. Mathematics School of Mathematical Sciences Projet APICS Universiteit

More information

arxiv: v1 [cs.sy] 17 Nov 2015

arxiv: v1 [cs.sy] 17 Nov 2015 Optimal H model approximation based on multiple input/output delays systems I. Pontes Duff a, C. Poussot-Vassal b, C. Seren b a ISAE SUPAERO & ONERA b ONERA - The French Aerospace Lab, F-31055 Toulouse,

More information

Efficient and Accurate Rectangular Window Subspace Tracking

Efficient and Accurate Rectangular Window Subspace Tracking Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu

More information

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity: Diagonalization We have seen that diagonal and triangular matrices are much easier to work with than are most matrices For example, determinants and eigenvalues are easy to compute, and multiplication

More information

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra Massachusetts Institute of Technology Department of Economics 14.381 Statistics Guido Kuersteiner Lecture Notes on Matrix Algebra These lecture notes summarize some basic results on matrix algebra used

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Optimal Scaling of Companion Pencils for the QZ-Algorithm

Optimal Scaling of Companion Pencils for the QZ-Algorithm Optimal Scaling of Companion Pencils for the QZ-Algorithm D Lemonnier, P Van Dooren 1 Introduction Computing roots of a monic polynomial may be done by computing the eigenvalues of the corresponding companion

More information

MATH 5640: Functions of Diagonalizable Matrices

MATH 5640: Functions of Diagonalizable Matrices MATH 5640: Functions of Diagonalizable Matrices Hung Phan, UMass Lowell November 27, 208 Spectral theorem for diagonalizable matrices Definition Let V = X Y Every v V is uniquely decomposed as u = x +

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Matrices and Linear Algebra

Matrices and Linear Algebra Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2

More information

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show MTH 0: Linear Algebra Department of Mathematics and Statistics Indian Institute of Technology - Kanpur Problem Set Problems marked (T) are for discussions in Tutorial sessions (T) If A is an m n matrix,

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0. Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

SQUARE ROOTS OF 2x2 MATRICES 1. Sam Northshield SUNY-Plattsburgh

SQUARE ROOTS OF 2x2 MATRICES 1. Sam Northshield SUNY-Plattsburgh SQUARE ROOTS OF x MATRICES Sam Northshield SUNY-Plattsburgh INTRODUCTION A B What is the square root of a matrix such as? It is not, in general, A B C D C D This is easy to see since the upper left entry

More information

BALANCING-RELATED MODEL REDUCTION FOR DATA-SPARSE SYSTEMS

BALANCING-RELATED MODEL REDUCTION FOR DATA-SPARSE SYSTEMS BALANCING-RELATED Peter Benner Professur Mathematik in Industrie und Technik Fakultät für Mathematik Technische Universität Chemnitz Computational Methods with Applications Harrachov, 19 25 August 2007

More information

Model Reduction of Linear Systems, an Interpolation point of View

Model Reduction of Linear Systems, an Interpolation point of View Université Catholique De Louvain Faculté des sciences appliquées Département d ingénierie mathématique Center for Systems Engineering and Applied Mechanics Model Reduction of Linear Systems, an Interpolation

More information

Applications of Controlled Invariance to the l 1 Optimal Control Problem

Applications of Controlled Invariance to the l 1 Optimal Control Problem Applications of Controlled Invariance to the l 1 Optimal Control Problem Carlos E.T. Dórea and Jean-Claude Hennet LAAS-CNRS 7, Ave. du Colonel Roche, 31077 Toulouse Cédex 4, FRANCE Phone : (+33) 61 33

More information

Zentrum für Technomathematik

Zentrum für Technomathematik Zentrum für Technomathematik Fachbereich 3 Mathematik und Informatik h 2 -norm optimal model reduction for large-scale discrete dynamical MIMO systems Angelika Bunse-Gerstner Dorota Kubalińska Georg Vossen

More information

Homework 6 Solutions. Solution. Note {e t, te t, t 2 e t, e 2t } is linearly independent. If β = {e t, te t, t 2 e t, e 2t }, then

Homework 6 Solutions. Solution. Note {e t, te t, t 2 e t, e 2t } is linearly independent. If β = {e t, te t, t 2 e t, e 2t }, then Homework 6 Solutions 1 Let V be the real vector space spanned by the functions e t, te t, t 2 e t, e 2t Find a Jordan canonical basis and a Jordan canonical form of T on V dened by T (f) = f Solution Note

More information

MULTIVARIABLE ZEROS OF STATE-SPACE SYSTEMS

MULTIVARIABLE ZEROS OF STATE-SPACE SYSTEMS Copyright F.L. Lewis All rights reserved Updated: Monday, September 9, 8 MULIVARIABLE ZEROS OF SAE-SPACE SYSEMS If a system has more than one input or output, it is called multi-input/multi-output (MIMO)

More information

2.3. VECTOR SPACES 25

2.3. VECTOR SPACES 25 2.3. VECTOR SPACES 25 2.3 Vector Spaces MATH 294 FALL 982 PRELIM # 3a 2.3. Let C[, ] denote the space of continuous functions defined on the interval [,] (i.e. f(x) is a member of C[, ] if f(x) is continuous

More information

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77 1/77 ECEN 605 LINEAR SYSTEMS Lecture 7 Solution of State Equations Solution of State Space Equations Recall from the previous Lecture note, for a system: ẋ(t) = A x(t) + B u(t) y(t) = C x(t) + D u(t),

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p

More information

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8. Linear Algebra M1 - FIB Contents: 5 Matrices, systems of linear equations and determinants 6 Vector space 7 Linear maps 8 Diagonalization Anna de Mier Montserrat Maureso Dept Matemàtica Aplicada II Translation:

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

Robust Control 2 Controllability, Observability & Transfer Functions

Robust Control 2 Controllability, Observability & Transfer Functions Robust Control 2 Controllability, Observability & Transfer Functions Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University /26/24 Outline Reachable Controllability Distinguishable

More information

SYSTEMTEORI - ÖVNING Stability of linear systems Exercise 3.1 (LTI system). Consider the following matrix:

SYSTEMTEORI - ÖVNING Stability of linear systems Exercise 3.1 (LTI system). Consider the following matrix: SYSTEMTEORI - ÖVNING 3 1. Stability of linear systems Exercise 3.1 (LTI system. Consider the following matrix: ( A = 2 1 Use the Lyapunov method to determine if A is a stability matrix: a: in continuous

More information

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND

More information

Symmetric and anti symmetric matrices

Symmetric and anti symmetric matrices Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal

More information

Control Systems. Frequency domain analysis. L. Lanari

Control Systems. Frequency domain analysis. L. Lanari Control Systems m i l e r p r a in r e v y n is o Frequency domain analysis L. Lanari outline introduce the Laplace unilateral transform define its properties show its advantages in turning ODEs to algebraic

More information

Case study: Approximations of the Bessel Function

Case study: Approximations of the Bessel Function Case study: Approximations of the Bessel Function DS Karachalios, IV Gosea, Q Zhang, AC Antoulas May 5, 218 arxiv:181339v1 [mathna] 31 Dec 217 Abstract The purpose of this note is to compare various approximation

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

Solution of Linear State-space Systems

Solution of Linear State-space Systems Solution of Linear State-space Systems Homogeneous (u=0) LTV systems first Theorem (Peano-Baker series) The unique solution to x(t) = (t, )x 0 where The matrix function is given by is called the state

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

MATH 5524 MATRIX THEORY Problem Set 4

MATH 5524 MATRIX THEORY Problem Set 4 MATH 5524 MATRIX THEORY Problem Set 4 Posted Tuesday 28 March 217. Due Tuesday 4 April 217. [Corrected 3 April 217.] [Late work is due on Wednesday 5 April.] Complete any four problems, 25 points each.

More information

Mathematical Olympiad Training Polynomials

Mathematical Olympiad Training Polynomials Mathematical Olympiad Training Polynomials Definition A polynomial over a ring R(Z, Q, R, C) in x is an expression of the form p(x) = a n x n + a n 1 x n 1 + + a 1 x + a 0, a i R, for 0 i n. If a n 0,

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Represent this system in terms of a block diagram consisting only of. g From Newton s law: 2 : θ sin θ 9 θ ` T

Represent this system in terms of a block diagram consisting only of. g From Newton s law: 2 : θ sin θ 9 θ ` T Exercise (Block diagram decomposition). Consider a system P that maps each input to the solutions of 9 4 ` 3 9 Represent this system in terms of a block diagram consisting only of integrator systems, represented

More information

Study Guide for Linear Algebra Exam 2

Study Guide for Linear Algebra Exam 2 Study Guide for Linear Algebra Exam 2 Term Vector Space Definition A Vector Space is a nonempty set V of objects, on which are defined two operations, called addition and multiplication by scalars (real

More information