Exercise 8: Frequency Response of MIMO Systems

Size: px
Start display at page:

Download "Exercise 8: Frequency Response of MIMO Systems"

Transcription

1 Exercise 8: Frequency Response of MIMO Systems 8 Singular Value Decomposition (SVD The Singular Value Decomposition plays a central role in MIMO frequency response analysis Let s recall some concepts from the course Lineare Algebra I/II : 8 Preliminary Definitions The induced norm A of a matrix that describes a linear function like is defined as y = A u (8 A = max u 0 y u = max y u = (8 Remark In the course Lineare Algebra I/II you learned that there are many different norms and the expression A could look too generic However, in this course we will always use the euclidean norm A = A This norm is defined as A = µ max = maximal eigenvalue of A A (83 where A is the conjugate transpose or Hermitian transpose of the matrix A This is defined as A = (conj(a T (84 Example Let Then it holds A = i + i i A = (conj(a T T + i = i i i = + i i Remark If we deal with A R n m holds of course One can list a few useful properties of the euclidean norm: (i Remember: A T A is always a quadratic matrix! A = A T (85 (ii If A is orthogonal: A = (86

2 (iii If A is symmetric: A = max( e i (87 (iv If A is invertible: A = µmin = minimal eigenvalue of A A (88 (v If A is invertible and symmetric: A = min( e i (89 where e i are the eigenvalues of A In order to define the SVD we have to go a step further Let s consider a Matrix A and the linear function 8 It holds A = max u = y y = max u = (A u (A u = max u = u A A u = max i µ(a A = max i σ i (80 where σ i are the singular values of matrix A They are defined as where µ i are the eigenvalues of A A Combining 8 and 8 one gets σ i = µ i (8 σ min (A y u σ max(a (8 8 Singular Value Decomposition Our goal is to write a general matrix A C p m as product of three matrices: U, Σ and V It holds A = U Σ V with U C p p, Σ R p m, V C m m (83 Remark U and V are orthogonal, Σ is a diagonal matrix

3 y = A u u Figure : Illustration of the singular values Kochrezept: Let A C p m be given: (I Compute all the eigenvalues and eigenvectors of the matrix A A C m m and sort them as µ µ µ r > µ r+ = = µ m = 0 (84 (II Compute an orthogonal basis from the eigenvectors v i and write it in a matrix as V = (v v m C m m (85 (III We have already found the singular values: they are defined as σ i = µ i for i =,, min{p, m} (86 We can then write Σ as σ 0 0 Σ = R p m, p < m (87 σ m 0 0 (IV One finds u,, u r from σ Σ = σ n 0 0 R p m, p > m ( u i = σ i A v i for all i =,, r (for σ i 0 (89 3

4 (V If r < p one has to complete the basis u,, u r (with ONB to obtain an orthogonal basis, with U orthogonal (VI If you followed the previous steps, you can write A = U Σ V Motivation for the computation of Σ, U und V A A = (U Σ V (U Σ V = V Σ U U Σ V = V Σ Σ V = V Σ V (80 This is nothing else than the diagonalization of the matrix A A The columns of V are the eigenvectors of A A and the σi the eigenvalues For U: A A = (U Σ V (U Σ V = U Σ V V Σ U = U Σ Σ U = U Σ U (8 This is nothing else than the diagonalization of the matrix A A The columns of U are the eigenvectors of A A and the σi the eigenvalues Remark In order to derive the previous two equations I used that: The matrix A A is symmetric, ie U = U (because U is othogonal V = V (because V is othogonal (A A = A (A = A A Remark Since the matrix A A is always symmetric and positive semidefinite, the singular values are always real numbers Remark The MATLAB command for the singular value decomposition is [U,S,V]=svd One can write A T as A =transpose(a and A as A =conj(transpose(a Those two are equivalent for real numbers Remark Although I ve reported detailed informations about the calculation of U and V, this won t be relevant for the exam It is however good and useful to know the reasons that are behind this topic 4

5 Example Let u be with u = The matrix M is given as u = M = cos(x sin(x 0 0 We know that the product of M and u defines a linear function y = M u 0 cos(x = 0 sin(x cos(x = sin(x We need the maximum of y In order to avoid square roots, one can use that the x that maximizes y should also maximize y y = 4 cos (x + 4 sin (x has maximum d y dx = 8 cos(x sin(x + sin(x cos(x =! 0 { x max = 0, π, π, 3π } Inserting back for the maximal y one gets: y max =, y max = The singular values can be calculated with M M: M M = M M 4 0 = 0 4 λ i = { 4, } 4 σ i = {, } As stated before, one can see that y [σ min, σ max ] The matrix U has eigenvectors of M M as coulmns and the matrix V has eigenvectors of M M as columns In this case M M = M M, hence the two matrices are equal Since their product is a diagonal matrix one should recall from the theory that the eigenvectors are easy to determine: they are nothing else than the standard basis vectors This means Interpretation: U = [ 0 0 ] [ 0, Σ = 0 ], V = [ ] 0 0 Referring to Figure, let s interprete these calculations One can see that the maximal amplification occurs at v = V (:, and has direction u = U(:,, ie the vector u is doubled (σ max The minimal amplification occurs at v = V (:, and has direction u = U(:,, ie the vector u is halved (σ min 5

6 05 u y = M u Figure : Illustration von der Singularwertzerlegung Example 3 Let 3 0 A = be given Question: Find the singular values of A and write down the matrix Σ Solution Let s compute A T A: A T A = ( One can see easily that the eigenvalues are = 3 λ = 6, λ = 9 ( The singular values are We write then σ = 4, σ = Σ =

7 Example 4 A transfer function G(s is given as ( s+3 s+ s+3 Find the singular values of G(s at ω = rad s s+ s+3 s+3 7

8 Solution The transfer function G(s evaluated at ω = rad s G(j = ( j+3 j+ j+3 j+ j+3 j+3 has the form In order to calculate the singular values, we have to compute the eigenvalues of H = G G: H = G G = = ( j+3 j+ ( 3 0 j+3 0 = j+ j+3 j+3 ( 3 3 ( j+3 j+ j+3 j+ j+3 j+3 For the eigenvalues it holds It follows det(h λ = det = ( 3 0 λ 0 0 ( 3 0 λ 3 λ 0 ( 0 = λ 6 0 λ + 5 ( 00 = λ 0 ( λ 5 0 λ = 0 and so λ = σ = σ =

9 8 Frequency Responses As we learned for SISO systems, if one excites a system with an harmonic signal u(t = h(t cos(ω t, (8 the answer after a big amount of time is still an harmonic function with equal frequency ω: y (t = P (j ω cos(ω t + (P (j ω (83 One can generalize this and apply it to MIMO systems With the assumption of p = m, ie equal number of inputs and outputs, one excite a system with µ cos(ω t + φ u(t = h(t (84 µ m cos(ω t + φ m and get Let s define two diagonal matrices and two vectors y (t = ν cos(ω t + ψ ν m cos(ω t + ψ m (85 Φ = diag(φ,, φ m R m m, Φ = diag(ψ,, ψ m R m m (86 µ = ( µ µ m T, ν = ( ν ν m T (87 With these one can compute the Laplace Transform of the two signals as: U(s = e Φ s ω µ s s + ω (88 and Y (s = e Ψ s s ω ν (89 s + ω With the general equation for a systems one gets Y (s = P (s U(s e Ψ s s Φ s s ω ν = P (s e s ω + ω µ s + ω e Ψ j ω ω ν = P (s e Φ j ω ω µ e Ψ j ν = P (s e Φ j µ (830 We then recall that the induced norm for the matrix of a linear transformation y = A u from 8 Here it holds P (j ω = e Ψ j ν max e Φ j µ 0 e Φ j µ = max e Ψ j ν e Φ j µ = (83 9

10 Since and One gets e Φ j µ = µ (83 e Ψ j ν = ν (833 P (j ω = max µ 0 ν µ = max µ = ν (834 Here one should get the feeling of why we introduced the singular value decomposition From the theory we ve learned, it is clear that and if µ σ min (P (j ω ν σ max (P (j ω (835 σ min (P (j ω ν µ σ max(p (j ω (836 with σ i singular values of P (j ω These two are worst case ranges and is important to notice that there is no exact formula for ν = f(µ 8 Maximal and minimal Gain You are given a singular value decomposition P (j ω = U Σ V (837 One can read out from this decomposition several informations: the maximal/minial gain will be reached with an excitation in the direction of the column vectors of U The response of the system will then be in the direction of the coulmn vectors of V Let s look at an example and try to understand how to use these informations: Example 5 We consider a system with m = inputs and p = 3 outputs We are given its singular value decomposition at ω = 5 rad: s Σ = 0 063, 0 0 ( V =, j j j j j U = j j j j j j For the singular value σ max = 0467 the eigenvectors are V (:, and U(:, : ( V =, V j =, (V =, j U = j, U = 0960, (U = j

11 The maximal gain is then reached with ( 0908 cos(5 t u(t = cos(5 t 068 The response of the system is then 075 cos(5 t cos(5 t 858 y(t = σ max 0960 cos(5 t 5548 = cos(5 t cos(5 t cos(5 t 474 Since the three signals y (t, y (t and y 3 (t are not in phase, the maximal gain will never be reached One can show that max y(t 0460 < 0467 = σ max t The reason for this difference stays in the phase deviation between y (t, y (t and y 3 (t The same analysis can be computed for σ min 8 Robustness and Disturbance Rejection Let s redefine the matrix norm as Remark It holds G(s = max (max (σ i (G(i ω (838 ω i G (s G (s G (s G (s (839 As we did for SISO systems, we can resume some important indicators for robustness and noise amplification: Robustness Noise Amplification SISO MIMO µ = min ( + L(j ω µ = min (σ min (I + L(j ω ω ω S = max ( S(j ω S = max (σ max(s(j ω ω ω

12 Example 6 Given the MIMO system P (s = ( s+3 s+ s+ 3 s+ Starting at t = 0, the system is excited with the following input signal: ( cos(t u(t = µ cos(t + ϕ Find the parameters ϕ and µ such that for steady-state conditions the output signal y (t y (t has y (t equal to zero

13 Solution For a system excited using a harmonic input signal µ cos(ωt + ϕ u(t = µ cos(ωt + ϕ the output signal y(t, after a transient phase, will also be harmonic and hence have the form ν cos(ωt + ψ u(t = ν cos(ωt + ψ As we have learned, it holds For the first component one gets e Ψ j ν = P (jω e Φ j µ e ψ j ν = P (jω e ϕ j µ + P (jω e ϕ j µ For y (t = 0 to hold we must have ν = 0 In the given case, some parameters can be easily copied from the signals: With the given transfer functions, one gets 0 = j µ j + eϕ j 0 = 3 j 0 + µ j 0 = 3 j 0 + µ j e ϕ j µ = ϕ = 0 ω = (cos(ϕ + j sin(ϕ 0 = µ (cos(ϕ + sin(ϕ + j Splitting the real to the imaginary part, one gets two equations: µ (cos(ϕ + sin(ϕ = 0 µ (sin(ϕ cos(ϕ 0 = 0 ( µ (sin(ϕ cos(ϕ 0 Adding and subtracting the two equations one can reach two better equations: One of the solutions (periodicity reads µ sin(ϕ + 5 = 0 µ cos(ϕ + 5 = 0 µ = 5 ϕ = arctan + π 3

14 Example 7 A linear time invariant MIMO system with transfer function is excited with the signal u(t = P (s = ( s+ s + s+0 s+ s + µ cos(ω t + ϕ µ cos(ω t + ϕ Because we bought a cheap signal generator, we cannot know exactly the constants µ, and ϕ, A friend of you just found out with some measurements, that the excitation frequency is ω = rad The cheap generator, cannot produce signals with magnitude of µ bigger than s 0, ie µ + µ 0 This works always at maximal power, ie at 0 Choose all possible responses of the system after infinite time 5 sin(t + 04 y (t = cos(t 5 sin(t + 04 y (t = cos( t sin(t y (t = sin(t cos(t + 04 y (t = cos(t cos(t + 04 y (t = 5 cos(t 0 sin(t + 4 y (t = sin(t

15 Solution 5 sin(t + 04 y (t = cos(t 5 sin(t + 04 y (t = cos( t sin(t y (t = sin(t cos(t + 04 y (t = cos(t cos(t + 04 y (t = 5 cos(t 0 sin(t + 4 y (t = sin(t + 34 Explanation We have to compute the singular values of the matrix P (j These are With what we have learned it follows σ max = 8305 σ min = σ min = 3863 ν 8305 = 0 σ max The first response has ν = 6 that is in this range The second response also has ν = 6 but the frequency in its second element changes and that isn t possible for linear systems The third response has ν = that is too small to be in the range The fourth response has ν = 36 that is too big to be in the range The fifth response has ν = 50 that is in the range The sixth response has ν = that is in the range 5

16 Example 8 A 3 linear time invariant MIMO system is excited with the input 3 sin(30 t u(t = 4 cos(30 t You have forgot your PC and you don t know the transfer function of the system Before coming to school, however, you have saved the Matlab plot of the singular values of the system on your phone (see Figure 3 Choose all the possible responses of the system Figure 3: Singular values behaviour 05 sin(30 t y (t = 05 cos(30 t 05 cos(30 t + 4 sin(30 t y (t = 3 cos(30 t cos(30 t + 0 sin(30 t y (t = 0 cos(30 t 0 cos(30 t + y (t = 0 4 cos(30 t cos(30 t + cos(30 t y (t = cos(30 t + 04 cos(30 t

17 Solution 05 sin(30 t y (t = 05 cos(30 t 05 cos(30 t + 4 sin(30 t y (t = 3 cos(30 t cos(30 t + 0 sin(30 t y (t = 0 cos(30 t 0 cos(30 t + 0 y (t = 4 cos(30 t cos(30 t + cos(30 t y (t = cos(30 t + 04 cos(30 t + 05 Explanation From the given input one can read µ = = 5 From the plot one can read at ω = 30 rad s σ min = 0 and σ max = It follows 5 σ min = 05 ν 5 = 5 σ max The first response has ν = 075 that is in the range The second response has ν = 9 that is too big to be in the range The third response has ν = 003 that is to small to be in the range The fourth response has ν = 0 that is in the range The fifth response has ν = that is in the range 7

Lecture 4: Analysis of MIMO Systems

Lecture 4: Analysis of MIMO Systems Lecture 4: Analysis of MIMO Systems Norms The concept of norm will be extremely useful for evaluating signals and systems quantitatively during this course In the following, we will present vector norms

More information

Singular Value Decomposition Analysis

Singular Value Decomposition Analysis Singular Value Decomposition Analysis Singular Value Decomposition Analysis Introduction Introduce a linear algebra tool: singular values of a matrix Motivation Why do we need singular values in MIMO control

More information

Full-State Feedback Design for a Multi-Input System

Full-State Feedback Design for a Multi-Input System Full-State Feedback Design for a Multi-Input System A. Introduction The open-loop system is described by the following state space model. x(t) = Ax(t)+Bu(t), y(t) =Cx(t)+Du(t) () 4 8.5 A =, B =.5.5, C

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

1 Singular Value Decomposition and Principal Component

1 Singular Value Decomposition and Principal Component Singular Value Decomposition and Principal Component Analysis In these lectures we discuss the SVD and the PCA, two of the most widely used tools in machine learning. Principal Component Analysis (PCA)

More information

Control Systems I. Lecture 5: Transfer Functions. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Control Systems I. Lecture 5: Transfer Functions. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich Control Systems I Lecture 5: Transfer Functions Readings: Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich October 20, 2017 E. Frazzoli (ETH) Lecture 5: Control Systems I 20/10/2017

More information

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich Control Systems I Lecture 6: Poles and Zeros Readings: Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich October 27, 2017 E. Frazzoli (ETH) Lecture 6: Control Systems I 27/10/2017

More information

Ordinary Differential Equations

Ordinary Differential Equations Ordinary Differential Equations for Engineers and Scientists Gregg Waterman Oregon Institute of Technology c 2017 Gregg Waterman This work is licensed under the Creative Commons Attribution 4.0 International

More information

Extreme Values and Positive/ Negative Definite Matrix Conditions

Extreme Values and Positive/ Negative Definite Matrix Conditions Extreme Values and Positive/ Negative Definite Matrix Conditions James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 8, 016 Outline 1

More information

Question: Total. Points:

Question: Total. Points: MATH 308 May 23, 2011 Final Exam Name: ID: Question: 1 2 3 4 5 6 7 8 9 Total Points: 0 20 20 20 20 20 20 20 20 160 Score: There are 9 problems on 9 pages in this exam (not counting the cover sheet). Make

More information

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Tutorials in Optimization. Richard Socher

Tutorials in Optimization. Richard Socher Tutorials in Optimization Richard Socher July 20, 2008 CONTENTS 1 Contents 1 Linear Algebra: Bilinear Form - A Simple Optimization Problem 2 1.1 Definitions........................................ 2 1.2

More information

Topic 15 Notes Jeremy Orloff

Topic 15 Notes Jeremy Orloff Topic 5 Notes Jeremy Orloff 5 Transpose, Inverse, Determinant 5. Goals. Know the definition and be able to compute the inverse of any square matrix using row operations. 2. Know the properties of inverses.

More information

Exercises * on Linear Algebra

Exercises * on Linear Algebra Exercises * on Linear Algebra Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 7 Contents Vector spaces 4. Definition...............................................

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

Name Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise.

Name Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise. Name Solutions Linear Algebra; Test 3 Throughout the test simplify all answers except where stated otherwise. 1) Find the following: (10 points) ( ) Or note that so the rows are linearly independent, so

More information

Some solutions of the written exam of January 27th, 2014

Some solutions of the written exam of January 27th, 2014 TEORIA DEI SISTEMI Systems Theory) Prof. C. Manes, Prof. A. Germani Some solutions of the written exam of January 7th, 0 Problem. Consider a feedback control system with unit feedback gain, with the following

More information

3. Mathematical Properties of MDOF Systems

3. Mathematical Properties of MDOF Systems 3. Mathematical Properties of MDOF Systems 3.1 The Generalized Eigenvalue Problem Recall that the natural frequencies ω and modes a are found from [ - ω 2 M + K ] a = 0 or K a = ω 2 M a Where M and K are

More information

EL2520 Control Theory and Practice

EL2520 Control Theory and Practice So far EL2520 Control Theory and Practice r Fr wu u G w z n Lecture 5: Multivariable systems -Fy Mikael Johansson School of Electrical Engineering KTH, Stockholm, Sweden SISO control revisited: Signal

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

21 Linear State-Space Representations

21 Linear State-Space Representations ME 132, Spring 25, UC Berkeley, A Packard 187 21 Linear State-Space Representations First, let s describe the most general type of dynamic system that we will consider/encounter in this class Systems may

More information

GQE ALGEBRA PROBLEMS

GQE ALGEBRA PROBLEMS GQE ALGEBRA PROBLEMS JAKOB STREIPEL Contents. Eigenthings 2. Norms, Inner Products, Orthogonality, and Such 6 3. Determinants, Inverses, and Linear (In)dependence 4. (Invariant) Subspaces 3 Throughout

More information

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions. The exam will cover Sections 6.-6.2 and 7.-7.4: True/False 30% Definitions 0% Computational 60% Skip Minors and Laplace Expansion in Section 6.2 and p. 304 (trajectories and phase portraits) in Section

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 5 Singular Value Decomposition We now reach an important Chapter in this course concerned with the Singular Value Decomposition of a matrix A. SVD, as it is commonly referred to, is one of the

More information

Numerical Linear Algebra

Numerical Linear Algebra Chapter 3 Numerical Linear Algebra We review some techniques used to solve Ax = b where A is an n n matrix, and x and b are n 1 vectors (column vectors). We then review eigenvalues and eigenvectors and

More information

Second-Order Homogeneous Linear Equations with Constant Coefficients

Second-Order Homogeneous Linear Equations with Constant Coefficients 15 Second-Order Homogeneous Linear Equations with Constant Coefficients A very important class of second-order homogeneous linear equations consists of those with constant coefficients; that is, those

More information

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003 MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003 1. True or False (28 points, 2 each) T or F If V is a vector space

More information

2.3 Oscillation. The harmonic oscillator equation is the differential equation. d 2 y dt 2 r y (r > 0). Its solutions have the form

2.3 Oscillation. The harmonic oscillator equation is the differential equation. d 2 y dt 2 r y (r > 0). Its solutions have the form 2. Oscillation So far, we have used differential equations to describe functions that grow or decay over time. The next most common behavior for a function is to oscillate, meaning that it increases and

More information

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one

More information

M.A.P. Matrix Algebra Procedures. by Mary Donovan, Adrienne Copeland, & Patrick Curry

M.A.P. Matrix Algebra Procedures. by Mary Donovan, Adrienne Copeland, & Patrick Curry M.A.P. Matrix Algebra Procedures by Mary Donovan, Adrienne Copeland, & Patrick Curry This document provides an easy to follow background and review of basic matrix definitions and algebra. Because population

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

Eigenvalues & Eigenvectors

Eigenvalues & Eigenvectors Eigenvalues & Eigenvectors Page 1 Eigenvalues are a very important concept in linear algebra, and one that comes up in other mathematics courses as well. The word eigen is German for inherent or characteristic,

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

MTH 464: Computational Linear Algebra

MTH 464: Computational Linear Algebra MTH 464: Computational Linear Algebra Lecture Outlines Exam 4 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University April 15, 2018 Linear Algebra (MTH 464)

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Solutions to the Homework Replaces Section 3.7, 3.8

Solutions to the Homework Replaces Section 3.7, 3.8 Solutions to the Homework Replaces Section 3.7, 3.8. Show that the period of motion of an undamped vibration of a mass hanging from a vertical spring is 2π L/g SOLUTION: With no damping, mu + ku = 0 has

More information

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of CHAPTER III APPLICATIONS The eigenvalues are λ =, An orthonormal basis of eigenvectors consists of, The eigenvalues are λ =, A basis of eigenvectors consists of, 4 which are not perpendicular However,

More information

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1 1 Rows first, columns second. Remember that. R then C. 1 A matrix is a set of real or complex numbers arranged in a rectangular array. They can be any size and shape (provided they are rectangular). A

More information

Tutorial 2 - Learning about the Discrete Fourier Transform

Tutorial 2 - Learning about the Discrete Fourier Transform Tutorial - Learning about the Discrete Fourier Transform This tutorial will be about the Discrete Fourier Transform basis, or the DFT basis in short. What is a basis? If we google define basis, we get:

More information

A Field Extension as a Vector Space

A Field Extension as a Vector Space Chapter 8 A Field Extension as a Vector Space In this chapter, we take a closer look at a finite extension from the point of view that is a vector space over. It is clear, for instance, that any is a linear

More information

Linear Algebra. and

Linear Algebra. and Instructions Please answer the six problems on your own paper. These are essay questions: you should write in complete sentences. 1. Are the two matrices 1 2 2 1 3 5 2 7 and 1 1 1 4 4 2 5 5 2 row equivalent?

More information

Abstract Algebra Study Sheet

Abstract Algebra Study Sheet Abstract Algebra Study Sheet This study sheet should serve as a guide to which sections of Artin will be most relevant to the final exam. When you study, you may find it productive to prioritize the definitions,

More information

Empirical Gramians and Balanced Truncation for Model Reduction of Nonlinear Systems

Empirical Gramians and Balanced Truncation for Model Reduction of Nonlinear Systems Empirical Gramians and Balanced Truncation for Model Reduction of Nonlinear Systems Antoni Ras Departament de Matemàtica Aplicada 4 Universitat Politècnica de Catalunya Lecture goals To review the basic

More information

Repeated Eigenvalues and Symmetric Matrices

Repeated Eigenvalues and Symmetric Matrices Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one

More information

LECTURE NOTE #10 PROF. ALAN YUILLE

LECTURE NOTE #10 PROF. ALAN YUILLE LECTURE NOTE #10 PROF. ALAN YUILLE 1. Principle Component Analysis (PCA) One way to deal with the curse of dimensionality is to project data down onto a space of low dimensions, see figure (1). Figure

More information

Eigenvectors and Hermitian Operators

Eigenvectors and Hermitian Operators 7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7)

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7) EEE582 Topical Outline A.A. Rodriguez Fall 2007 GWC 352, 965-3712 The following represents a detailed topical outline of the course. It attempts to highlight most of the key concepts to be covered and

More information

Solutions to the Homework Replaces Section 3.7, 3.8

Solutions to the Homework Replaces Section 3.7, 3.8 Solutions to the Homework Replaces Section 3.7, 3.8 1. Our text (p. 198) states that µ ω 0 = ( 1 γ2 4km ) 1/2 1 1 2 γ 2 4km How was this approximation made? (Hint: Linearize 1 x) SOLUTION: We linearize

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Chapter 6 Eigenvalues and Eigenvectors 6. Introduction to Eigenvalues Eigenvalues are the key to a system of n differential equations : dy=dt ay becomes dy=dt Ay. Now A is a matrix and y is a vector.y.t/;

More information

Exam. 135 minutes, 15 minutes reading time

Exam. 135 minutes, 15 minutes reading time Exam August 6, 208 Control Systems II (5-0590-00) Dr. Jacopo Tani Exam Exam Duration: 35 minutes, 5 minutes reading time Number of Problems: 35 Number of Points: 47 Permitted aids: 0 pages (5 sheets) A4.

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier Spring 0 CIS 55 Fundamentals of Linear Algebra and Optimization Jean Gallier Homework 5 & 6 + Project 3 & 4 Note: Problems B and B6 are for extra credit April 7 0; Due May 7 0 Problem B (0 pts) Let A be

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

Module 02 CPS Background: Linear Systems Preliminaries

Module 02 CPS Background: Linear Systems Preliminaries Module 02 CPS Background: Linear Systems Preliminaries Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html August

More information

2t t dt.. So the distance is (t2 +6) 3/2

2t t dt.. So the distance is (t2 +6) 3/2 Math 8, Solutions to Review for the Final Exam Question : The distance is 5 t t + dt To work that out, integrate by parts with u t +, so that t dt du The integral is t t + dt u du u 3/ (t +) 3/ So the

More information

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in 806 Problem Set 8 - Solutions Due Wednesday, 4 November 2007 at 4 pm in 2-06 08 03 Problem : 205+5+5+5 Consider the matrix A 02 07 a Check that A is a positive Markov matrix, and find its steady state

More information

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein Matrix Mathematics Theory, Facts, and Formulas with Application to Linear Systems Theory Dennis S. Bernstein PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD Contents Special Symbols xv Conventions, Notation,

More information

Large Scale Data Analysis Using Deep Learning

Large Scale Data Analysis Using Deep Learning Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset

More information

Solutions of the Sample Problems for the Third In-Class Exam Math 246, Fall 2017, Professor David Levermore

Solutions of the Sample Problems for the Third In-Class Exam Math 246, Fall 2017, Professor David Levermore Solutions of the Sample Problems for the Third In-Class Exam Math 6 Fall 07 Professor David Levermore Compute the Laplace transform of ft t e t ut from its definition Solution The definition of the Laplace

More information

Elementary Linear Algebra

Elementary Linear Algebra Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Topic 5 Notes Jeremy Orloff. 5 Homogeneous, linear, constant coefficient differential equations

Topic 5 Notes Jeremy Orloff. 5 Homogeneous, linear, constant coefficient differential equations Topic 5 Notes Jeremy Orloff 5 Homogeneous, linear, constant coefficient differential equations 5.1 Goals 1. Be able to solve homogeneous constant coefficient linear differential equations using the method

More information

EC Control Engineering Quiz II IIT Madras

EC Control Engineering Quiz II IIT Madras EC34 - Control Engineering Quiz II IIT Madras Linear algebra Find the eigenvalues and eigenvectors of A, A, A and A + 4I Find the eigenvalues and eigenvectors of the following matrices: (a) cos θ sin θ

More information

Linear Algebra - Part II

Linear Algebra - Part II Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T

More information

Differential Equations 2280 Sample Midterm Exam 3 with Solutions Exam Date: 24 April 2015 at 12:50pm

Differential Equations 2280 Sample Midterm Exam 3 with Solutions Exam Date: 24 April 2015 at 12:50pm Differential Equations 228 Sample Midterm Exam 3 with Solutions Exam Date: 24 April 25 at 2:5pm Instructions: This in-class exam is 5 minutes. No calculators, notes, tables or books. No answer check is

More information

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I) CS 106 Spring 2004 Homework 1 Elena Davidson 8 April 2004 Problem 1.1 Let B be a 4 4 matrix to which we apply the following operations: 1. double column 1, 2. halve row 3, 3. add row 3 to row 1, 4. interchange

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Frequency Response of Linear Time Invariant Systems

Frequency Response of Linear Time Invariant Systems ME 328, Spring 203, Prof. Rajamani, University of Minnesota Frequency Response of Linear Time Invariant Systems Complex Numbers: Recall that every complex number has a magnitude and a phase. Example: z

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Classify a transfer function to see which order or ramp it can follow and with which expected error.

Classify a transfer function to see which order or ramp it can follow and with which expected error. Dr. J. Tani, Prof. Dr. E. Frazzoli 5-059-00 Control Systems I (Autumn 208) Exercise Set 0 Topic: Specifications for Feedback Systems Discussion: 30.. 208 Learning objectives: The student can grizzi@ethz.ch,

More information

Differential Equations

Differential Equations This document was written and copyrighted by Paul Dawkins. Use of this document and its online version is governed by the Terms and Conditions of Use located at. The online version of this document is

More information

Basic Procedures for Common Problems

Basic Procedures for Common Problems Basic Procedures for Common Problems ECHE 550, Fall 2002 Steady State Multivariable Modeling and Control 1 Determine what variables are available to manipulate (inputs, u) and what variables are available

More information

Contents. 1 State-Space Linear Systems 5. 2 Linearization Causality, Time Invariance, and Linearity 31

Contents. 1 State-Space Linear Systems 5. 2 Linearization Causality, Time Invariance, and Linearity 31 Contents Preamble xiii Linear Systems I Basic Concepts 1 I System Representation 3 1 State-Space Linear Systems 5 1.1 State-Space Linear Systems 5 1.2 Block Diagrams 7 1.3 Exercises 11 2 Linearization

More information

Lecture 7 (Weeks 13-14)

Lecture 7 (Weeks 13-14) Lecture 7 (Weeks 13-14) Introduction to Multivariable Control (SP - Chapters 3 & 4) Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 7 (Weeks 13-14) p.

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our

More information

MATH 350: Introduction to Computational Mathematics

MATH 350: Introduction to Computational Mathematics MATH 350: Introduction to Computational Mathematics Chapter V: Least Squares Problems Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2011 fasshauer@iit.edu MATH

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Linear Algebra. Min Yan

Linear Algebra. Min Yan Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................

More information

Math Fall Final Exam

Math Fall Final Exam Math 104 - Fall 2008 - Final Exam Name: Student ID: Signature: Instructions: Print your name and student ID number, write your signature to indicate that you accept the honor code. During the test, you

More information

Deep Learning Book Notes Chapter 2: Linear Algebra

Deep Learning Book Notes Chapter 2: Linear Algebra Deep Learning Book Notes Chapter 2: Linear Algebra Compiled By: Abhinaba Bala, Dakshit Agrawal, Mohit Jain Section 2.1: Scalars, Vectors, Matrices and Tensors Scalar Single Number Lowercase names in italic

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

Systems of Linear ODEs

Systems of Linear ODEs P a g e 1 Systems of Linear ODEs Systems of ordinary differential equations can be solved in much the same way as discrete dynamical systems if the differential equations are linear. We will focus here

More information

Laboratory handout 1 Mathematical preliminaries

Laboratory handout 1 Mathematical preliminaries laboratory handouts, me 340 2 Laboratory handout 1 Mathematical preliminaries In this handout, an expression on the left of the symbol := is defined in terms of the expression on the right. In contrast,

More information

Diagonalizing Matrices

Diagonalizing Matrices Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B

More information

Lecture 10: Powers of Matrices, Difference Equations

Lecture 10: Powers of Matrices, Difference Equations Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each

More information

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL. Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is

More information

Designing Information Devices and Systems I Spring 2016 Official Lecture Notes Note 21

Designing Information Devices and Systems I Spring 2016 Official Lecture Notes Note 21 EECS 6A Designing Information Devices and Systems I Spring 26 Official Lecture Notes Note 2 Introduction In this lecture note, we will introduce the last topics of this semester, change of basis and diagonalization.

More information

Chemistry 365: Normal Mode Analysis David Ronis McGill University

Chemistry 365: Normal Mode Analysis David Ronis McGill University Chemistry 365: Normal Mode Analysis David Ronis McGill University 1. Quantum Mechanical Treatment Our starting point is the Schrodinger wave equation: Σ h 2 2 2m i N i=1 r 2 i + U( r 1,..., r N ) Ψ( r

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Sample Quantum Chemistry Exam 2 Solutions

Sample Quantum Chemistry Exam 2 Solutions Chemistry 46 Fall 7 Dr. Jean M. Standard Name SAMPE EXAM Sample Quantum Chemistry Exam Solutions.) ( points) Answer the following questions by selecting the correct answer from the choices provided. a.)

More information

Eigenvalues, Eigenvectors and the Jordan Form

Eigenvalues, Eigenvectors and the Jordan Form EE/ME 701: Advanced Linear Systems Eigenvalues, Eigenvectors and the Jordan Form Contents 1 Introduction 3 1.1 Review of basic facts about eigenvectors and eigenvalues..... 3 1.1.1 Looking at eigenvalues

More information

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2 Norwegian University of Science and Technology Department of Mathematical Sciences TMA445 Linear Methods Fall 07 Exercise set Please justify your answers! The most important part is how you arrive at an

More information