Lecture 3. Chapter 4: Elements of Linear System Theory. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.

Similar documents
Lecture 7 (Weeks 13-14)

Lecture 6. Chapter 8: Robust Stability and Performance Analysis for MIMO Systems. Eugenio Schuster.

Problem Set 5 Solutions 1

Control Systems. Laplace domain analysis

Kalman Decomposition B 2. z = T 1 x, where C = ( C. z + u (7) T 1, and. where B = T, and

Multivariable Control. Lecture 03. Description of Linear Time Invariant Systems. John T. Wen. September 7, 2006

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7)

EL2520 Control Theory and Practice

FEL3210 Multivariable Feedback Control

FEL3210 Multivariable Feedback Control

Lecture 5: Linear Systems. Transfer functions. Frequency Domain Analysis. Basic Control Design.

5. Observer-based Controller Design

Introduction. Performance and Robustness (Chapter 1) Advanced Control Systems Spring / 31

Control Systems. Frequency domain analysis. L. Lanari

6.241 Dynamic Systems and Control

Controls Problems for Qualifying Exam - Spring 2014

Multivariable Control. Lecture 05. Multivariable Poles and Zeros. John T. Wen. September 14, 2006

Chapter 20 Analysis of MIMO Control Loops

Robust Control 2 Controllability, Observability & Transfer Functions

Control Systems I. Lecture 7: Feedback and the Root Locus method. Readings: Jacopo Tani. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Robust Multivariable Control

Linear State Feedback Controller Design

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ

Linear System Theory

Control Systems I. Lecture 5: Transfer Functions. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Balanced Truncation 1

EEE582 Homework Problems

CONTROL DESIGN FOR SET POINT TRACKING

FEL3210 Multivariable Feedback Control

MEM 355 Performance Enhancement of Dynamical Systems MIMO Introduction

Singular Value Decomposition Analysis

Design Methods for Control Systems

Topic # Feedback Control. State-Space Systems Closed-loop control using estimators and regulators. Dynamics output feedback

Controllability, Observability, Full State Feedback, Observer Based Control

Time Response Analysis (Part II)

Richiami di Controlli Automatici

Control Systems I. Lecture 9: The Nyquist condition

Raktim Bhattacharya. . AERO 632: Design of Advance Flight Control System. Norms for Signals and Systems

MULTIVARIABLE ZEROS OF STATE-SPACE SYSTEMS

Professor Fearing EE C128 / ME C134 Problem Set 7 Solution Fall 2010 Jansen Sheng and Wenjie Chen, UC Berkeley

Linear Systems. Linear systems?!? (Roughly) Systems which obey properties of superposition Input u(t) output

Intro. Computer Control Systems: F8

Perspective. ECE 3640 Lecture 11 State-Space Analysis. To learn about state-space analysis for continuous and discrete-time. Objective: systems

Modeling and Analysis of Dynamic Systems

Lecture 12. Upcoming labs: Final Exam on 12/21/2015 (Monday)10:30-12:30

João P. Hespanha. January 16, 2009

Topic # Feedback Control Systems

Some solutions of the written exam of January 27th, 2014

ECEN 605 LINEAR SYSTEMS. Lecture 20 Characteristics of Feedback Control Systems II Feedback and Stability 1/27

State will have dimension 5. One possible choice is given by y and its derivatives up to y (4)

Control Systems I. Lecture 9: The Nyquist condition

Full-State Feedback Design for a Multi-Input System

Topic # Feedback Control Systems

Chapter 5. Standard LTI Feedback Optimization Setup. 5.1 The Canonical Setup

Lec 6: State Feedback, Controllability, Integral Action

6.241 Dynamic Systems and Control

An Internal Stability Example

Theory of Robust Control

Contents lecture 5. Automatic Control III. Summary of lecture 4 (II/II) Summary of lecture 4 (I/II) u y F r. Lecture 5 H 2 and H loop shaping

Control Systems Design

EE263: Introduction to Linear Dynamical Systems Review Session 6

CDS Solutions to the Midterm Exam

Control Systems Design, SC4026. SC4026 Fall 2010, dr. A. Abate, DCSC, TU Delft

Outline. Classical Control. Lecture 1

3.1 Overview 3.2 Process and control-loop interactions

10 Transfer Matrix Models

Problem Set 4 Solutions 1

Lecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster.

Systems and Control Theory Lecture Notes. Laura Giarré

STABILITY ANALYSIS. Asystemmaybe stable, neutrallyormarginallystable, or unstable. This can be illustrated using cones: Stable Neutral Unstable

Engineering Tripos Part IIB Nonlinear Systems and Control. Handout 4: Circle and Popov Criteria

Zeros and zero dynamics

Analysis of SISO Control Loops

Fall 線性系統 Linear Systems. Chapter 08 State Feedback & State Estimators (SISO) Feng-Li Lian. NTU-EE Sep07 Jan08

(Continued on next page)

Hankel Optimal Model Reduction 1

CDS Solutions to Final Exam

Nonlinear Control. Nonlinear Control Lecture # 18 Stability of Feedback Systems

ECE504: Lecture 9. D. Richard Brown III. Worcester Polytechnic Institute. 04-Nov-2008

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Problem Set 4 Solution 1

Control Systems. System response. L. Lanari

Problem 2 (Gaussian Elimination, Fundamental Spaces, Least Squares, Minimum Norm) Consider the following linear algebraic system of equations:

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77

Chapter Robust Performance and Introduction to the Structured Singular Value Function Introduction As discussed in Lecture 0, a process is better desc

Control Systems Design

Lecture 4: Analysis of MIMO Systems

Lecture 1: Feedback Control Loop

Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science : MULTIVARIABLE CONTROL SYSTEMS by A.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

Topic # Feedback Control

Control Systems Design, SC4026. SC4026 Fall 2009, dr. A. Abate, DCSC, TU Delft

9.5 The Transfer Function

Root Locus. Motivation Sketching Root Locus Examples. School of Mechanical Engineering Purdue University. ME375 Root Locus - 1

Solution via Laplace transform and matrix exponential

Part IB Paper 6: Information Engineering LINEAR SYSTEMS AND CONTROL. Glenn Vinnicombe HANDOUT 5. An Introduction to Feedback Control Systems

Some of the different forms of a signal, obtained by transformations, are shown in the figure. jwt e z. jwt z e

6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski. Solutions to Problem Set 1 1. Massachusetts Institute of Technology

Model Uncertainty and Robust Stability for Multivariable Systems

Transcription:

Lecture 3 Chapter 4: Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 3 p. 1/77

3.1 System Descriptions [4.1] Let f(u) be a liner operator, u 1 and u 2 two independent variables, and α 1 and α 2 two real scalars, then (3.1) f(α 1 u 1 +α 2 u 2 ) = α 1 f(u 1 )+α 2 f(u 2 ) 3.1.1 State Space Representation (3.2) ẋ(t) = Ax(t)+Bu(t) (3.3) or: [ẋ (3.4) y ] = y(t) = Cx(t)+Du(t) [ ][ ] [ A B x, G = s C D u A B C D ] Lecture 3 p. 2/77

System (3.2) (3.3) is not a unique description of the input-output behaviour of a linear system. Define new states q = Sx, i.e. x = S 1 q. Equivalent state-space realization (i.e., with same input-output behaviour): (3.5) A q = SAS 1, B q = SB, C q = CS 1, D q = D Dynamical system response x(t) for t t 0 (3.6) x(t) = e A(t t 0) x(t 0 )+ t t 0 e A(t τ) Bu(τ)dτ For a system with disturbances d and noise n: (3.7) (3.8) ẋ = Ax+Bu+Ed y = Cx+Du+Fd+n Lecture 3 p. 3/77

(3.9) e At = I + k=1 (At) k k! Let A q = SAS 1 = Λ = diag{λ i } be diagonal then e At = S 1 {diag(e λ it )}S where e λ it is the mode associated with eigenvalue λ i (A). Lecture 3 p. 4/77

3.1.2 Impulse response The impulse response matrix is { 0 t < 0 (3.10) g(t) = Ce At B +Dδ(t) t 0 With initial state x(0) = 0, the dynamic response to an arbitrary input u(t) is (3.11) y(t) = g(t) u(t) = t 0 g(t τ)u(τ)dτ where denotes the convolution operator. Lecture 3 p. 5/77

3.1.3 Transfer function representation (Laplace) (3.12) G(s) = 0 g(t)e st dt Laplace transforms of (3.2) and (3.3) become for x(0) = 0 sx(s) = Ax(s)+Bu(s) (3.13) x(s) = (si A) 1 Bu(s) y(s) = Cx(s)+Du(s) (3.14) y(s) = (C(sI A) 1 B +D) u(s) }{{} G(s) where G(s) is the transfer function matrix. Lecture 3 p. 6/77

Equivalently, (3.15) G(s) = 1 [Cadj(sI A)B +Ddet(sI A)] det(si A) From Appendix A.2.1 (3.16) det(si A) = n λ i (si A) = n (s λ i (A)) i=1 i=1 Lecture 3 p. 7/77

3.1.4 State-space realizations [4.1.6] Inverse system. For a square G(s) we have (3.17) G 1 s = [ A BD 1 C BD 1 D 1 C D 1 ] If D = 0, set D = εi. Be careful not to introduce RHP zeros with this modification. Improper systems cannot be represented in state space form. Lecture 3 p. 8/77

Realization of SISO transfer functions. (3.18) G(s) = β n 1s n 1 + +β 1 s+β 0 s n +a n 1 s n 1 + +a 1 s+a 0 y(s) = G(s)u(s) corresponds to (3.19) y n (t)+a n 1 y n 1 (t)+ +a 1 y (t)+a 0 y(t) = β n 1 u n 1 (t)+ +β 1 u (t)+β 0 u(t) where y n 1 (t) and u n 1 (t) represent n 1 th order derivatives, etc. Lecture 3 p. 9/77

Write this as y n = ( a n 1 y n 1 +β n 1 u n 1 )+ +( a 1 y +β 1 u )+( a 0 y +β 0 u) }{{} } {{ x n } x 2 n 1 With the notation ẋ x (t) = dx/dt, we get ẋ n = a 0 x 1 +β 0 u ẋ n 1 = a 1 x 1 +x n +β 1 u. ẋ 1 = a n 1 x 1 +x 2 +β n 1 u Lecture 3 p. 10/77

corresponding to the realization (observer canonical form) a n 1 1 0 0 0 β n 1 a n 2 0 1 0 0 β n 2 A =..... (3.20). a 2 0 0 1 0, B =. β 2 a 1 0 0 0 1 β 1 a 0 0 0 0 0 β 0 C = [1 0 0 0 0] Example: To obtain the state-space realization of G(s) = s a s+a, first bring out a constant term by division to get G(s) = s a s+a = 2a s+a +1 Thus D = 1. Then (3.20) yields A = a,b = 2a and C = 1. Lecture 3 p. 11/77

Example: Ideal PID-controller (3.21) K(s) = K c (1+ 1 τ I s +τ τ I τ D s 2 +τ I s+1 Ds) = K c τ I s Improper no realization Proper PID controller (3.22) K(s) = K c (1+ 1 τ I s + τ Ds ),ǫ 0.1 1+ǫτ D s Four common realizations (3.23) D = K c 1+ǫ ǫ Lecture 3 p. 12/77

1. Diagonalized form (Jordan canonical form) (3.24) A = [ 0 0 0 1 ǫτ D ], B = [ ] K c /τ I K c /(ǫ 2, C = [1 1] τ D ) 2. Observability canonical form (3.25) A = [ 0 1 0 1 ǫτ D ], B = [ γ1 γ 2 ], C = [1 0] (3.26) where γ 1 = K c ( 1 τ I 1 ǫ 2 τ D ), γ 2 = K c ǫ 3 τ 2 D Lecture 3 p. 13/77

3. Controllability canonical form (3.27) A = [ 0 0 1 1 ǫτ D ], B = [ ] 1, C = [γ 0 1 γ 2 ] 4. Observer canonical form in (3.20) (3.28) A = [ ] 1 ǫτ D 1, B = 0 0 [ β1 β 0 ], C = [1 0] (3.29) where β 0 = K c ǫτ I τ D, β 1 = K c ǫ 2 τ D τ I ǫ 2 τ I τ D Note: Transfer function offers more immediate insight. Lecture 3 p. 14/77

3.2 State controllability and state observability State controllability. The dynamical system ẋ = Ax+Bu, or equivalently the pair (A,B), is said to be state controllable if, for any initial state x(0) = x 0, any time t 1 > 0 and any final state x 1, there exists an input u(t) such that x(t 1 ) = x 1. Otherwise the system is said to be state uncontrollable. 1. The pair: (A,B) is state controllable if and only if the controllability matrix (3.30) C = [ B AB A 2 B A n 1 B ] has rank n (full row rank). Here n is the number of states. Lecture 3 p. 15/77

2. From (3.6) one can verify that for x(t 1 ) = x 1 (3.31) u(t) = B T e AT (t 1 t) W c (t 1 ) 1 (e At 1 x 0 x 1 ) where W c (t) is the Gramian matrix at time t, (3.32) W c (t) = t 0 e Aτ BB T e ATτ dτ Thus (A,B) is state controllable if and only if W c (t) has full rank (and thus is positive definite) for any t > 0. For a stable system (A is stable) check only P = W c ( ), (3.33) P = 0 e Aτ BB T e ATτ dτ Lecture 3 p. 16/77

P may also be obtained as the solution to the Lyapunov equation (3.34) AP +PA T = BB T 3. Let p i be the i th eigenvalue of A and q i the corresponding left eigenvector, qi HA = p iqi H. Then the system is state controllable if and only if qi HB 0, i. Example: A = [ 2 2 0 4 The transfer function has only one state. ], B = [ 1 1 ], C = [1 0], D = 0 G(s) = C(sI A) 1 B = 1 s+4 Lecture 3 p. 17/77

1. The controllability matrix has two linearly dependent rows: C = [B AB] = [ 1 4 1 4 2. The controllability Gramian is singular P = [ ] 0.125 0.125 0.125 0.125 3. p 1 = 2 and p 2 = 4, q 1 = [0.707 0.707] T and q 2 = [0 1] T. q H 1 B = 0, q H 2 B = 1 the first mode (eigenvalue) is not state controllable. ]. Lecture 3 p. 18/77

Controllability is a system-theoretic concept important for computation and realizations; but no practical insight: 1. It says nothing about how the states behave, e.g. it does not imply that one can hold (as t ) the states at a given value. 2. Required inputs may be very large with sudden changes. 3. Some states may be of no practical importance. 4. Existence result which provides no degree of controllability. Lecture 3 p. 19/77

Definition State observability. The dynamical system ẋ = Ax+Bu, y = Cx+Du (or the pair (A,C)) is said to be state observable if, for any time t 1 > 0, the initial state x(0) = x 0 can be determined from the time history of the input u(t) and the output y(t) in the interval [0,t 1 ]. Otherwise the system, or (A,C), is said to be state unobservable. 1. (A,C) is state observable if and only if the observability matrix C CA (3.35) O = has rank n (full column rank).. CA n 1 Lecture 3 p. 20/77

2. For a stable system the observability Gramian (3.36) Q = 0 e ATτ C T Ce Aτ dτ must have full rank n (and thus be positive definite). Q can also be found as the solution to the following Lyapunov equation (3.37) A T Q+QA = C T C 3. Let p i be the i th eigenvalue of A and t i the corresponding eigenvector, At i = p i t i. Then the system is state observable if and only if Ct i 0, i. Observability is a system theoretical concept but may not give practical insight. Lecture 3 p. 21/77

Kalman s decomposition By performing an appropriate coordinate transformation, any system can be reduce to a decomposition indicating the state that are or aren t controllable and/or observable. ẋ 1 ẋ 2 ẋ 3 ẋ 4 = y = [C 1 C 2 0 0] A 11 A 12 0 0 0 A 22 0 0 A 31 A 32 A 33 A 34 0 A 42 0 A 44 x 1 x 2 x 3 x 4 x 1 x 2 x 3 x 4 + B 1 0 B 3 0 u Lecture 3 p. 22/77

u s 1 x 1 y s 2 x 2 s 3 x 3 s 4 x 4 A state-space realization (A,B,C,D) of G(s) is said to be a minimal realization of G(s) if A has the smallest possible dimension (i.e., the fewest number of states). The smallest dimension is called the McMillan degree of G(s). A mode is hidden if it is not state controllable or observable and thus not appear in the minimial realization. The state-space realization is minimal if and only if (A,B) is controllable and (A, C) is observable. Lecture 3 p. 23/77

3.3 Stability [4.3] Definition A system is (internally) stable if none of its components contains hidden unstable modes and the injection of bounded external signals at any place in the system results in bounded output signals measured anywhere in the system. internal, i.e. all the states must be stable not only inputs/outputs. Definition State stabilizable, state detectable and hidden unstable modes. A system is state stabilizable if all unstable modes are state controllable. A system is state detectable if all unstable modes are state observable. A system with unstabilizable or undetectable modes is said to contain hidden unstable modes. Lecture 3 p. 24/77

3.4 Poles [4.4] Definition Poles. The poles p i of a system with state-space description (3.2) (3.3) are the eigenvalues λ i (A),i = 1,...,n of the matrix A. The pole or characteristic polynomial φ(s) is defined as φ(s) = det(si A) = n i=1 (s p i). Thus the poles are the roots of the characteristic equation (3.38) φ(s) = det(si A) = 0 Lecture 3 p. 25/77

3.4.1 Poles and stability Theorem 1 A linear dynamic system ẋ = Ax+Bu is stable if and only if all the poles are in the open left-half plane (LHP), that is, Re{λ i (A)} < 0, i. A matrix A with such a property is said to be stable or Hurwitz. 3.4.2 Poles from transfer functions Theorem 2 The pole polynomial φ(s) corresponding to a minimal realization of a system with transfer function G(s), is the least common denominator of all non-identically-zero minors of all orders of G(s). Lecture 3 p. 26/77

Example: (3.39) G(s) = 1 1.25(s+1)(s+2) [ ] s 1 s 6 s 2 The minors of order 1 are the four elements all have (s + 1)(s + 2) in the denominator. Minor of order 2 (3.40) detg(s) = (s 1)(s 2)+6s 1.25 2 (s+1) 2 (s+2) 2 = 1 1.25 2 (s+1)(s+2) Least common denominator of all the minors: (3.41) φ(s) = (s+1)(s+2) Minimal realization has two poles: s = 1; s = 2. Lecture 3 p. 27/77

Example: Consider the 2 3 system, with 3 inputs and 2 outputs, G(s) = 1 (s+1)(s+2)(s 1) (3.42) [ (s 1)(s+2) 0 (s 1) 2 (s+1)(s+2) (s 1)(s+1) (s 1)(s+1) ] Minors of order 1: (3.43) 1 s+1, s 1 (s+1)(s+2), 1 s 1, 1 s+2, 1 s+2 Minor of order 2 corresponding to the deletion of column 2: Lecture 3 p. 28/77

M 2 = (s 1)(s+2)(s 1)(s+1)+(s+1)(s+2)(s 1)2 ((s+1)(s+2)(s 1)) 2 = (3.44) = 2 (s+1)(s+2) The other two minors of order two are (3.45) M 1 = (s 1) (s+1)(s+2) 2, M 3 = 1 (s+1)(s+2) Least common denominator: (3.46) φ(s) = (s+1)(s+2) 2 (s 1) The system therefore has four poles: s = 1, s = 1 and two at s = 2. Note MIMO-poles are essentially the poles of the elements. A procedure is needed to determine multiplicity. Lecture 3 p. 29/77

3.5 Zeros [4.5] SISO system: zeros z i are the solutions to G(z i ) = 0. In general, zeros are values of s at which G(s) loses rank. Example [ Y = ] s+2 s 2 +7s+12 U Compute the response when u(t) = e 2t, y(0) = 0, ẏ(0) = 1 Lecture 3 p. 30/77

L{u(t)} = 1 s+2 s 2 Y sy(0) ẏ(0)+7sy 7y(0)+12Y = 1 s 2 Y + 7sY +12Y +1 = 1 Y(s) = 0 Assumption: g(s) has a zero z, g(z) = 0. Then for input u(t) = u 0 e zt the output is y(t) 0, t > 0. (with appropriate initial conditions) Lecture 3 p. 31/77

3.5.1 Zeros from state-space realizations [4.5.1] The state-space equations of a system can be written as: [ ] [ ] [ ] x 0 si A B (3.47) P(s) =, P(s) = u y C D The zeros are then the values s = z for which the polynomial system matrix P(s) loses rank, resulting in zero output for some non-zero input [ zi A B C D ][ x z u z ] = 0 Lecture 3 p. 32/77

The zeros are the solutions of [ zi A B det C D ] = 0 MATLAB zero = tzero(a,b,c,d) Lecture 3 p. 33/77

3.5.2 Zeros from transfer functions [4.5.2] Definition Zeros. z i is a zero of G(s) if the rank of G(z i ) is less than the normal rank of G(s). The zero polynomial is defined as z(s) = n z i=1 (s z i) where n z is the number of finite zeros of G(s). Theorem The zero polynomial z(s), corresponding to a minimal realization of the system, is the greatest common divisor of all the numerators of all order-r minors of G(s), where r is the normal rank of G(s), provided that these minors have been adjusted in such a way as to have the pole polynomial φ(s) as their denominators. Lecture 3 p. 34/77

Example (3.48) G(s) = 1 s+2 [ s 1 4 The normal rank of G(s) is 2. Minor of order 2: detg(s) = 2(s 1)2 18 (s+2) 2 Pole polynomial: φ(s) = s+2. Zero polynomial: z(s) = s 4. 4.5 2(s 1) = 2 s 4 s+2. Note Multivariable zeros have no relationship with the zeros of the transfer function elements. ] Lecture 3 p. 35/77

Example (3.49) G(s) = 1 1.25(s+1)(s+2) Minor of order 2 is the determinant (3.50) [ ] s 1 s 6 s 2 detg(s) = (s 1)(s 2)+6s 1.25 2 (s+1) 2 (s+2) 2 = 1 1.25 2 (s+1)(s+2) φ(s) = 1.25 2 (s+1)(s+2) Zero polynomial = numerator of (3.50) no multivariable zeros. Lecture 3 p. 36/77

Example (3.51) G(s) = [ s 1 s+1 The normal rank of G(s) is 1 s 2 s+2 no value of s for which G(s) = 0 G(s) has no zeros. ] Lecture 3 p. 37/77

3.6 More on poles and zeros[4.6] 3.6.1 Directions of poles and zeros Let G(s) = C(sI A) 1 B +D. Zero directions. Let G(s) have a zero at s = z. Then G(s) loses rank at s = z, and there exist non-zero vectors u z and y z such that (3.52) G(z)u z = 0, yz H G(z) = 0 u z = input zero direction y z = output zero direction y z gives information about which output (or combination of outputs) may be difficult to control. Lecture 3 p. 38/77

Pole directions. Let G(s) have a pole at s = p. Then G(p) is infinite, and we may write (3.53) G(p)u p =, y H p G(p) = u p = input pole direction y p = output pole direction. SVD: G(z/p) = UΣV H u z = last column in V, y z = last column of U (corresponding to the zero singular value of G(z)) u p = first column in V, y p = first column of U (corresponding to the infinite singular value of G(p)) Lecture 3 p. 39/77

Example Plant in (3.48) has a RHP-zero at z = 4 and a LHP-pole at p = 2. [ ] 3 4 (3.54) G(z) = G(4) = 1 6 4.5 6 = 1 [ 0.55 0.83 6 0.83 0.55 u z = [ ] 0.80 0.60 For pole directions consider (3.55) ][ 9.01 0 0 0 y z = [ ] 0.83 0.55 ][ 0.6 0.8 0.8 0.6 G(p+ǫ) = G( 2+ǫ) = 1 ǫ 2 [ 3+ǫ 4 4.5 2( 3+ǫ) ] ] H Lecture 3 p. 40/77

The SVD as ǫ 0 yields G( 2+ǫ) = 1 [ ][ ][ ] H 0.55 0.83 9.01 0 0.6 0.8 ǫ 2 0.83 0.55 0 0 0.8 0.6 (3.56) u p = [ 0.60 0.80 ] y p = [ ] 0.55 0.83 Note Locations of poles and zeros are independent of input and output scalings, their directions are not. Lecture 3 p. 41/77

3.6.2 Remarks on poles and zeros [4.6.2] 1. For square systems the poles and zeros of G(s) are essentially the poles and zeros of detg(s). This fails when zero and pole in different parts of the system cancel when forming detg(s). [ ] (s+2)/(s+1) 0 (3.57) G(s) = 0 (s+1)/(s+2) detg(s) = 1, although the system obviously has poles at 1 and 2 and (multivariable) zeros at 1 and 2. 2. System (3.57) has poles and zeros at the same locations (at 1 and 2). Their directions are different. They do not cancel or otherwise interact. Lecture 3 p. 42/77

3. There are no zeros if the outputs contain direct information about all the states; that is, if from y we can directly obtain x (e.g. C = I and D = 0); 4. Zeros usually appear when there are fewer inputs or outputs than states 5. Moving poles. (a) feedback control (G(I +KG) 1 ) moves the poles, (b) series compensation (GK, feedforward control) can cancel poles in G by placing zeros in K (but not move them), and (c) parallel compensation (G+K) cannot affect the poles in G. Lecture 3 p. 43/77

6. Moving zeros. (a) With feedback, the zeros of G(I +KG) 1 are the zeros of G plus the poles of K., i.e. the zeros are unaffected by feedback. (b) Series compensation can counter the effect of zeros in G by placing poles in K to cancel them, but cancellations are not possible for RHP-zeros due to internal stability (see Section 3.7). (c) The only way to move zeros is by parallel compensation, y = (G+K)u, which, if y is a physical output, can only be accomplished by adding an extra input (actuator). Lecture 3 p. 44/77

Example. Effect of feedback on poles and zeros. SISO plant G(s) = z(s)/φ(s) and K(s) = k. (3.58) T(s) = L(s) 1+L(s) = kg(s) 1+kG(s) = kz(s) φ(s)+kz(s) = kz cl(s) φ cl (s) Note the following: 1. Zero polynomial: z cl (s) = z(s) zero locations are unchanged. 2. Pole locations are changed by feedback. For example, (3.59) k 0 φ cl (s) φ(s) (3.60) k φ cl (s) z(s). z(s) where roots of z(s) move with k to infinity (complex pattern) (cf. root locus) Lecture 3 p. 45/77

3.7 Internal stability of feedback systems [4.7] Note: Checking the pole of S or T is not sufficient to determine internal stability Example (Figure 1). In forming L = GK we cancel the term (s 1) (a RHP pole-zero cancellation) to obtain (3.61) L = GK = k s, and S = (I +L) 1 = s s+k S(s) is stable, i.e. transfer function from d y to y is stable. However, the transfer function from d y to u is unstable: (3.62) u = K(I +GK) 1 d y = k(s+1) (s 1)(s+k) d y Lecture 3 p. 46/77

r d u K G + + u + y - k(s+1) s 1 s(s 1) + s+1 + d y Figure 1: Internally unstable system Lecture 3 p. 47/77

y K + d u + + d y + G u Figure 2: Block diagram used to check internal stability of feedback system For internal stability consider (3.63) (3.64) u = (I +KG) 1 d u K(I +GK) 1 d y y = G(I +KG) 1 d u +(I +GK) 1 d y Lecture 3 p. 48/77

Theorem 4.4 The feedback system in Figure 2 is internally stable if and only if all four closed-loop transfer matrices in (3.63) and (3.64) are stable. Theorem 4.5 Assume there are no RHP pole-zero cancellations between G(s) and K(s). Then the feedback system in Figure 2 is internally stable if and only if one of the four closed-loop transfer function matrices in (3.63) and (3.64) is stable. Lecture 3 p. 49/77

Implications of the internal stability requirement 1. If G(s) has a RHP-zero at z, then L = GK, T = GK(I +GK) 1, SG = (I +GK) 1 G, L I = KG and T I = KG(I +KG) 1 will each have a RHP-zero at z. 2. If G(s) has a RHP-pole at p, then L = GK and L I = KG also have a RHP-pole at p, while S = (I +GK) 1,KS = K(I +GK) 1 and S I = (I +KG) 1 have a RHP-zero at p. Lecture 3 p. 50/77

Exercise: Interpolation constraints. Prove for SISO feedback systems when the plant G(s) has a RHP-zero z or a RHP-pole p: (3.65) G(z) = 0 L(z) = 0 T(z) = 0,S(z) = 1 (3.66)G 1 (p) = 0 L(p) = T(p) = 1,S(p) = 0 Remark Perfect control implies S 0 and T 1. RHP-zero perfect control impossible. RHP-pole perfect control possible. RHP-poles cause problems when tight (high gain) control is not possible. Lecture 3 p. 51/77

3.8 Stabilizing controllers [4.8] Stable plants Lemma For a stable plant G(s) the negative feedback system in Figure 2 is internally stable if and only if Q = K(I +GK) 1 is stable. Proof: The four transfer functions in (3.63) and (3.64) are (3.67) (3.68) (3.69) K(I +GK) 1 = Q (I +GK) 1 = I GQ (I +KG) 1 = I QG (3.70) G(I +KG) 1 = G(I QG) which are clearly all stable if and only if G and Q are stable. Lecture 3 p. 52/77

Consequences: All stabilizing negative feedback controllers for the stable plant G(s) are given by (3.71) K = (I QG) 1 Q = Q(I GQ) 1 where the parameter Q is any stable transfer function matrix. (Identical to the internal model control (IMC) parameterization of stabilizing controllers.) Lecture 3 p. 53/77

r + K plant d y Q G - + + y G model - + Figure 3: The internal model control (IMC) structure Lecture 3 p. 54/77

3.9 Stability analysis in the frequency domain Generalization of Nyquist s stability test for SISO systems. 3.9.1 Open and closed-loop characteristic polynomials r + - L y Figure 4: Negative feedback system Lecture 3 p. 55/77

Open Loop: (3.72) L(s) = C ol (si A ol ) 1 B ol +D ol Poles of L(s) are the roots of the open-loop characteristic polynomial (3.73) φ ol (s) = det(si A ol ) Assume no RHP pole-zero cancellations between G(s) and K(s). Then from Theorem 4.5 internal stability of the closed-loop system is equivalent to the stability of S(s) = (I +L(s)) 1. The realization of S(s) can be derived as follow: (3.74) (3.75) ẋ = A ol x+b ol (r y) e = r y = r C ol x D ol (r y) Lecture 3 p. 56/77

or (3.76) r y = (I +D ol ) 1 (r C ol x) and (3.77) ẋ = (A ol B ol (I +D ol ) 1 C ol )x+b ol (I +D ol ) 1 r Therefore the state matrix of S(s) is: (3.78) A cl = A ol B ol (I +D ol ) 1 C ol And the closed-loop characteristic polynomial is (3.79) φ cl (s) = det(si A cl ) = det(si A ol +B ol (I+D ol ) 1 C ol ) Lecture 3 p. 57/77

Relationship between characteristic polynomials From (3.72) we get (3.80) det(i +L(s)) = det(i +C ol (si A ol ) 1 B ol +D ol ) Schur s formula yields (with A 11 = I +D ol,a 12 = C ol,a 22 = si A ol,a 21 = B ol ) (3.81) det(i +L(s)) = φ cl(s) φ ol (s) c where c = det(i +D ol ) is a constant (cf. SISO result from RSI). Lecture 3 p. 58/77

Side calculation: [ ] I +D det ol C ol B ol si A ol ] = det[i +D ol ]det [si A ol +B ol (I +D ol ) 1 C ol ] = det[si A ol ]det [I +D ol +C ol (si A ol ) 1 B ol Schur s formula: det [ A 11 A 12 A 21 A 22 ] = det(a 11 ) det(a 22 A 21 A 1 11 A 12) = det(a 22 ) det(a 11 A 12 A 1 22 A 21) Lecture 3 p. 59/77

3.9.2 MIMO Nyquist stability criteria Theorem: Generalized (MIMO) Nyquist theorem. Let P ol denote the number of open-loop unstable poles in L(s). The closed-loop system with loop transfer function L(s) and negative feedback is stable if and only if the Nyquist plot of det(i +L(s)) i) makes P ol anti-clockwise encirclements of the origin, and ii) does not pass through the origin. Note By Nyquist plot of det(i +L(s)) we mean the image of det(i +L(s)) as s goes clockwise around the Nyquist D-contour. Lecture 3 p. 60/77

3.9.4 Small gain theorem [4.9.4] (3.82) ρ(l(jω)) = max i λ i (L(jω)) Theorem: Spectral radius stability condition. Consider a system with a stable loop transfer function L(s). Then the closed-loop system is stable if (3.83) ρ(l(jω)) = max i λ i (L(jω)) < 1 ω Lecture 3 p. 61/77

Proof: Assume the system is unstable. Therefore det(i +L(s)) encircles the origin, and there is an eigenvalue, λ i (L(jω)) which is larger than 1 at some frequency. If det(i +L(s)) does encircle the origin, then there must exists a gain ǫ (0,1] and a frequency ω such that (3.84) or (3.85) (3.86) det(i +ǫl(jω )) = 0 λ i (I +ǫl(jω )) = 0 i = 1+ǫλ i (L(jω )) = 0 for some i (3.87) = λ i (L(jω )) = 1 ǫ for some i (3.88) (3.89) λ i (L(jω )) 1 for some i = ρ(l(jω )) 1 Lecture 3 p. 62/77

Interpretation: If the system gain is less than 1 in all directions (all eigenvalues) and for all frequencies ( ω), then all signal deviations will eventually die out, and the system is stable. Spectral radius theorem is conservative because phase information is not considered. Small Gain Theorem. Consider a system with a stable loop transfer function L(s). Then the closed-loop system is stable if (3.90) L(jω) < 1 ω where L denotes any matrix norm satisfying AB A B, for example the singular value σ(l). Note The small gain theorem is generally more conservative than the spectral radius condition in (3.83). Lecture 3 p. 63/77

3.10 System norms [4.10] w G z Figure 5: System G Figure 5: System with stable transfer function matrix G(s) and impulse response matrix g(t). Question: given information about the allowed input signals w(t), how large can the outputs z(t) become? We use the 2-norm, (3.91) z(t) 2 = i z i (τ) 2 dτ Lecture 3 p. 64/77

and consider three inputs: 1. w(t) is a series of unit impulses. 2. w(t) is any signal satisfying w(t) 2 = 1. 3. w(t) is any signal satisfying w(t) 2 = 1, but w(t) = 0 for t 0, and we only measure z(t) for t 0. The relevant system norms in the three cases are the H 2, H, and Hankel norms, respectively. Lecture 3 p. 65/77

3.10.1 H 2 norm [4.10.1] G(s) strictly proper. For the H 2 norm we use the Frobenius norm spatially (for the matrix) and integrate over frequency, i.e., (3.92) G(s) 2 = 1 2π tr(g(jω) H G(jω)) dω }{{} G(jω) 2 F = ij G ij(jω) 2 G(s) must be strictly proper, otherwise the H 2 norm is infinite. By Parseval s theorem, (3.92) is equal to the H 2 norm of the impulse response (3.93) G(s) 2 = g(t) 2 = 0 tr(g T (τ)g(τ)) dτ }{{} g(τ) 2 F = ij g ij(τ) 2 Lecture 3 p. 66/77

Note that G(s) and g(t) are dynamic systems while G(jω) and g(τ) are constant matrices (for given value of ω or τ). We can change the order of integration and summation in (3.93) to get (3.94) G(s) 2 = g(t) 2 = g ij (τ) 2 dτ ij where g ij (t) is the ij th element of the impulse response matrix, g(t). Thus H 2 norm can be interpreted as the 2-norm output resulting from applying unit impulses δ j (t) to each input, one after another (allowing the output to settle to zero before applying an impulse to the next input). Thus G(s) 2 = m i=1 z i(t) 2 2 where z i (t) is the output vector resulting from applying a unit impulse δ i (t) to the i th input. 0 Lecture 3 p. 67/77

Numerical computations of theh 2 norm. Consider G(s) = C(sI A) 1 B. Then (3.95) G(s) 2 = tr(b T QB) or G(s) 2 = tr(cpc T ) where Q = observability Gramian and P = controllability Gramian Lecture 3 p. 68/77

3.10.2 H norm [4.10.2] G(s) proper. For the H norm we use the singular value (induced 2-norm) spatially (for the matrix) and pick out the peak value as a function of frequency (3.96) G(s) = max ω σ(g(jω)) The H norm is the peak of the transfer function magnitude. Lecture 3 p. 69/77

Time domain performance interpretations of theh norm. Worst-case steady-state gain for sinusoidal inputs at any frequency. Induced (worst-case) 2-norm in the time domain: (3.97) G(s) = max w(t) 0 z(t) 2 w(t) 2 = max w(t) 2 =1 z(t) 2 (In essence, (3.97) arises because the worst input signal w(t) is a sinusoid with frequency ω and a direction which gives σ(g(jω )) as the maximum gain.) Lecture 3 p. 70/77

Numerical computation of theh norm. Consider G(s) = C(sI A) 1 B +D H norm is the smallest value of γ such that the Hamiltonian matrix H has no eigenvalues on the imaginary axis, where [ ] A+BR H = 1 D T C BR 1 B T (3.98) C T (I +DR 1 D T )C (A+BR 1 D T C) T and R = γ 2 I D T D Lecture 3 p. 71/77

3.10.3 Difference between theh 2 andh norms Frobenius norm in terms of singular values (3.99) G(s) 2 = 1 2π i σ 2 i (G(jω))dω Thus when optimizing performance in terms of the different norms: H : push down peak of largest singular value. H 2 : push down whole thing (all singular values over all frequencies). Lecture 3 p. 72/77

Example (3.100) H 2 norm: G(s) = 1 s+a G(s) 2 = ( 1 2π = ( 1 2πa G(jω) 2 dω) 1 2 }{{} 1 ω 2 +a 2 [ tan 1 ( ω ] 1 a ) )1 2 = 2a Alternatively: Consider the impulse response ( ) 1 (3.101) g(t) = L 1 = e at,t 0 s+a Lecture 3 p. 73/77

to get (3.102) g(t) 2 = (e at ) 2 dt = as expected from Parseval s theorem. H norm: 0 1 2a (3.103) G(s) = max ω G(jω) = max ω 1 (ω 2 +a 2 ) 1 2 = 1 a Lecture 3 p. 74/77

Example There is no general relationship between the H 2 and H norms. (3.104) f 1 (s) = 1 ǫs+1, f 2(s) = ǫs s 2 +ǫs+1 (3.105) f 1 = 1 f 1 2 = f 2 = 1 f 2 2 = 0 Why is theh norm so popular? In robust control convenient for representing unstructured model uncertainty, and because it satisfies the multiplicative property: (3.106) A(s)B(s) A(s) B(s) What is wrong with theh 2 norm? It is not an induced norm and does not satisfy the multiplicative property. Lecture 3 p. 75/77

Example Consider again G(s) = 1/(s+a) in (3.100), for which G(s) 2 = 1/2a. G(s)G(s) 2 = 0 L 1 1 [( s+a )2 ] }{{} te at 2 (3.107) = 1 a 1 2a = 1 a G(s) 2 2 Lecture 3 p. 76/77

for a < 1, (3.108) G(s)G(s) 2 > G(s) 2 G(s) 2 which does not satisfy the multiplicative property. H norm does satisfy the multiplicative property G(s)G(s) = 1 a 2 = G(s) G(s) Lecture 3 p. 77/77