EE363 Review Session 1: LQR, Controllability and Observability

Similar documents
Problem 1 Cost of an Infinite Horizon LQR

EE363 homework 2 solutions

Lecture 4 Continuous time linear quadratic regulator

EE221A Linear System Theory Final Exam

Announcements Monday, September 18

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Lecture 5 Linear Quadratic Stochastic Control

Lecture 2 and 3: Controllability of DT-LTI systems

Announcements September 19

Announcements Wednesday, August 30

Announcements Wednesday, August 30

Lecture 19 Observability and state estimation

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

Observability and state estimation

Announcements Wednesday, November 15

Announcements Wednesday, September 05

Matrix-Vector Products and the Matrix Equation Ax = b

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation

Announcements Monday, November 13

CS173 Lecture B, November 3, 2015

Announcements Wednesday, September 20

1 Similarity transform 2. 2 Controllability The PBH test for controllability Observability The PBH test for observability...

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

Announcements Wednesday, November 01

Calculus II. Monday, March 13th. WebAssign 7 due Friday March 17 Problem Set 6 due Wednesday March 15 Midterm 2 is Monday March 20

Announcements Monday, September 17

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

A Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley

Announcements Monday, November 26

1. Type your solutions. This homework is mainly a programming assignment.

CDS 110b: Lecture 2-1 Linear Quadratic Regulators

Homework Solution # 3

Announcements Wednesday, October 04

Robotics. Control Theory. Marc Toussaint U Stuttgart

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Announcements Monday, November 13

Math 220 F11 Lecture Notes

Announcements Wednesday, November 01

Announcements Monday, October 02

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

Course Outline. Higher Order Poles: Example. Higher Order Poles. Amme 3500 : System Dynamics & Control. State Space Design. 1 G(s) = s(s + 2)(s +10)

Lecture 4 and 5 Controllability and Observability: Kalman decompositions

First Derivative Test

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

Unit D Student Success Sheet (SSS) Factoring (section 2.1)

Two-Sample Inference for Proportions and Inference for Linear Regression

Announcements Monday, November 19

Announcements Monday, September 25

4.5 Integration of Rational Functions by Partial Fractions

Controllability. Chapter Reachable States. This chapter develops the fundamental results about controllability and pole assignment.

Updated: January 16, 2016 Calculus II 7.4. Math 230. Calculus II. Brian Veitch Fall 2015 Northern Illinois University

6.241 Dynamic Systems and Control

4F3 - Predictive Control

Section Instructors: by now you should be scheduled into one of the following Sections:

Solution for Homework 5

12/11/2013 BW 12/11 BW 12/11 BW 12/11 BW 12/11. 5 th Period: Enter silently as ALWAYS

Lecture 10: Powers of Matrices, Difference Equations

Math 1553 Introduction to Linear Algebra. School of Mathematics Georgia Institute of Technology

Linear-Quadratic Optimal Control: Full-State Feedback

Module 02 CPS Background: Linear Systems Preliminaries

Linear System Theory

Announcements Wednesday, September 06

Module 03 Linear Systems Theory: Necessary Background

EE263 homework 9 solutions

Linear algebra and differential equations (Math 54): Lecture 10

Physics 1A, Lecture 3: One Dimensional Kinema:cs Summer Session 1, 2011

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

5. Observer-based Controller Design

Problem Set 1. This week. Please read all of Chapter 1 in the Strauss text.

Lecture 9: Discrete-Time Linear Quadratic Regulator Finite-Horizon Case

Review Solutions for Exam 1

Partial Fractions. June 27, In this section, we will learn to integrate another class of functions: the rational functions.

Welcome to Math 102 Section 102

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

Name Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise.

MATH 310, REVIEW SHEET 2

Announcements Wednesday, October 25

We ll start today by learning how to change a decimal to a fraction on our calculator! Then we will pick up our Unit 1-5 Review where we left off!

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

Z i Q ij Z j. J(x, φ; U) = X T φ(t ) 2 h + where h k k, H(t) k k and R(t) r r are nonnegative definite matrices (R(t) is uniformly in t nonsingular).

Linear Systems. Manfred Morari Melanie Zeilinger. Institut für Automatik, ETH Zürich Institute for Dynamic Systems and Control, ETH Zürich

Linear Quadratic Regulator (LQR) Design I

Math 416, Spring 2010 Matrix multiplication; subspaces February 2, 2010 MATRIX MULTIPLICATION; SUBSPACES. 1. Announcements

Algebra 2 Summer Packet

MAT01B1: Integration of Rational Functions by Partial Fractions

Announcements Monday, November 26

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Math 369 Exam #2 Practice Problem Solutions

What if the characteristic equation has a double root?

Announcements Wednesday, September 27

MAT188H1S LINEAR ALGEBRA: Course Information as of February 2, Calendar Description:

Math 21b: Linear Algebra Spring 2018

EE C128 / ME C134 Feedback Control Systems

CALCULUS 2 FIVES SHEET F=Fun at home I=Incunabula V=Variety E=Endeavors S=Specimens Thursday, 12/3

Suppose that we have a specific single stage dynamic system governed by the following equation:

HW2 - Due 01/30. Each answer must be mathematically justified. Don t forget your name.

Transcription:

EE363 Review Session : LQR, Controllability and Observability In this review session we ll work through a variation on LQR in which we add an input smoothness cost, in addition to the usual penalties on the state and input We will also (briefly) review the concepts of controllability and observability from EE263 If you haven t seen this before, or, if you don t remember them, please read through EE263 lectures 8 and 9 As always, the TAs are more than happy to help if you have any questions Announcements: TA office hours: Tuesday 3-5pm, Packard 07, Wednesday 7-9pm, Packard 277, Thursday 4-6pm, Packard 277 Homework is due on Fridays Homework is graded on a scale of 0-0 Representing quadratic functions as quadratic forms Let s first go over a method for representing quadratic functions that you might find useful for the homework This representation can often simplify the algebra for LQR problems, especially when there are linear, as well as quadratic terms in the costs Consider the following quadratic function in u and v, f(u, v) = u T Fu + v T Gv + 2u T Sv + 2f T u + 2g T v + s, where F > 0 and G > 0 We can write this as a pretty, symmetric quadratic form, f(u, v) = u v T F S f S T G g f T g T s u v

Now we can apply many of the results we know for quadratic forms to quadratic functions For example, suppose we want to minimze f(u, v), over v We know (see homework, problem 3), y min y x T Q Q 2 y Q 22 x Q T 2 = x T (Q 22 Q T 2 Q Q 2)x Applying this result to our quadratic form representation of f(u, v), we get v min f(u, v) = v T ( G g g T s S T f T F S f ) v Example: Represent the following quadratic function in w and z, f(z, w) = z T Qz + 2q T z + w T Rw + (Az + Bw) T P(Az + Bw) + 2p T (Az + Bw) + s, as a quadratic form (This quadratic function arises, for example, in the Bellman equation for an LQR problem) Solution We can write f(z, w) as f(z, w) = w z R + B T PB A T PB B T p B T PA Q + A T PA q + A T p p T B q T + p T A s T This means, for example, that we can derive the Riccati recursion for this problem by directly applying our Shur complement minimization formula above This gives, min w f(z, w) = LQR with smoothness penalty T ( z Q + A T PA q + A T p q T + p T A s B T PA p T (R + B T PB) A B T PB B T p ) z Consider the following linear dynamical system x t+ = Ax t + Bu t, x 0 = x init In conventional LQR we choose, u,, so that x 0, x, and, u, are small In this problem we d like to add an additional penalty on the smoothness of the input This means that we want the differences, u, u 2 u, to be small as well To do this, we define the following cost function, J = N τ=0 N x T τ Qx τ + u T τ Ru τ + (u τ u τ ) T R(uτ u τ ) + x T N Q fx N, τ=0 2 w z

where Q = Q T 0, R = R T > 0, R = RT > 0, and u = 0 Notice that J is still a quadratic function of,,u N, although the form of the quadratic is more complicated, compared with conventional LQR This means that we can solve for,,u N as a giant least squares problem and, if we use appropriate numerical linear algebra methods, we can solve this very fast (no slower than a Riccati recursion) Another way of solving this problem is by transforming it into a standard LQR problem Let s define w t = u t u t, t = 0,,N (with u = 0) Then we can write xt+ x t+ = where We can write J as where x t = u t xt J = u t Q = N τ=0 = = Axt + Bu t A B 0 I u t xt = Ã x t + Bw t, A B, Ã = 0 I 0 + + u t w t B I, B = B I x T τ Q x τ + w T τ Rw τ + x T N Q f x N, Q 0 0 R Qf 0, Qf = 0 0 So we see that by defining a new state x t that consists of the original state x t, and the previous input u t, we have transformed this problem into a conventional LQR problem This means that we can apply the standard LQR Riccati equation to find the optimal w 0,,w N, and hence,,u N I should point out here, that there nothing profound about anything we have done here State augmentation (increasing the size of the state), is a common trick to deal with situations where the variables are coupled across several time periods There is also no reason why we should compare every problem we encounter with the conventional LQR problem It s much more important to learn and understand dynamic programming and the Bellman equation, rather than memorizing LQR formulas you can always look these up, or re-derive them if you need to Controllability Now let s briefly review the concepts of controllability and observability for the discrete-time linear dynamical system x t+ = Ax t + Bu t, t = 0,,, 3 w t

where x t R n and u t R m Similar concepts exist for continuous time systems for more details, please refer to EE263 lectures 8 and 9 We say that a state z is reachable in t steps, if we can find an input sequence, u,,u t that steers the state from x 0 = 0 to x t = z We can write x t = B AB A t B u t, and so the set of states reachable in t steps is simply R(C t ), where C t = B AB A t B By the Cayley-Hamilton theorem (EE263 lecture 2), we can express each A k for k n as a linear combination of A 0,,A n, and so for t n, R(C t ) = R(C n ) This means that any state that can be reached, can be reached by t = n, and similarly, any state that cannot be reached by t = n is not reachable (in any number of steps) We call C = C n the controllability matrix The set of reachable states is exactly R(C) We say that the linear dynamical system is reachable or controllable if all the states are reachable (ie, R(C) = R n ) Thus, the system is controllable if and only if Rank(C) = n We will often say that the pair of matrices (A, B) is controllable this means that the controllability matrix formed from A and B is full rank Example: Suppose the system is controllable Is there always an input that steers the state from an initial state x init at time t = 0 to some desired state x final? Solution Yes We can write x t = A t x 0 + B AB A t B Thus, we can find an input that steers the state from x init to x final in time t, if and only if x final A t x init R(C t ) Since the system is controllable, we know that Rank(C n ) = n, which implies that R(C n ) = R n This means that for any x init and any x final, we must have x final A n x init R(C n ) Thus, if the system is controllable, we can always find an input transfers the state from any x init to any x final in time n Example: Suppose there is an input that steers the state from an initial state x init at time t = 0 to some state x final at time t = N (where N > n) Is there always an input that steers the state from x init at time t = 0 to x final at time t = n? (We do not assume that (A, B) is controllable) Solution No Let, 0 0 A =, B =, x 0 init =, x 0 final = 4 u t

Notice that this system is not controllable, and R(C N ) = span{(, )} for any N From the previous example, we know that we can drive the system from x init at time t = 0 to x final at time t = N if and only if x final A N x init R(C N ) Now 0 x final A 3 x init = R(C 0 3 ) but, x final A n x init = / R(C n ) The point is that when the system is controllable everything works out a state transfer from any x init to any x final is possible However, when the system is not controllable we must be very careful Many results that seem perfectly plausible are simply not true For instance, the result above is true if the initial state x init = 0, but it not true in general for any x init In the lectures you will see that we always state the controllability and observability assumptions very clearly Observability We ll review observability for the discrete-time linear dynamical system x t+ = Ax t + Bu t, y t = Cx t + Du t, t = 0,,, where x t R n, u t R m, and y R p Let s consider the problem of finding x 0, given,, u t and y 0,,y t We can write where, This implies O t = C CA CA t y 0 y t, T t = O t x 0 = = O t x 0 + T t u t, D 0 0 0 CB D 0 0 CAB CB D 0 CA t 2 B CA t 3 B CA t 4 B D y 0 y t T t and so we see that x 0 can be determined uniquely if and only if N(O t ) = 0 By the Cayley-Hamilton theorem, we know that each A k can be written as a linear combination of 5 u t,

A 0,, A n Thus, for t n, N(O t ) = N(O), where C CA O = O n = CA n is called the observability matrix We say that the system is observable if N(O) = 0 We will also often say that a pair of matrices (C, A) is observable this means that the observability matrix formed from C and A has zero nullspace 6