Lecture 4 Continuous time linear quadratic regulator

Size: px
Start display at page:

Download "Lecture 4 Continuous time linear quadratic regulator"

Transcription

1 EE363 Winter Lecture 4 Continuous time linear quadratic regulator continuous-time LQR problem dynamic programming solution Hamiltonian system and two point boundary value problem infinite horizon LQR direct solution of ARE via Hamiltonian 4 1

2 Continuous-time LQR problem continuous-time system ẋ = Ax + Bu, x(0) = x 0 problem: choose u : [0, T R m to minimize J = T 0 ( x(τ) T Qx(τ) + u(τ) T Ru(τ) ) dτ + x(t) T Q f x(t) T is time horizon Q = Q T 0, Q f = Q T f 0, R = RT > 0 are state cost, final state cost, and input cost matrices... an infinite-dimensional problem: (trajectory u : [0, T R m is variable) Continuous time linear quadratic regulator 4 2

3 Dynamic programming solution we ll solve LQR problem using dynamic programming for 0 t T we define the value function V t : R n R by V t (z) = min u T t ( x(τ) T Qx(τ) + u(τ) T Ru(τ) ) dτ + x(t) T Q f x(t) subject to x(t) = z, ẋ = Ax + Bu minimum is taken over all possible signals u : [t, T R m V t (z) gives the minimum LQR cost-to-go, starting from state z at time t V T (z) = z T Q f z Continuous time linear quadratic regulator 4 3

4 fact: V t is quadratic, i.e., V t (z) = z T P t z, where P t = P T t 0 similar to discrete-time case: P t can be found from a differential equation running backward in time from t = T the LQR optimal u is easily expressed in terms of P t Continuous time linear quadratic regulator 4 4

5 we start with x(t) = z let s take u(t) = w R m, a constant, over the time interval [t, t + h, where h > 0 is small cost incurred over [t, t + h is t+h t ( x(τ) T Qx(τ) + w T Rw ) dτ h(z T Qz + w T Rw) and we end up at x(t + h) z + h(az + Bw) Continuous time linear quadratic regulator 4 5

6 min-cost-to-go from where we land is approximately V t+h (z + h(az + Bw)) = (z + h(az + Bw)) T P t+h (z + h(az + Bw)) (z + h(az + Bw)) T (P t + hp t )(z + h(az + Bw)) ( ) z T P t z + h (Az + Bw) T P t z + z T P t (Az + Bw) + z T P t z (dropping h 2 and higher terms) cost incurred plus min-cost-to-go is approximately z T P t z+h ( ) z T Qz + w T Rw + (Az + Bw) T P t z + z T P t (Az + Bw) + z T P t z minimize over w to get (approximately) optimal w: 2hw T R + 2hz T P t B = 0 Continuous time linear quadratic regulator 4 6

7 so w = R 1 B T P t z thus optimal u is time-varying linear state feedback: u lqr (t) = K t x(t), K t = R 1 B T P t Continuous time linear quadratic regulator 4 7

8 HJ equation now let s substitute w into HJ equation: z T P( t z z T P t z+ ) +h z T Qz + w T Rw + (Az + Bw ) T P t z + z T P t (Az + Bw ) + z T P t z yields, after simplification, P t = A T P t + P t A P t BR 1 B T P t + Q which is the Riccati differential equation for the LQR problem we can solve it (numerically) using the final condition P T = Q f Continuous time linear quadratic regulator 4 8

9 Summary of cts-time LQR solution via DP 1. solve Riccati differential equation P t = A T P t + P t A P t BR 1 B T P t + Q, P T = Q f (backward in time) 2. optimal u is u lqr (t) = K t x(t), K t := R 1 B T P t DP method readily extends to time-varying A, B, Q, R, and tracking problem Continuous time linear quadratic regulator 4 9

10 Steady-state regulator usually P t rapidly converges as t decreases below T limit P ss satisfies (cts-time) algebraic Riccati equation (ARE) A T P + PA PBR 1 B T P + Q = 0 a quadratic matrix equation P ss can be found by (numerically) integrating the Riccati differential equation, or by direct methods for t not close to horizon T, LQR optimal input is approximately a linear, constant state feedback u(t) = K ss x(t), K ss = R 1 B T P ss Continuous time linear quadratic regulator 4 10

11 Derivation via discretization let s discretize using small step size h > 0, with Nh = T x((k + 1)h) x(kh) + hẋ(kh) = (I + ha)x(kh) + hbu(kh) J h 2 N 1 k=0 ( x(kh) T Qx(kh) + u(kh) T Ru(kh) ) x(nh)t Q f x(nh) this yields a discrete-time LQR problem, with parameters à = I + ha, B = hb, Q = hq, R = hr, Qf = Q f Continuous time linear quadratic regulator 4 11

12 solution to discrete-time LQR problem is u(kh) = K k x(kh), K k = ( R + B T 1 Pk+1 B) BT Pk+1 à P k 1 = Q + ÃT Pk à ÃT Pk B( R + 1 BT Pk B) BT Pk à substituting and keeping only h 0 and h 1 terms yields P k 1 = hq + P k + ha T Pk + h P k A h P k BR 1 B T Pk which is the same as 1 h ( P k P k 1 ) = Q + A T Pk + P k A P k BR 1 B T Pk letting h 0 we see that P k P kh, where P = Q + A T P + PA PBR 1 B T P Continuous time linear quadratic regulator 4 12

13 similarly, we have K k = ( R + B T 1 Pk+1 B) BT Pk+1 Ã = (hr + h 2 B T Pk+1 B) 1 hb T Pk+1 (I + ha) R 1 B T P kh as h 0 Continuous time linear quadratic regulator 4 13

14 Derivation using Lagrange multipliers pose as constrained problem: minimize J = 1 2 T 0 x(τ)t Qx(τ) + u(τ) T Ru(τ) dτ x(t)t Q f x(t) subject to ẋ(t) = Ax(t) + Bu(t), t [0, T optimization variable is function u : [0, T R m infinite number of equality constraints, one for each t [0, T introduce Lagrange multiplier function λ : [0, T R n and form L = J + T 0 λ(τ) T (Ax(τ) + Bu(τ) ẋ(τ)) dτ Continuous time linear quadratic regulator 4 14

15 Optimality conditions (note: you need distribution theory to really make sense of the derivatives here... ) from u(t) L = Ru(t) + B T λ(t) = 0 we get u(t) = R 1 B T λ(t) to find x(t) L, we use T 0 λ(τ) T ẋ(τ) dτ = λ(t) T x(t) λ(0) T x(0) T 0 λ(τ) T x(τ) dτ from x(t) L = Qx(t) + A T λ(t) + λ(t) = 0 we get λ(t) = A T λ(t) Qx(t) from x(t) L = Q f x(t) λ(t) = 0, we get λ(t) = Q f x(t) Continuous time linear quadratic regulator 4 15

16 Co-state equations optimality conditions are ẋ = Ax + Bu, x(0) = x 0, λ = A T λ Qx, λ(t) = Q f x(t) using u(t) = R 1 B T λ(t), can write as d dt [ x(t) λ(t) = [ A BR 1 B T Q A T [ x(t) λ(t) 2n 2n matrix above is called Hamiltonian for problem with conditions x(0) = x 0, λ(t) = Q f x(t), called two-point boundary value problem Continuous time linear quadratic regulator 4 16

17 as in discrete-time case, we can show that λ(t) = P t x(t), where P t = A T P t + P t A P t BR 1 B T P t + Q, P T = Q f in other words, value function P t gives simple relation between x and λ to show this, we show that λ = Px satisfies co-state equation λ = A T λ Qx λ = d dt (Px) = Px + Pẋ = (Q + A T P + PA PBR 1 B T P)x + P(Ax BR 1 B T λ) = Qx A T Px + PBR 1 B T Px PBR 1 B T Px = Qx A T λ Continuous time linear quadratic regulator 4 17

18 Solving Riccati differential equation via Hamiltonian the (quadratic) Riccati differential equation P = A T P + PA PBR 1 B T P + Q and the (linear) Hamiltonian differential equation d dt [ x(t) λ(t) = [ A BR 1 B T Q A T [ x(t) λ(t) are closely related λ(t) = P t x(t) suggests that P should have the form P t = λ(t)x(t) 1 (but this doesn t make sense unless x and λ are scalars) Continuous time linear quadratic regulator 4 18

19 consider the Hamiltonian matrix (linear) differential equation d dt [ X(t) Y (t) = [ A BR 1 B T Q A T [ X(t) Y (t) where X(t), Y (t) R n n then, Z(t) = Y (t)x(t) 1 satisfies Riccati differential equation Ż = AT Z + ZA ZBR 1 B T Z + Q hence we can solve Riccati DE by solving (linear) matrix Hamiltonian DE, with final conditions X(T) = I, Y (T) = Q f, and forming P(t) = Y (t)x(t) 1 Continuous time linear quadratic regulator 4 19

20 Ż = d dt Y X 1 = Ẏ X 1 Y X 1 ẊX 1 = ( QX A T Y )X 1 Y X 1 ( AX BR 1 B T Y ) X 1 = Q A T Z ZA + ZBR 1 B T Z where we use two identities: d dt (F(t)G(t)) = F(t)G(t) + F(t)Ġ(t) d dt ( F(t) 1 ) = F(t) 1 F(t)F(t) 1 Continuous time linear quadratic regulator 4 20

21 Infinite horizon LQR we now consider the infinite horizon cost function J = 0 x(τ) T Qx(τ) + u(τ) T Ru(τ) dτ we define the value function as V (z) = min u subject to x(0) = z, ẋ = Ax + Bu 0 x(τ) T Qx(τ) + u(τ) T Ru(τ) dτ we assume that (A, B) is controllable, so V is finite for all z can show that V is quadratic: V (z) = z T Pz, where P = P T 0 Continuous time linear quadratic regulator 4 21

22 optimal u is u(t) = Kx(t), where K = R 1 B T P (i.e., a constant linear state feedback) HJ equation is ARE Q + A T P + PA PBR 1 B T P = 0 which together with P 0 characterizes P can solve as limiting value of Riccati DE, or via direct method Continuous time linear quadratic regulator 4 22

23 Closed-loop system with K LQR optimal state feedback gain, closed-loop system is ẋ = Ax + Bu = (A + BK)x fact: closed-loop system is stable when (Q,A) observable and (A, B) controllable we denote eigenvalues of A + BK, called closed-loop eigenvalues, as λ 1,..., λ n with assumptions above, Rλ i < 0 Continuous time linear quadratic regulator 4 23

24 Solving ARE via Hamiltonian [ A BR 1 B T Q A T [ I P = [ A BR 1 B T P Q A T P = [ A + BK Q A T P and so [ I 0 P I [ A BR 1 B T Q A T [ I 0 P I = [ A + BK BR 1 B T 0 (A + BK) T where 0 in lower left corner comes from ARE note that [ I 0 P I 1 = [ I 0 P I Continuous time linear quadratic regulator 4 24

25 we see that: eigenvalues of Hamiltonian H are λ 1,..., λ n and λ 1,..., λ n hence, closed-loop eigenvalues are the eigenvalues of H with negative real part Continuous time linear quadratic regulator 4 25

26 let s assume A + BK is diagonalizable, i.e., T 1 (A + BK)T = Λ = diag(λ 1,..., λ n ) then we have T T ( A BK) T T T = Λ, so [ [ T 1 0 A + BK BR 1 B T 0 T T 0 (A + BK) T [ Λ T = 1 BR 1 B T T T 0 Λ [ T 0 0 T T Continuous time linear quadratic regulator 4 26

27 putting it together we get [ [ [ T 1 0 I 0 I 0 0 T T H P I P I [ [ T 1 0 T 0 = T T P T T H PT [ Λ T = 1 BR 1 B T T T 0 Λ T T [ T 0 0 T T and so thus, the n columns of H [ T PT [ T PT the stable eigenvalues λ 1,..., λ n = [ T PT Λ are the eigenvectors of H associated with Continuous time linear quadratic regulator 4 27

28 Solving ARE via Hamiltonian find eigenvalues of H, and let λ 1,..., λ n denote the n stable ones (there are exactly n stable and n unstable ones) find associated eigenvectors v 1,..., v n, and partition as [ v1 v n = [ X Y R 2n n P = Y X 1 is unique PSD solution of the ARE (this is very close to the method used in practice, which does not require A + BK to be diagonalizable) Continuous time linear quadratic regulator 4 28

EE363 homework 2 solutions

EE363 homework 2 solutions EE363 Prof. S. Boyd EE363 homework 2 solutions. Derivative of matrix inverse. Suppose that X : R R n n, and that X(t is invertible. Show that ( d d dt X(t = X(t dt X(t X(t. Hint: differentiate X(tX(t =

More information

6. Linear Quadratic Regulator Control

6. Linear Quadratic Regulator Control EE635 - Control System Theory 6. Linear Quadratic Regulator Control Jitkomut Songsiri algebraic Riccati Equation (ARE) infinite-time LQR (continuous) Hamiltonian matrix gain margin of LQR 6-1 Algebraic

More information

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear

More information

Lecture 5 Linear Quadratic Stochastic Control

Lecture 5 Linear Quadratic Stochastic Control EE363 Winter 2008-09 Lecture 5 Linear Quadratic Stochastic Control linear-quadratic stochastic control problem solution via dynamic programming 5 1 Linear stochastic system linear dynamical system, over

More information

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28 OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from

More information

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University .. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if

More information

Suppose that we have a specific single stage dynamic system governed by the following equation:

Suppose that we have a specific single stage dynamic system governed by the following equation: Dynamic Optimisation Discrete Dynamic Systems A single stage example Suppose that we have a specific single stage dynamic system governed by the following equation: x 1 = ax 0 + bu 0, x 0 = x i (1) where

More information

9 Controller Discretization

9 Controller Discretization 9 Controller Discretization In most applications, a control system is implemented in a digital fashion on a computer. This implies that the measurements that are supplied to the control system must be

More information

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation EE363 Winter 2008-09 Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation partially observed linear-quadratic stochastic control problem estimation-control separation principle

More information

Problem 1 Cost of an Infinite Horizon LQR

Problem 1 Cost of an Infinite Horizon LQR THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 5 Ahmad F. Taha October 12, 215 Homework Instructions: 1. Type your solutions in the LATEX homework

More information

EE363 Review Session 1: LQR, Controllability and Observability

EE363 Review Session 1: LQR, Controllability and Observability EE363 Review Session : LQR, Controllability and Observability In this review session we ll work through a variation on LQR in which we add an input smoothness cost, in addition to the usual penalties on

More information

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018 EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal

More information

EE363 homework 7 solutions

EE363 homework 7 solutions EE363 Prof. S. Boyd EE363 homework 7 solutions 1. Gain margin for a linear quadratic regulator. Let K be the optimal state feedback gain for the LQR problem with system ẋ = Ax + Bu, state cost matrix Q,

More information

ECEEN 5448 Fall 2011 Homework #4 Solutions

ECEEN 5448 Fall 2011 Homework #4 Solutions ECEEN 5448 Fall 2 Homework #4 Solutions Professor David G. Meyer Novemeber 29, 2. The state-space realization is A = [ [ ; b = ; c = [ which describes, of course, a free mass (in normalized units) with

More information

5. Observer-based Controller Design

5. Observer-based Controller Design EE635 - Control System Theory 5. Observer-based Controller Design Jitkomut Songsiri state feedback pole-placement design regulation and tracking state observer feedback observer design LQR and LQG 5-1

More information

1 Steady State Error (30 pts)

1 Steady State Error (30 pts) Professor Fearing EECS C28/ME C34 Problem Set Fall 2 Steady State Error (3 pts) Given the following continuous time (CT) system ] ẋ = A x + B u = x + 2 7 ] u(t), y = ] x () a) Given error e(t) = r(t) y(t)

More information

Homework Solution # 3

Homework Solution # 3 ECSE 644 Optimal Control Feb, 4 Due: Feb 17, 4 (Tuesday) Homework Solution # 3 1 (5%) Consider the discrete nonlinear control system in Homework # For the optimal control and trajectory that you have found

More information

MODERN CONTROL DESIGN

MODERN CONTROL DESIGN CHAPTER 8 MODERN CONTROL DESIGN The classical design techniques of Chapters 6 and 7 are based on the root-locus and frequency response that utilize only the plant output for feedback with a dynamic controller

More information

Lecture 6. Foundations of LMIs in System and Control Theory

Lecture 6. Foundations of LMIs in System and Control Theory Lecture 6. Foundations of LMIs in System and Control Theory Ivan Papusha CDS270 2: Mathematical Methods in Control and System Engineering May 4, 2015 1 / 22 Logistics hw5 due this Wed, May 6 do an easy

More information

EE221A Linear System Theory Final Exam

EE221A Linear System Theory Final Exam EE221A Linear System Theory Final Exam Professor C. Tomlin Department of Electrical Engineering and Computer Sciences, UC Berkeley Fall 2016 12/16/16, 8-11am Your answers must be supported by analysis,

More information

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Module 07 Controllability and Controller Design of Dynamical LTI Systems Module 07 Controllability and Controller Design of Dynamical LTI Systems Ahmad F. Taha EE 5143: Linear Systems and Control Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ataha October

More information

Topic # Feedback Control Systems

Topic # Feedback Control Systems Topic #17 16.31 Feedback Control Systems Deterministic LQR Optimal control and the Riccati equation Weight Selection Fall 2007 16.31 17 1 Linear Quadratic Regulator (LQR) Have seen the solutions to the

More information

EE C128 / ME C134 Final Exam Fall 2014

EE C128 / ME C134 Final Exam Fall 2014 EE C128 / ME C134 Final Exam Fall 2014 December 19, 2014 Your PRINTED FULL NAME Your STUDENT ID NUMBER Number of additional sheets 1. No computers, no tablets, no connected device (phone etc.) 2. Pocket

More information

Course Outline. Higher Order Poles: Example. Higher Order Poles. Amme 3500 : System Dynamics & Control. State Space Design. 1 G(s) = s(s + 2)(s +10)

Course Outline. Higher Order Poles: Example. Higher Order Poles. Amme 3500 : System Dynamics & Control. State Space Design. 1 G(s) = s(s + 2)(s +10) Amme 35 : System Dynamics Control State Space Design Course Outline Week Date Content Assignment Notes 1 1 Mar Introduction 2 8 Mar Frequency Domain Modelling 3 15 Mar Transient Performance and the s-plane

More information

Linear Quadratic Optimal Control Topics

Linear Quadratic Optimal Control Topics Linear Quadratic Optimal Control Topics Finite time LQR problem for time varying systems Open loop solution via Lagrange multiplier Closed loop solution Dynamic programming (DP) principle Cost-to-go function

More information

ECSE.6440 MIDTERM EXAM Solution Optimal Control. Assigned: February 26, 2004 Due: 12:00 pm, March 4, 2004

ECSE.6440 MIDTERM EXAM Solution Optimal Control. Assigned: February 26, 2004 Due: 12:00 pm, March 4, 2004 ECSE.6440 MIDTERM EXAM Solution Optimal Control Assigned: February 26, 2004 Due: 12:00 pm, March 4, 2004 This is a take home exam. It is essential to SHOW ALL STEPS IN YOUR WORK. In NO circumstance is

More information

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function: Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 24: H2 Synthesis Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology May 4, 2011 E. Frazzoli (MIT) Lecture 24: H 2 Synthesis May

More information

Lecture 2 and 3: Controllability of DT-LTI systems

Lecture 2 and 3: Controllability of DT-LTI systems 1 Lecture 2 and 3: Controllability of DT-LTI systems Spring 2013 - EE 194, Advanced Control (Prof Khan) January 23 (Wed) and 28 (Mon), 2013 I LTI SYSTEMS Recall that continuous-time LTI systems can be

More information

2 The Linear Quadratic Regulator (LQR)

2 The Linear Quadratic Regulator (LQR) 2 The Linear Quadratic Regulator (LQR) Problem: Compute a state feedback controller u(t) = Kx(t) that stabilizes the closed loop system and minimizes J := 0 x(t) T Qx(t)+u(t) T Ru(t)dt where x and u are

More information

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules Advanced Control State Regulator Scope design of controllers using pole placement and LQ design rules Keywords pole placement, optimal control, LQ regulator, weighting matrixes Prerequisites Contact state

More information

CONTROL DESIGN FOR SET POINT TRACKING

CONTROL DESIGN FOR SET POINT TRACKING Chapter 5 CONTROL DESIGN FOR SET POINT TRACKING In this chapter, we extend the pole placement, observer-based output feedback design to solve tracking problems. By tracking we mean that the output is commanded

More information

Linear Quadratic Regulator (LQR) Design I

Linear Quadratic Regulator (LQR) Design I Lecture 7 Linear Quadratic Regulator LQR) Design I Dr. Radhakant Padhi Asst. Proessor Dept. o Aerospace Engineering Indian Institute o Science - Bangalore LQR Design: Problem Objective o drive the state

More information

Dynamic interpretation of eigenvectors

Dynamic interpretation of eigenvectors EE263 Autumn 2015 S. Boyd and S. Lall Dynamic interpretation of eigenvectors invariant sets complex eigenvectors & invariant planes left eigenvectors modal form discrete-time stability 1 Dynamic interpretation

More information

EE363 homework 8 solutions

EE363 homework 8 solutions EE363 Prof. S. Boyd EE363 homework 8 solutions 1. Lyapunov condition for passivity. The system described by ẋ = f(x, u), y = g(x), x() =, with u(t), y(t) R m, is said to be passive if t u(τ) T y(τ) dτ

More information

Linear dynamical systems with inputs & outputs

Linear dynamical systems with inputs & outputs EE263 Autumn 215 S. Boyd and S. Lall Linear dynamical systems with inputs & outputs inputs & outputs: interpretations transfer function impulse and step responses examples 1 Inputs & outputs recall continuous-time

More information

Robotics. Control Theory. Marc Toussaint U Stuttgart

Robotics. Control Theory. Marc Toussaint U Stuttgart Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,

More information

Australian Journal of Basic and Applied Sciences, 3(4): , 2009 ISSN Modern Control Design of Power System

Australian Journal of Basic and Applied Sciences, 3(4): , 2009 ISSN Modern Control Design of Power System Australian Journal of Basic and Applied Sciences, 3(4): 4267-4273, 29 ISSN 99-878 Modern Control Design of Power System Atef Saleh Othman Al-Mashakbeh Tafila Technical University, Electrical Engineering

More information

EE C128 / ME C134 Feedback Control Systems

EE C128 / ME C134 Feedback Control Systems EE C128 / ME C134 Feedback Control Systems Lecture Additional Material Introduction to Model Predictive Control Maximilian Balandat Department of Electrical Engineering & Computer Science University of

More information

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin LQR, Kalman Filter, and LQG Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin May 2015 Linear Quadratic Regulator (LQR) Consider a linear system

More information

Steady State Kalman Filter

Steady State Kalman Filter Steady State Kalman Filter Infinite Horizon LQ Control: ẋ = Ax + Bu R positive definite, Q = Q T 2Q 1 2. (A, B) stabilizable, (A, Q 1 2) detectable. Solve for the positive (semi-) definite P in the ARE:

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Linear Quadratic Regulator (LQR) I

Linear Quadratic Regulator (LQR) I Optimal Control, Guidance and Estimation Lecture Linear Quadratic Regulator (LQR) I Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore Generic Optimal Control Problem

More information

Module 03 Linear Systems Theory: Necessary Background

Module 03 Linear Systems Theory: Necessary Background Module 03 Linear Systems Theory: Necessary Background Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html September

More information

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt +

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt + ECE 553, Spring 8 Posted: May nd, 8 Problem Set #7 Solution Solutions: 1. The optimal controller is still the one given in the solution to the Problem 6 in Homework #5: u (x, t) = p(t)x k(t), t. The minimum

More information

Extensions and applications of LQ

Extensions and applications of LQ Extensions and applications of LQ 1 Discrete time systems 2 Assigning closed loop pole location 3 Frequency shaping LQ Regulator for Discrete Time Systems Consider the discrete time system: x(k + 1) =

More information

Formula Sheet for Optimal Control

Formula Sheet for Optimal Control Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming

More information

LINEAR QUADRATIC GAUSSIAN

LINEAR QUADRATIC GAUSSIAN ECE553: Multivariable Control Systems II. LINEAR QUADRATIC GAUSSIAN.: Deriving LQG via separation principle We will now start to look at the design of controllers for systems Px.t/ D A.t/x.t/ C B u.t/u.t/

More information

The Inverted Pendulum

The Inverted Pendulum Lab 1 The Inverted Pendulum Lab Objective: We will set up the LQR optimal control problem for the inverted pendulum and compute the solution numerically. Think back to your childhood days when, for entertainment

More information

Linear Quadratic Regulator (LQR) Design II

Linear Quadratic Regulator (LQR) Design II Lecture 8 Linear Quadratic Regulator LQR) Design II Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore Outline Stability and Robustness properties

More information

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear

More information

Lecture 2: Discrete-time Linear Quadratic Optimal Control

Lecture 2: Discrete-time Linear Quadratic Optimal Control ME 33, U Berkeley, Spring 04 Xu hen Lecture : Discrete-time Linear Quadratic Optimal ontrol Big picture Example onvergence of finite-time LQ solutions Big picture previously: dynamic programming and finite-horizon

More information

Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem

Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem Noboru Sakamoto, Branislav Rehak N.S.: Nagoya University, Department of Aerospace

More information

4F3 - Predictive Control

4F3 - Predictive Control 4F3 Predictive Control - Lecture 2 p 1/23 4F3 - Predictive Control Lecture 2 - Unconstrained Predictive Control Jan Maciejowski jmm@engcamacuk 4F3 Predictive Control - Lecture 2 p 2/23 References Predictive

More information

Lecture 19 Observability and state estimation

Lecture 19 Observability and state estimation EE263 Autumn 2007-08 Stephen Boyd Lecture 19 Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time

More information

Homogeneous and particular LTI solutions

Homogeneous and particular LTI solutions Homogeneous and particular LTI solutions Daniele Carnevale Dipartimento di Ing. Civile ed Ing. Informatica (DICII), University of Rome Tor Vergata Fondamenti di Automatica e Controlli Automatici A.A. 2014-2015

More information

1 (30 pts) Dominant Pole

1 (30 pts) Dominant Pole EECS C8/ME C34 Fall Problem Set 9 Solutions (3 pts) Dominant Pole For the following transfer function: Y (s) U(s) = (s + )(s + ) a) Give state space description of the system in parallel form (ẋ = Ax +

More information

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Control Systems I. Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback. Readings: Emilio Frazzoli

Control Systems I. Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback. Readings: Emilio Frazzoli Control Systems I Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback Readings: Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich October 13, 2017 E. Frazzoli (ETH)

More information

Observability and state estimation

Observability and state estimation EE263 Autumn 2015 S Boyd and S Lall Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time observability

More information

Robust Control 5 Nominal Controller Design Continued

Robust Control 5 Nominal Controller Design Continued Robust Control 5 Nominal Controller Design Continued Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University 4/14/2003 Outline he LQR Problem A Generalization to LQR Min-Max

More information

Denis ARZELIER arzelier

Denis ARZELIER   arzelier COURSE ON LMI OPTIMIZATION WITH APPLICATIONS IN CONTROL PART II.2 LMIs IN SYSTEMS CONTROL STATE-SPACE METHODS PERFORMANCE ANALYSIS and SYNTHESIS Denis ARZELIER www.laas.fr/ arzelier arzelier@laas.fr 15

More information

CDS 110b: Lecture 2-1 Linear Quadratic Regulators

CDS 110b: Lecture 2-1 Linear Quadratic Regulators CDS 110b: Lecture 2-1 Linear Quadratic Regulators Richard M. Murray 11 January 2006 Goals: Derive the linear quadratic regulator and demonstrate its use Reading: Friedland, Chapter 9 (different derivation,

More information

EE263: Introduction to Linear Dynamical Systems Review Session 6

EE263: Introduction to Linear Dynamical Systems Review Session 6 EE263: Introduction to Linear Dynamical Systems Review Session 6 Outline diagonalizability eigen decomposition theorem applications (modal forms, asymptotic growth rate) EE263 RS6 1 Diagonalizability consider

More information

A Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley

A Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley A Tour of Reinforcement Learning The View from Continuous Control Benjamin Recht University of California, Berkeley trustable, scalable, predictable Control Theory! Reinforcement Learning is the study

More information

6 OUTPUT FEEDBACK DESIGN

6 OUTPUT FEEDBACK DESIGN 6 OUTPUT FEEDBACK DESIGN When the whole sate vector is not available for feedback, i.e, we can measure only y = Cx. 6.1 Review of observer design Recall from the first class in linear systems that a simple

More information

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10) Subject: Optimal Control Assignment- (Related to Lecture notes -). Design a oil mug, shown in fig., to hold as much oil possible. The height and radius of the mug should not be more than 6cm. The mug must

More information

Riccati Equations in Optimal Control Theory

Riccati Equations in Optimal Control Theory Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics 4-21-2008 Riccati Equations in Optimal Control Theory James Bellon Follow this

More information

Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 2015

Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 2015 Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 15 Asymptotic approach from time-varying to constant gains Elimination of cross weighting

More information

OPTIMAL CONTROL SYSTEMS

OPTIMAL CONTROL SYSTEMS SYSTEMS MIN-MAX Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University OUTLINE MIN-MAX CONTROL Problem Definition HJB Equation Example GAME THEORY Differential Games Isaacs

More information

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004 MER42 Advanced Control Lecture 9 Introduction to Kalman Filtering Linear Quadratic Gaussian Control (LQG) G. Hovland 24 Announcement No tutorials on hursday mornings 8-9am I will be present in all practical

More information

Chapter Introduction. A. Bensoussan

Chapter Introduction. A. Bensoussan Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688

More information

Pontryagin s maximum principle

Pontryagin s maximum principle Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin

More information

HW3 - Due 02/06. Each answer must be mathematically justified. Don t forget your name. 1 2, A = 2 2

HW3 - Due 02/06. Each answer must be mathematically justified. Don t forget your name. 1 2, A = 2 2 HW3 - Due 02/06 Each answer must be mathematically justified Don t forget your name Problem 1 Find a 2 2 matrix B such that B 3 = A, where A = 2 2 If A was diagonal, it would be easy: we would just take

More information

Diagonalization. P. Danziger. u B = A 1. B u S.

Diagonalization. P. Danziger. u B = A 1. B u S. 7., 8., 8.2 Diagonalization P. Danziger Change of Basis Given a basis of R n, B {v,..., v n }, we have seen that the matrix whose columns consist of these vectors can be thought of as a change of basis

More information

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL 1. Optimal regulator with noisy measurement Consider the following system: ẋ = Ax + Bu + w, x(0) = x 0 where w(t) is white noise with Ew(t) = 0, and x 0 is a stochastic

More information

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory

More information

SUCCESSIVE POLE SHIFTING USING SAMPLED-DATA LQ REGULATORS. Sigeru Omatu

SUCCESSIVE POLE SHIFTING USING SAMPLED-DATA LQ REGULATORS. Sigeru Omatu SUCCESSIVE POLE SHIFING USING SAMPLED-DAA LQ REGULAORS oru Fujinaka Sigeru Omatu Graduate School of Engineering, Osaka Prefecture University, 1-1 Gakuen-cho, Sakai, 599-8531 Japan Abstract: Design of sampled-data

More information

ECE 680 Fall Test #2 Solutions. 1. Use Dynamic Programming to find u(0) and u(1) that minimize. J = (x(2) 1) u 2 (k) x(k + 1) = bu(k),

ECE 680 Fall Test #2 Solutions. 1. Use Dynamic Programming to find u(0) and u(1) that minimize. J = (x(2) 1) u 2 (k) x(k + 1) = bu(k), ECE 68 Fall 211 Test #2 Solutions 1. Use Dynamic Programming to find u() and u(1) that minimize subject to 1 J (x(2) 1) 2 + 2 u 2 (k) k x(k + 1) bu(k), where b. Let J (x(k)) be the minimum cost of transfer

More information

Linear Quadratic Regulator (LQR) II

Linear Quadratic Regulator (LQR) II Optimal Control, Guidance and Estimation Lecture 11 Linear Quadratic Regulator (LQR) II Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore Outline Summary o LQR design

More information

Optimal Linear Feedback Control for Incompressible Fluid Flow

Optimal Linear Feedback Control for Incompressible Fluid Flow Optimal Linear Feedback Control for Incompressible Fluid Flow Miroslav K. Stoyanov Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the

More information

Advanced Mechatronics Engineering

Advanced Mechatronics Engineering Advanced Mechatronics Engineering German University in Cairo 21 December, 2013 Outline Necessary conditions for optimal input Example Linear regulator problem Example Necessary conditions for optimal input

More information

Here represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.

Here represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities. 19 KALMAN FILTER 19.1 Introduction In the previous section, we derived the linear quadratic regulator as an optimal solution for the fullstate feedback control problem. The inherent assumption was that

More information

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T 1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.

More information

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis Topic # 16.30/31 Feedback Control Systems Analysis of Nonlinear Systems Lyapunov Stability Analysis Fall 010 16.30/31 Lyapunov Stability Analysis Very general method to prove (or disprove) stability of

More information

Lecture 20: Linear Dynamics and LQG

Lecture 20: Linear Dynamics and LQG CSE599i: Online and Adaptive Machine Learning Winter 2018 Lecturer: Kevin Jamieson Lecture 20: Linear Dynamics and LQG Scribes: Atinuke Ademola-Idowu, Yuanyuan Shi Disclaimer: These notes have not been

More information

Consensus Control of Multi-agent Systems with Optimal Performance

Consensus Control of Multi-agent Systems with Optimal Performance 1 Consensus Control of Multi-agent Systems with Optimal Performance Juanjuan Xu, Huanshui Zhang arxiv:183.941v1 [math.oc 6 Mar 18 Abstract The consensus control with optimal cost remains major challenging

More information

Automatic Control 2. Nonlinear systems. Prof. Alberto Bemporad. University of Trento. Academic year

Automatic Control 2. Nonlinear systems. Prof. Alberto Bemporad. University of Trento. Academic year Automatic Control 2 Nonlinear systems Prof. Alberto Bemporad University of Trento Academic year 2010-2011 Prof. Alberto Bemporad (University of Trento) Automatic Control 2 Academic year 2010-2011 1 / 18

More information

Deterministic Kalman Filtering on Semi-infinite Interval

Deterministic Kalman Filtering on Semi-infinite Interval Deterministic Kalman Filtering on Semi-infinite Interval L. Faybusovich and T. Mouktonglang Abstract We relate a deterministic Kalman filter on semi-infinite interval to linear-quadratic tracking control

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 8: Solutions of State-space Models Readings: DDV, Chapters 10, 11, 12 (skip the parts on transform methods) Emilio Frazzoli Aeronautics and Astronautics Massachusetts

More information

ACM/CMS 107 Linear Analysis & Applications Fall 2016 Assignment 4: Linear ODEs and Control Theory Due: 5th December 2016

ACM/CMS 107 Linear Analysis & Applications Fall 2016 Assignment 4: Linear ODEs and Control Theory Due: 5th December 2016 ACM/CMS 17 Linear Analysis & Applications Fall 216 Assignment 4: Linear ODEs and Control Theory Due: 5th December 216 Introduction Systems of ordinary differential equations (ODEs) can be used to describe

More information

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

Hamilton-Jacobi-Bellman Equation Feb 25, 2008 Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS

More information

Linear System Theory

Linear System Theory Linear System Theory Wonhee Kim Chapter 6: Controllability & Observability Chapter 7: Minimal Realizations May 2, 217 1 / 31 Recap State space equation Linear Algebra Solutions of LTI and LTV system Stability

More information

Appendix A Solving Linear Matrix Inequality (LMI) Problems

Appendix A Solving Linear Matrix Inequality (LMI) Problems Appendix A Solving Linear Matrix Inequality (LMI) Problems In this section, we present a brief introduction about linear matrix inequalities which have been used extensively to solve the FDI problems described

More information

Module 05 Introduction to Optimal Control

Module 05 Introduction to Optimal Control Module 05 Introduction to Optimal Control Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html October 8, 2015

More information

ME8281-Advanced Control Systems Design

ME8281-Advanced Control Systems Design ME8281 - Advanced Control Systems Design Spring 2016 Perry Y. Li Department of Mechanical Engineering University of Minnesota Spring 2016 Lecture 4 - Outline 1 Homework 1 to be posted by tonight 2 Transition

More information

Stochastic and Adaptive Optimal Control

Stochastic and Adaptive Optimal Control Stochastic and Adaptive Optimal Control Robert Stengel Optimal Control and Estimation, MAE 546 Princeton University, 2018! Nonlinear systems with random inputs and perfect measurements! Stochastic neighboring-optimal

More information

Principles of Optimal Control Spring 2008

Principles of Optimal Control Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 16.323 Principles of Optimal Control Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.323 Lecture

More information

Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE)

Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE) Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE) AHMED KHAMIS Idaho State University Department of Electrical Engineering Pocatello, Idaho USA khamahme@isu.edu D.

More information