OPTIMAL CONTROL SYSTEMS

Similar documents
Robust Control 5 Nominal Controller Design Continued

Riccati Equations and Inequalities in Robust Control

6.241 Dynamic Systems and Control

Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

Robust Control 9 Design Summary & Examples

9 Controller Discretization

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

Lecture 5 Linear Quadratic Stochastic Control

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

Robust Control 2 Controllability, Observability & Transfer Functions

Problem 1 Cost of an Infinite Horizon LQR

The norms can also be characterized in terms of Riccati inequalities.

Variable Structure Control ~ Disturbance Rejection. Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Chapter 3. LQ, LQG and Control System Design. Dutch Institute of Systems and Control

Lecture 4 Continuous time linear quadratic regulator

6.241 Dynamic Systems and Control

Topic # Feedback Control Systems

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

ECEEN 5448 Fall 2011 Homework #4 Solutions

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

EEE 184: Introduction to feedback systems

Steady State Kalman Filter

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Semidefinite Programming Duality and Linear Time-invariant Systems

Stability, Pole Placement, Observers and Stabilization

Game Theory Extra Lecture 1 (BoB)

Dynamical systems: basic concepts

Deterministic Dynamic Programming

Controllability, Observability, Full State Feedback, Observer Based Control

Linear Quadratic Optimal Control Topics

Control Systems I. Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback. Readings: Emilio Frazzoli

Extensions and applications of LQ

Risk-Sensitive and Robust Mean Field Games

MEM 355 Performance Enhancement of Dynamical Systems MIMO Introduction

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures

Min-Max Certainty Equivalence Principle and Differential Games

Linear Quadratic Regulator (LQR) Design II

Lecture 2 and 3: Controllability of DT-LTI systems

Linear dynamical systems with inputs & outputs

10 Transfer Matrix Models

Linear System Fundamentals

6. Linear Quadratic Regulator Control

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

Decentralized control with input saturation

Denis ARZELIER arzelier

EE C128 / ME C134 Final Exam Fall 2014

POLE PLACEMENT. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 19

Linear Quadratic Zero-Sum Two-Person Differential Games

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

EE263: Introduction to Linear Dynamical Systems Review Session 6

State Space Design. MEM 355 Performance Enhancement of Dynamical Systems

Homework Solution # 3

Controls Problems for Qualifying Exam - Spring 2014

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

Module 03 Linear Systems Theory: Necessary Background

Stationary trajectories, singular Hamiltonian systems and ill-posed Interconnection

Contents lecture 5. Automatic Control III. Summary of lecture 4 (II/II) Summary of lecture 4 (I/II) u y F r. Lecture 5 H 2 and H loop shaping

Robotics. Control Theory. Marc Toussaint U Stuttgart

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

Robust Multivariable Control

TRACKING AND DISTURBANCE REJECTION

MEM 355 Performance Enhancement of Dynamical Systems

( ) ( = ) = ( ) ( ) ( )

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77

EECS C128/ ME C134 Final Wed. Dec. 14, am. Closed book. One page, 2 sides of formula sheets. No calculators.

Module 05 Introduction to Optimal Control

L2 gains and system approximation quality 1

6 OUTPUT FEEDBACK DESIGN

Résonance et contrôle en cavité ouverte

EE 380. Linear Control Systems. Lecture 10

Stabilization of persistently excited linear systems

Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications

Georgia Institute of Technology Nonlinear Controls Theory Primer ME 6402

Fall 線性系統 Linear Systems. Chapter 08 State Feedback & State Estimators (SISO) Feng-Li Lian. NTU-EE Sep07 Jan08

14 Riccati Equations and their Solution

2 The Linear Quadratic Regulator (LQR)

Linear System Theory

Class 27: Block Diagrams

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

EE Control Systems LECTURE 9

Computational Issues in Nonlinear Dynamics and Control

MEM 355 Performance Enhancement of Dynamical Systems

Low Gain Feedback. Properties, Design Methods and Applications. Zongli Lin. July 28, The 32nd Chinese Control Conference

Dynamical Systems & Lyapunov Stability

EECS C128/ ME C134 Final Thu. May 14, pm. Closed book. One page, 2 sides of formula sheets. No calculators.

Observability. It was the property in Lyapunov stability which allowed us to resolve that

Observability and state estimation

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7)

Suppose that we have a specific single stage dynamic system governed by the following equation:

Math Ordinary Differential Equations

Department of Mathematics IIT Guwahati

w T 1 w T 2. w T n 0 if i j 1 if i = j

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

4F3 - Predictive Control

Professor Fearing EE C128 / ME C134 Problem Set 7 Solution Fall 2010 Jansen Sheng and Wenjie Chen, UC Berkeley

Transcription:

SYSTEMS MIN-MAX Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University

OUTLINE MIN-MAX CONTROL Problem Definition HJB Equation Example GAME THEORY Differential Games Isaacs Equations EXAMPLES Heating Tank

PROBLEM DEFINITION PROBLEM SETUP Consider the control system where x R n is the state u U R m is the control ẋ = f (x, u, w w W R q is an external disturbance We seek a control u that minimizes the cost J for the worst possible disturbance w and T J = l (x (T + L (x (t, u (t, w (t dt 0

HJB EQUATION HJB EQUATION We seek both u (t and w (t such that {u, w } = arg min Define the cost to go ( V (x, t = min max l (x (T + u U w W The principle of optimality leads to V (x, t t + H ( x, max u U w W T t J (u, w L (x (τ, u (τ, w (τ dτ V (x, t = 0 x where H (x, λ = min max u U w W { L (x, u, w + λ T f (x, u, w }

EXAMPLE EXAMPLE Consider the linear dynamics and cost 1 J(u, w = lim T 2T T Suppose (A, B and (A, E are stabilizable Then (A, C is detectable 0 ẋ = Ax + Bu + Ew y = Cx u (t = 1 ρ BT Sx (t, [ ] y T (ty(t + ρu T (tu(t γ 2 w T (tw(t dt w(t = 1 γ 2 ET Sx(t ( 1 A T S + SA S ρ BBT 1 γ 2 EET S = C T C

EXAMPLE EXAMPLE CONT D: SOME CALCULATIONS { 1 H ( = min max x T C T Cx + ρu T u γ 2 w T w } + λ T (Ax + Bu + Ew u w 2 u = 1 ρ BT λ, w = 1 γ 2 ET λ H = 1 2 xt C T Cx + λ T Ax 1 2ρ λt BB T λ + 1 2γ 2 λt EE T λ ẋ = H λ = Ax + 1 γ 2 EET λ 1 ρ BBT λ [ ẋ λ λ = H x = CT Cx + A T λ ] [ 1 A = γ EE T 1 ] [ 2 ρ BBT x C T C A T λ ]

EXAMPLE EXAMPLE CONT D Consider the Hamiltonian matrix [ A H = C T C 1 EE T 1 ] γ 2 ρ BBT A T Now, under the above assumptions: There exists a γ min such that so that there are no eigenvalues of H on the imaginary axis provided γ > γ min. When γ we approach the standard LQR solution. For γ = γ min, the controller is the full state feedback H controller. All other values of γ produce valid min-max controllers. The condition Reλ(H 0 is equivalent to stability of the matrix The closed loop system matrix is A + 1 γ 2 EET 1 ρ BBT A 1 ρ BBT The term EE T /γ 2 is destabilizing, so the system is guaranteed some margin of stability.

DIFFERENTIAL GAMES TWO PERSON DIFFERENTIAL GAME Once again consider the system ẋ = f (x, u, w However, now consider both u and w to be controls associated, respectively, with two players A and B. Player A wants to minimize the cost T J = l (x (T + L (x (t, u (t, w (t dt 0 and player B wants to maximize it. This is a two person, zero-sum, differential game.

DIFFERENTIAL GAMES DEFINITIONS Instead of the cost to go define two value functions: DEFINITION (VALUE FUNCTIONS The lower value function is ( T l (x (T + V A (x, t = min u U max w W and the upper value function is ( V B (x, t = max min l (x (T + w W u U t T t L (x (τ, u (τ, w (τ dτ L (x (τ, u (τ, w (τ dτ if V A = V B, a unique solution exists. This solution is called a pure strategy. If V A V B, a unique strategy does not exist. In general, the player who plays second has an advantage.

ISAACS EQUATIONS ISAACS EQUATIONS Define H A (x, λ = min H B (x, λ = min Then we get Isaacs equations H A ( x, VA (x, t x V A (x, t t V B (x, t t max u U w W max u U w W { L (x, u, w + λ T f (x, u, w } { L (x, u, w + λ T f (x, u, w } + H A ( x, + H B ( x, H B ( x, VA (x, t = 0 x VB (x, t = 0 x V A (x, t V B (x, t VB (x, t x and the game is said to satisfy the Isaacs condition. Notice that if u,w are optimal,the the pair (u, w is a saddle point in the sense that J (u, w J (u, w J (u, w

HEATING TANK WELL-STIRRED HEATING TANK STARTUP ẋ = x + w + u, x (0 = x 0, 0 u 2, w 1 x, dimensionless tank temperature u, dimensionless heat input w, dimensionless input water flow T J = (x x 2 dt, T fixed 0 H = (x x 2 + λ ( x + w + u {u, w } = arg min max H (x, u, w, λ u w u = (1 sgn (λ, w = sgn (λ λ = H x λ = λ 2 (x x, λ (T = 0

HEATING TANK EXAMPLE, CONT D ẋ = x + 1 λ = λ 2 (x x x (T = x, λ (T = 0 0.5 0.0 x,λ 0.5 1.0 with x = 0.8 0.0 0.5 1.0 1.5 2.0