Game Theory Extra Lecture 1 (BoB)

Size: px
Start display at page:

Download "Game Theory Extra Lecture 1 (BoB)"

Transcription

1 Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder, pp , , ,

2 To understand Definition of difference/differential games Optimal Control Dynamic programming and HJBI equation Open-loop and feedback Nash Equilibria Copies from Baser/Olsder Material 2

3 Difference Games State equation u i determined by player i x k+1 = f(x k,u 1,u 2,...,u N ), x 1 given Observations y i k = hi k (x k), i = 1,...,N Information structure, for each player a subset η i k of {y 1 1,...,y 1 k;...;y N 1,...,y N k ;u 1 1,...,u 1 k 1;...;u N 1,...,u N k 1} Goal: Player i wants to minimize some functional, L i, of x and u given his available information. Solution concepts: As before, Nash, Stackelberg, etc 3

4 Differential Games d dt x(t) = f(x(t),u 1(t),u 2 (t),...,u N (t)), x(t 0 ) given Goal: to minimize L i = T 0 g i (x,u 1,...,u N ) dt T either fixed, or given by an implicit equation T = min t {t;l(x(t)) 0} 4

5 Example, Robust Control H control Assume u 1,u 2 and y are related via ẋ = Ax+B 1 u 1 +B 2 u 2 z = C 1 x y = C 2 x+du 2 Typical H question: Is z 2 (t)+u 2 1(t)dt γ 2 u 2 2(t)dt? Differential game between controller, u 1 (t) function of y([0,t]) disturbance, u 2 (t) function of x([0,t]) and u 1 ([0,t]) Performance criterium min u 1 max u 2 0 (z 2 +u 2 1 γ 2 u 2 2)dt 5

6 Robust/H Control The interpretation is that u 1 is the control signal and u 2 is a worstcase disturbance signal. Introduce the L 2 (or energy) norm of the signal w w 2 = ( 0 ) 1/2 w(t) 2 dt The H norm of a linear system G(s) is defined by G = sup w 0 Gw w 6

7 Robustness The H norm measures the largest amplification of energy by the system. Minimization of the norm is clearly interesting if w is a disturbance signal. Another motivation is given by the so called small-gain theorem Theorem A closed loop system is stable for all perturbations with norm < γ if and only if G K 1/γ G Κ 7

8 H control G K < γ K : z 2 γ 2 w 2 < 0 min u max w ( z 2 γ 2 w 2 ) < 0 This is exactly (the upper value of) an affine quadratic game The relation between H control and game theory was noted rather late (end of 80s). For details see Section 6.6 8

9 Pursuit Evasion Games 9

10 Tools from One-Person Optimization Dynamic programming The (maximum) minimum principle Baser/Olsder Ch

11 Dynamic Programming, discrete time x k+1 = f k (x k,u k ), u k U k K L(u) = g k (x k+1,u k,x k ) k=1 f k,g k,u k,k,x 1 given. Want u k = γ k (x k ) that minimizes L Idea: Generalize the problem; Calculate the value function V(k,x) = min γ k,...,γ K for all initial conditions x k = x. K i=k g i (x i+1,u i,x k ) 11

12 Principle of Optimality An optimal control sequence u 1,...,u K should be sequentially optimal. The only coupling between the optimization problems on time horizons [1,k 1] and [k,k] is via the state x k. This leads to V(k,x) = min u k U k [g k (f k (x,u k ),u k,x)+v(k +1,f k (x,u k ))] with the final condition V(K,x) = min u K g k (x K+1,u K,x K ). 12

13 Example Affine quadratic problems has the solution x k+1 = A k x k +B k u k +c k L(u) = 1 K (x 2 k+1qx k+1 +u kru k ) k=1 V(k,x) = 1 2 x S k x+x s k +q k u k = P k S k+1 A k x k P k (s k+1 +S k+1 c k ) Formulas for P k,s k,s k are given on p

14 Dynamic Programming, continuous-time The same reasoning leads to a PDE equation called the Hamilton- Jacobi-Bellman (HJB) equation ẋ(t) = f(x(t),u(t)) u(t) = γ(x(t)) L(u) = T 0 g(x(t),u(t)dt+q(t,x(t)) The final time T can be either fixed known, or given implicitly by T = min{t : l(x(t)) = 0} t 0 14

15 HJB equation T V(t,x) = min g(x(s),u(s))ds+q(t,x(t)) {u(s),s [t,t]} t V(T,x) = q(t,x) along l(x) = 0 Principle of optimality shows that if V is C 1 then V(t,x) [ ] V(t,x) = min f(x,u)+g(x,u) t u U x (+ final condition when t = T) 15

16 Theorem of Sufficiency Theorem 5.3 p 237 If a C 1 function V can be found that satisfies the HJB equation and boundary conditions above then it generates the optimal strategy u through the pointwise optimization problem defined by the right hand side. Proof see p. 237 Example: Affine-quadratic problems has a quadratic function V(t,x) = 1 2 x S(t)x+k (t)x+m(t), see p

17 The Minimum Principle Introduce the costate vector p (t) = V(t,x (t))/ x where x denotes the optimal trajectory corresponding to u, i.e. ẋ (t) = f(x (t),u (t)). Also define (the Hamiltonian) Theorem 5.4 H(t,p,x,u) = g(x,u)+p (t)f(x,u) p (T) = q(t(x ),x ) along l(t,x) = 0 x ṗ (t) = H(t,p,x,u ) x u = arg min u U H(t,p,x,u) See also the discrete-time counterpart Theorem

18 N person difference games How does this translate to difference games with N > 1 players? x k+1 = f k (x k,u 1 k,...,u N k ) K J i (u 1,...,u N ) = gk(x i k+1,u 1 k,...,u N k,x k ) k=1 The control laws u i = γ i constitute a Nash equilibrium if Open-loop information: u i k J i (γ ) J i ({γ i,γ i }) i is an open-loop function of k Closed-loop information: u i k (x k) is allowed to be a function of x k. Hence changing u i will result in changes in u j 18

19 See Section Open Loop Dynamic Games Idea: Player i solves a one-player problem. His choice u i will not influence the other u j, so these can be treated as given functions of time. Then the results from the N = 1 apply directly 19

20 Open Loop Theorems Theorem 6.1 If u = γ provides an open-loop Nash equilibrium, then there exists a sequence of costate vectors p i such that x k+1 = f k (x k,u k) γ i k = arg min u i k p i k = x k f k H i k(p i k+1,{γ i k,u i k},x k) [ p i k+1 + ( ) ] gk i x k+1 where H i = g i +p i f. For details see p ( ) gk i x k The affine case results in coupled Riccati equations. The special zero-sum case for N = 2 is given in Theorem 6.3, and 6.4. Only one Riccati equation must be solved. The condition for existence of Nash equilibrium has the form I B 2 k S k+1b 2 k > 0. 20

21 Feedback Solutions Theorem 6.6 The set of strategies γ provides a feedback Nash equilibrium if and only if there exists functions V i (k,x) such that V i (k,x) = arg min[gk( i,u i i k)+v i (k +1, u i k = g i k( i,γ i k )+V i (k +1, f i k (x,γ i k )) f i k (x,u i k))] where f i k (x,ui k ) = f k(x,{γ i (x),u i k }) 21

22 Feedback Solutions, Zero-sum case N = 2 See Corollary 6.2 p. 282 The set of strategies γ 1,γ 2 provides a feedback saddle-point solution if and only if there exists functions V(k,x) such that for all k V(k,x) = min u 1 k = max u 2 k max u 2 k min u 1 k [ gk (f k (x,u 1 k,u 2 k),u 1 k,u 2 k,x)+v(k +1,f k (x,u 1 k,u 2 k) ] [ gk (f k (x,u 1 k,u 2 k),u 1 k,u 2 k,x)+v(k +1,f k (x,u 1 k,u 2 k) ] = g k (f k (x,γ 1 k,γ 2 k ),γ 1 k (x),γ 2 k (x),x)+v(k +1,f k (x,γ 1 k (x),γ 2 k (x))) 22

23 Differential Games Open-loop Nash Equilibria Want to present counterparts for continuous time. Existence of smooth cost function V is more problematic. L i (u 1,...,u N ) = = ẋ(t) = f(x(t),u 1 (t),...,u N (t)) T 0 g i (x(t),u 1 (t),...,u N (t))dt+q i (x(t)) 23

24 Open-loop Nash Equilibria Theorem 6.11 (Some smoothness assumptions) If u i = γ i (t,x 0 ) provides an open-loop Nash equilibrium (with corresponding x ) then there exist N costate functions p i (t) such that ẋ (t) = f(x (t),u 1 (t),...,u N (t)) γ i = arg min u i U i H i (p i (t),x (t),{u i,u i }) ṗ i (t) = x Hi (p i (t),x,u (t)) ṗ i (T) = x qi (x (T)) where H i (p i,x,u) = g i (x,u)+p i f(x,u). 24

25 The Affine-Quadratic Case f(x,u) = A(t)x+ B i (t)u i +c(t) g i (x,u) = 1 2 x Q i (t)x+ u j R ij u j q i (x(t)) = 1 2 x (T)Q i fx(t) with Q i 0,Q i f 0,Rii > 0. 25

26 The Affine-Quadratic Case Theorem 6.12+p.316 There is an open-loop Nash equilibrium solution to the affine-quadratic game for any T [0,T f ] if and only if the following coupled matrix Riccati differential equations have solutions for any T [0,T f ] M i +M i A+A M i +Q i M i j B j R jj (t) 1 B j M j = 0 M i (T) = Q i f The NE is given by u i = R ii (t) 1 B i [M i x (t) + m i ] where the feedforward signal m i is given by (6.51) 26

27 P1 minimizes, P2 maximizes. L(u 1,u 2 ) = 1 2 N = 2 Zero-Sum Case T with Q 0, Q f 0. 0 x Qx+u 1 u 1 u 2 u 2 dt+ 1 2 x (T)Q f x(t) Assume there exists a unique bounded symmetric solution S( ) to the matrix equation Ṡ +A S +SA+Q+SB 2 B 2 S = 0, S(T) = Q f on the interval [0,T]. 27

28 Then there exists a solution to Ṁ +A M +MA+Q M(B 1 N 1 N 2 B 2 )M = 0;M(T) = Q f and the game admits a unique open-loop saddle-point given by γ 1 (t,x 0 ) = B 1 (t) [M(t)x (t)+m(t)] γ 2 (t,x 0 ) = B 2 (t) [M(t)x (t)+m(t)] 28

29 Closed-loop Feedback Nash Equilibria Will only give the result for N = 2 and zero-sum situation Corollary 6.6, p326 A pair of strategies {γ i } provides a feedback saddle-point solution if there exists a function V satisfying the PDE V(t,x) [ ] V(t,x) = minmax f(x,u 1,u 2 )+g(x,u 1,u 2 ) t u 1 u 2 x [ ] V(t,x) = maxmin f(x,u 1,u 2 )+g(x,u 1,u 2 ) u 2 u 1 x [ ] V(t,x) = f(x,γ 1,γ 2 )+g(x,γ 1,γ 2 ) x The value of the game is V(0,x 0 ) This is the famous Hamilton-Jacobi-Bellman-Isaacs equation obtained by Isaacs in 1950s. 29

30 N = 2, Zero-sum Affine Case ẋ = Ax+B 1 u 1 +B 2 u 2 +c L = 1 2 T 0 x Qx+u 1 u 1 u 2 u 2 dt Theorem 6.17 If there exists a unique symmetric bounded solution to the Riccati equation Ż +A Z +ZA+Q Z(B 1 B 1 B 2 B 2 )Z = 0, Z(T) = Q f then the two-player zero-sum game admits a unique feedback saddle-point given by (for details see p. 327) γ i (t,x) = ( 1) i B i (t) [Z(t)x(t)+ζ(t)] 30

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28 OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from

More information

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

Hamilton-Jacobi-Bellman Equation Feb 25, 2008 Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS

More information

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018 EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal

More information

Formula Sheet for Optimal Control

Formula Sheet for Optimal Control Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games

EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

Risk-Sensitive and Robust Mean Field Games

Risk-Sensitive and Robust Mean Field Games Risk-Sensitive and Robust Mean Field Games Tamer Başar Coordinated Science Laboratory Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL - 6181 IPAM

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Conference on the occasion of Giovanni Colombo and Franco Ra... https://events.math.unipd.it/oscgc18/ Optimization,

More information

Homework Solution # 3

Homework Solution # 3 ECSE 644 Optimal Control Feb, 4 Due: Feb 17, 4 (Tuesday) Homework Solution # 3 1 (5%) Consider the discrete nonlinear control system in Homework # For the optimal control and trajectory that you have found

More information

Nonlinear Control. Nonlinear Control Lecture # 25 State Feedback Stabilization

Nonlinear Control. Nonlinear Control Lecture # 25 State Feedback Stabilization Nonlinear Control Lecture # 25 State Feedback Stabilization Backstepping η = f a (η)+g a (η)ξ ξ = f b (η,ξ)+g b (η,ξ)u, g b 0, η R n, ξ, u R Stabilize the origin using state feedback View ξ as virtual

More information

OPTIMAL CONTROL SYSTEMS

OPTIMAL CONTROL SYSTEMS SYSTEMS MIN-MAX Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University OUTLINE MIN-MAX CONTROL Problem Definition HJB Equation Example GAME THEORY Differential Games Isaacs

More information

Chap. 3. Controlled Systems, Controllability

Chap. 3. Controlled Systems, Controllability Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :

More information

An Introduction to Noncooperative Games

An Introduction to Noncooperative Games An Introduction to Noncooperative Games Alberto Bressan Department of Mathematics, Penn State University Alberto Bressan (Penn State) Noncooperative Games 1 / 29 introduction to non-cooperative games:

More information

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory Constrained Innite-Time Nonlinear Quadratic Optimal Control V. Manousiouthakis D. Chmielewski Chemical Engineering Department UCLA 1998 AIChE Annual Meeting Outline Unconstrained Innite-Time Nonlinear

More information

Linear Quadratic Zero-Sum Two-Person Differential Games

Linear Quadratic Zero-Sum Two-Person Differential Games Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,

More information

Robust control and applications in economic theory

Robust control and applications in economic theory Robust control and applications in economic theory In honour of Professor Emeritus Grigoris Kalogeropoulos on the occasion of his retirement A. N. Yannacopoulos Department of Statistics AUEB 24 May 2013

More information

Lecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster.

Lecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster. Lecture 8 Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture

More information

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4. Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation

More information

An Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game

An Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game An Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game Marzieh Khakestari (Corresponding author) Institute For Mathematical Research, Universiti Putra Malaysia, 43400

More information

Multi-Pursuer Single-Evader Differential Games with Limited Observations*

Multi-Pursuer Single-Evader Differential Games with Limited Observations* American Control Conference (ACC) Washington DC USA June 7-9 Multi-Pursuer Single- Differential Games with Limited Observations* Wei Lin Student Member IEEE Zhihua Qu Fellow IEEE and Marwan A. Simaan Life

More information

Computational Issues in Nonlinear Dynamics and Control

Computational Issues in Nonlinear Dynamics and Control Computational Issues in Nonlinear Dynamics and Control Arthur J. Krener ajkrener@ucdavis.edu Supported by AFOSR and NSF Typical Problems Numerical Computation of Invariant Manifolds Typical Problems Numerical

More information

Static and Dynamic Optimization (42111)

Static and Dynamic Optimization (42111) Static and Dynamic Optimization (421) Niels Kjølstad Poulsen Build. 0b, room 01 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark Email:

More information

The Linear Quadratic Regulator

The Linear Quadratic Regulator 10 The Linear Qadratic Reglator 10.1 Problem formlation This chapter concerns optimal control of dynamical systems. Most of this development concerns linear models with a particlarly simple notion of optimality.

More information

Min-Max Certainty Equivalence Principle and Differential Games

Min-Max Certainty Equivalence Principle and Differential Games Min-Max Certainty Equivalence Principle and Differential Games Pierre Bernhard and Alain Rapaport INRIA Sophia-Antipolis August 1994 Abstract This paper presents a version of the Certainty Equivalence

More information

Numerical Optimal Control Overview. Moritz Diehl

Numerical Optimal Control Overview. Moritz Diehl Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize

More information

Alberto Bressan. Department of Mathematics, Penn State University

Alberto Bressan. Department of Mathematics, Penn State University Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i

More information

Advanced Mechatronics Engineering

Advanced Mechatronics Engineering Advanced Mechatronics Engineering German University in Cairo 21 December, 2013 Outline Necessary conditions for optimal input Example Linear regulator problem Example Necessary conditions for optimal input

More information

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33

More information

pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre

pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre HJ pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre Control of State Constrained Dynamical Systems 25-29 September 2017, Padova Main

More information

On the Stability of the Best Reply Map for Noncooperative Differential Games

On the Stability of the Best Reply Map for Noncooperative Differential Games On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University

More information

Model Predictive Regulation

Model Predictive Regulation Preprints of the 19th World Congress The International Federation of Automatic Control Model Predictive Regulation Cesar O. Aguilar Arthur J. Krener California State University, Bakersfield, CA, 93311,

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 5. Input-Output Stability DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Input-Output Stability y = Hu H denotes

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f

More information

José C. Geromel. Australian National University Canberra, December 7-8, 2017

José C. Geromel. Australian National University Canberra, December 7-8, 2017 5 1 15 2 25 3 35 4 45 5 1 15 2 25 3 35 4 45 5 55 Differential LMI in Optimal Sampled-data Control José C. Geromel School of Electrical and Computer Engineering University of Campinas - Brazil Australian

More information

ESC794: Special Topics: Model Predictive Control

ESC794: Special Topics: Model Predictive Control ESC794: Special Topics: Model Predictive Control Nonlinear MPC Analysis : Part 1 Reference: Nonlinear Model Predictive Control (Ch.3), Grüne and Pannek Hanz Richter, Professor Mechanical Engineering Department

More information

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University. Lecture 4 Chapter 4: Lyapunov Stability Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 4 p. 1/86 Autonomous Systems Consider the autonomous system ẋ

More information

Optimal Control Theory

Optimal Control Theory Optimal Control Theory The theory Optimal control theory is a mature mathematical discipline which provides algorithms to solve various control problems The elaborate mathematical machinery behind optimal

More information

Game Theoretic Continuous Time Differential Dynamic Programming

Game Theoretic Continuous Time Differential Dynamic Programming Game heoretic Continuous ime Differential Dynamic Programming Wei Sun, Evangelos A. heodorou and Panagiotis siotras 3 Abstract In this work, we derive a Game heoretic Differential Dynamic Programming G-DDP

More information

Control of Dynamical System

Control of Dynamical System Control of Dynamical System Yuan Gao Applied Mathematics University of Washington yuangao@uw.edu Spring 2015 1 / 21 Simplified Model for a Car To start, let s consider the simple example of controlling

More information

Lecture Note 7: Switching Stabilization via Control-Lyapunov Function

Lecture Note 7: Switching Stabilization via Control-Lyapunov Function ECE7850: Hybrid Systems:Theory and Applications Lecture Note 7: Switching Stabilization via Control-Lyapunov Function Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio

More information

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information

Input to state Stability

Input to state Stability Input to state Stability Mini course, Universität Stuttgart, November 2004 Lars Grüne, Mathematisches Institut, Universität Bayreuth Part IV: Applications ISS Consider with solutions ϕ(t, x, w) ẋ(t) =

More information

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function: Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic

More information

Output Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x)

Output Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x) Output Feedback and State Feedback EL2620 Nonlinear Control Lecture 10 Exact feedback linearization Input-output linearization Lyapunov-based control design methods ẋ = f(x,u) y = h(x) Output feedback:

More information

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear

More information

SUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS

SUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS ICIC Express Letters ICIC International c 2008 ISSN 1881-803X Volume 2, Number 4, December 2008 pp. 371-377 SUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS Ma. Aracelia

More information

Riccati Equations and Inequalities in Robust Control

Riccati Equations and Inequalities in Robust Control Riccati Equations and Inequalities in Robust Control Lianhao Yin Gabriel Ingesson Martin Karlsson Optimal Control LP4 2014 June 10, 2014 Lianhao Yin Gabriel Ingesson Martin Karlsson (LTH) H control problem

More information

Converse Lyapunov theorem and Input-to-State Stability

Converse Lyapunov theorem and Input-to-State Stability Converse Lyapunov theorem and Input-to-State Stability April 6, 2014 1 Converse Lyapunov theorem In the previous lecture, we have discussed few examples of nonlinear control systems and stability concepts

More information

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems State-Feedback Optimal Controllers for Deterministic Nonlinear Systems Chang-Hee Won*, Abstract A full-state feedback optimal control problem is solved for a general deterministic nonlinear system. The

More information

Suboptimality of minmax MPC. Seungho Lee. ẋ(t) = f(x(t), u(t)), x(0) = x 0, t 0 (1)

Suboptimality of minmax MPC. Seungho Lee. ẋ(t) = f(x(t), u(t)), x(0) = x 0, t 0 (1) Suboptimality of minmax MPC Seungho Lee In this paper, we consider particular case of Model Predictive Control (MPC) when the problem that needs to be solved in each sample time is the form of min max

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Z i Q ij Z j. J(x, φ; U) = X T φ(t ) 2 h + where h k k, H(t) k k and R(t) r r are nonnegative definite matrices (R(t) is uniformly in t nonsingular).

Z i Q ij Z j. J(x, φ; U) = X T φ(t ) 2 h + where h k k, H(t) k k and R(t) r r are nonnegative definite matrices (R(t) is uniformly in t nonsingular). 2. LINEAR QUADRATIC DETERMINISTIC PROBLEM Notations: For a vector Z, Z = Z, Z is the Euclidean norm here Z, Z = i Z2 i is the inner product; For a vector Z and nonnegative definite matrix Q, Z Q = Z, QZ

More information

Lecture 15: H Control Synthesis

Lecture 15: H Control Synthesis c A. Shiriaev/L. Freidovich. March 12, 2010. Optimal Control for Linear Systems: Lecture 15 p. 1/14 Lecture 15: H Control Synthesis Example c A. Shiriaev/L. Freidovich. March 12, 2010. Optimal Control

More information

Prof. Krstic Nonlinear Systems MAE281A Homework set 1 Linearization & phase portrait

Prof. Krstic Nonlinear Systems MAE281A Homework set 1 Linearization & phase portrait Prof. Krstic Nonlinear Systems MAE28A Homework set Linearization & phase portrait. For each of the following systems, find all equilibrium points and determine the type of each isolated equilibrium. Use

More information

Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley

Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability Goal of This

More information

Mean-Field Games with non-convex Hamiltonian

Mean-Field Games with non-convex Hamiltonian Mean-Field Games with non-convex Hamiltonian Martino Bardi Dipartimento di Matematica "Tullio Levi-Civita" Università di Padova Workshop "Optimal Control and MFGs" Pavia, September 19 21, 2018 Martino

More information

Chain differentials with an application to the mathematical fear operator

Chain differentials with an application to the mathematical fear operator Chain differentials with an application to the mathematical fear operator Pierre Bernhard I3S, University of Nice Sophia Antipolis and CNRS, ESSI, B.P. 145, 06903 Sophia Antipolis cedex, France January

More information

Pontryagin s maximum principle

Pontryagin s maximum principle Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin

More information

Chapter Introduction. A. Bensoussan

Chapter Introduction. A. Bensoussan Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688

More information

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations 2005 American Control Conference June 8-10, 2005. Portland, OR, USA WeB10.3 A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations Arik Melikyan, Andrei Akhmetzhanov and Naira

More information

Computing Minmax; Dominance

Computing Minmax; Dominance Computing Minmax; Dominance CPSC 532A Lecture 5 Computing Minmax; Dominance CPSC 532A Lecture 5, Slide 1 Lecture Overview 1 Recap 2 Linear Programming 3 Computational Problems Involving Maxmin 4 Domination

More information

ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko

ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko Continuous-time optimization involves maximization wrt to an innite dimensional object, an entire

More information

Locally optimal controllers and application to orbital transfer (long version)

Locally optimal controllers and application to orbital transfer (long version) 9th IFAC Symposium on Nonlinear Control Systems Toulouse, France, September 4-6, 13 FrA1.4 Locally optimal controllers and application to orbital transfer (long version) S. Benachour V. Andrieu Université

More information

Optimal Control. Lecture 20. Control Lyapunov Function, Optimal Estimation. John T. Wen. April 5. Ref: Papers by R. Freeman (on-line), B&H Ch.

Optimal Control. Lecture 20. Control Lyapunov Function, Optimal Estimation. John T. Wen. April 5. Ref: Papers by R. Freeman (on-line), B&H Ch. Optimal Control Lecture 20 Control Lyapunov Function, Optimal Estimation John T. Wen April 5 Ref: Papers by R. Freeman (on-line), B&H Ch. 12 Outline Summary of Control Lyapunov Function and example Introduction

More information

Modeling and Analysis of Dynamic Systems

Modeling and Analysis of Dynamic Systems Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability

More information

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University .. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if

More information

Max Plus Eigenvector Methods for Nonlinear H Problems: Error Analysis

Max Plus Eigenvector Methods for Nonlinear H Problems: Error Analysis Max Plus Eigenvector Methods for Nonlinear H Problems: Error Analysis William M. McEneaney August 28, 2003 Abstract The H problem for a nonlinear system is considered. The corresponding dynamic programming

More information

Dynamic Consistency for Stochastic Optimal Control Problems

Dynamic Consistency for Stochastic Optimal Control Problems Dynamic Consistency for Stochastic Optimal Control Problems Cadarache Summer School CEA/EDF/INRIA 2012 Pierre Carpentier Jean-Philippe Chancelier Michel De Lara SOWG June 2012 Lecture outline Introduction

More information

LMI Methods in Optimal and Robust Control

LMI Methods in Optimal and Robust Control LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 15: Nonlinear Systems and Lyapunov Functions Overview Our next goal is to extend LMI s and optimization to nonlinear

More information

POD for Parametric PDEs and for Optimality Systems

POD for Parametric PDEs and for Optimality Systems POD for Parametric PDEs and for Optimality Systems M. Kahlbacher, K. Kunisch, H. Müller and S. Volkwein Institute for Mathematics and Scientific Computing University of Graz, Austria DMV-Jahrestagung 26,

More information

Direct Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.

Direct Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Direct Methods Moritz Diehl Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Leuven Belgium Overview Direct Single Shooting Direct Collocation Direct Multiple

More information

Math-Net.Ru All Russian mathematical portal

Math-Net.Ru All Russian mathematical portal Math-Net.Ru All Russian mathematical portal Anna V. Tur, A time-consistent solution formula for linear-quadratic discrete-time dynamic games with nontransferable payoffs, Contributions to Game Theory and

More information

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory Constrained Innite-time Optimal Control Donald J. Chmielewski Chemical Engineering Department University of California Los Angeles February 23, 2000 Stochastic Formulation - Min Max Formulation - UCLA

More information

arxiv: v1 [math.pr] 18 Oct 2016

arxiv: v1 [math.pr] 18 Oct 2016 An Alternative Approach to Mean Field Game with Major and Minor Players, and Applications to Herders Impacts arxiv:161.544v1 [math.pr 18 Oct 216 Rene Carmona August 7, 218 Abstract Peiqi Wang The goal

More information

Tilburg University. Linear Quadratic Games Engwerda, Jacob. Publication date: Link to publication

Tilburg University. Linear Quadratic Games Engwerda, Jacob. Publication date: Link to publication Tilburg University Linear Quadratic Games Engwerda, Jacob Publication date: 26 Link to publication Citation for published version (APA): Engwerda, J. C. (26). Linear Quadratic Games: An Overview. (CentER

More information

A Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices. Abstract

A Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices. Abstract A Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices Kenji Fujiwara School of Economics, Kwansei Gakuin University Abstract We develop the following Stackelberg game model of

More information

The Important State Coordinates of a Nonlinear System

The Important State Coordinates of a Nonlinear System The Important State Coordinates of a Nonlinear System Arthur J. Krener 1 University of California, Davis, CA and Naval Postgraduate School, Monterey, CA ajkrener@ucdavis.edu Summary. We offer an alternative

More information

DECENTRALIZED CONTROL OF INTERCONNECTED SINGULARLY PERTURBED SYSTEMS. A Dissertation by. Walid Sulaiman Alfuhaid

DECENTRALIZED CONTROL OF INTERCONNECTED SINGULARLY PERTURBED SYSTEMS. A Dissertation by. Walid Sulaiman Alfuhaid DECENTRALIZED CONTROL OF INTERCONNECTED SINGULARLY PERTURBED SYSTEMS A Dissertation by Walid Sulaiman Alfuhaid Master of Engineering, Arkansas State University, 2011 Bachelor of Engineering, King Fahd

More information

Lecture 2: Discrete-time Linear Quadratic Optimal Control

Lecture 2: Discrete-time Linear Quadratic Optimal Control ME 33, U Berkeley, Spring 04 Xu hen Lecture : Discrete-time Linear Quadratic Optimal ontrol Big picture Example onvergence of finite-time LQ solutions Big picture previously: dynamic programming and finite-horizon

More information

2.152 Course Notes Contraction Analysis MIT, 2005

2.152 Course Notes Contraction Analysis MIT, 2005 2.152 Course Notes Contraction Analysis MIT, 2005 Jean-Jacques Slotine Contraction Theory ẋ = f(x, t) If Θ(x, t) such that, uniformly x, t 0, F = ( Θ + Θ f x )Θ 1 < 0 Θ(x, t) T Θ(x, t) > 0 then all solutions

More information

A DIFFERENTIAL GAMES APPROACH FOR ANALYSIS OF SPACECRAFT POST-DOCKING OPERATIONS

A DIFFERENTIAL GAMES APPROACH FOR ANALYSIS OF SPACECRAFT POST-DOCKING OPERATIONS A DIFFERENTIAL GAMES APPROACH FOR ANALYSIS OF SPACECRAFT POST-DOCKING OPERATIONS By TAKASHI HIRAMATSU A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT

More information

POSITIVE CONTROLS. P.O. Box 513. Fax: , Regulator problem."

POSITIVE CONTROLS. P.O. Box 513. Fax: ,   Regulator problem. LINEAR QUADRATIC REGULATOR PROBLEM WITH POSITIVE CONTROLS W.P.M.H. Heemels, S.J.L. van Eijndhoven y, A.A. Stoorvogel y Technical University of Eindhoven y Dept. of Electrical Engineering Dept. of Mathematics

More information

Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries!

Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries! Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, 2017 Copyright 2017 by Robert Stengel. All rights reserved. For educational use only. http://www.princeton.edu/~stengel/mae546.html

More information

Computing Minmax; Dominance

Computing Minmax; Dominance Computing Minmax; Dominance CPSC 532A Lecture 5 Computing Minmax; Dominance CPSC 532A Lecture 5, Slide 1 Lecture Overview 1 Recap 2 Linear Programming 3 Computational Problems Involving Maxmin 4 Domination

More information

The first order quasi-linear PDEs

The first order quasi-linear PDEs Chapter 2 The first order quasi-linear PDEs The first order quasi-linear PDEs have the following general form: F (x, u, Du) = 0, (2.1) where x = (x 1, x 2,, x 3 ) R n, u = u(x), Du is the gradient of u.

More information

Answers to exercises in Linear Quadratic Dynamic Optimization and Differential Games.

Answers to exercises in Linear Quadratic Dynamic Optimization and Differential Games. Answers to exercises in Linear Quadratic Dynamic Optimization and Differential Games. Jacob Engwerda Tilburg University Dept. of Econometrics and O.R. P.O. Box:95, 5 LE Tilburg, The Netherlands e-mail:engwerda@uvt.nl

More information

Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries!

Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries! Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, 2018 Copyright 2018 by Robert Stengel. All rights reserved. For educational use only. http://www.princeton.edu/~stengel/mae546.html

More information

Region of attraction approximations for polynomial dynamical systems

Region of attraction approximations for polynomial dynamical systems Region of attraction approximations for polynomial dynamical systems Milan Korda EPFL Lausanne Didier Henrion LAAS-CNRS Toulouse & CTU Prague Colin N. Jones EPFL Lausanne Region of Attraction (ROA) ẋ(t)

More information

Value Function and Optimal Trajectories for some State Constrained Control Problems

Value Function and Optimal Trajectories for some State Constrained Control Problems Value Function and Optimal Trajectories for some State Constrained Control Problems Hasnaa Zidani ENSTA ParisTech, Univ. of Paris-saclay "Control of State Constrained Dynamical Systems" Università di Padova,

More information

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS René Carmona Department of Operations Research & Financial Engineering PACM Princeton University CEMRACS - Luminy, July 17, 217 MFG WITH MAJOR AND MINOR PLAYERS

More information

Game of Defending a Target: A Linear Quadratic Differential Game Approach

Game of Defending a Target: A Linear Quadratic Differential Game Approach Game of Defending a Target: A Linear Quadratic Differential Game Approach Dongxu Li Jose B. Cruz, Jr. Department of Electrical and Computer Engineering, The Ohio State University, 5 Dreese Lab, 5 Neil

More information

Computational Issues in Nonlinear Control and Estimation. Arthur J Krener Naval Postgraduate School Monterey, CA 93943

Computational Issues in Nonlinear Control and Estimation. Arthur J Krener Naval Postgraduate School Monterey, CA 93943 Computational Issues in Nonlinear Control and Estimation Arthur J Krener Naval Postgraduate School Monterey, CA 93943 Modern Control Theory Modern Control Theory dates back to first International Federation

More information

Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem

Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem Noboru Sakamoto, Branislav Rehak N.S.: Nagoya University, Department of Aerospace

More information

A Note on Open Loop Nash Equilibrium in Linear-State Differential Games

A Note on Open Loop Nash Equilibrium in Linear-State Differential Games Applied Mathematical Sciences, vol. 8, 2014, no. 145, 7239-7248 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.49746 A Note on Open Loop Nash Equilibrium in Linear-State Differential

More information

Stabilization and Passivity-Based Control

Stabilization and Passivity-Based Control DISC Systems and Control Theory of Nonlinear Systems, 2010 1 Stabilization and Passivity-Based Control Lecture 8 Nonlinear Dynamical Control Systems, Chapter 10, plus handout from R. Sepulchre, Constructive

More information

Robotics. Control Theory. Marc Toussaint U Stuttgart

Robotics. Control Theory. Marc Toussaint U Stuttgart Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,

More information

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

On the Bellman equation for control problems with exit times and unbounded cost functionals 1 On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road

More information