Game Theory Extra Lecture 1 (BoB)
|
|
- Domenic Melton
- 6 years ago
- Views:
Transcription
1 Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder, pp , , ,
2 To understand Definition of difference/differential games Optimal Control Dynamic programming and HJBI equation Open-loop and feedback Nash Equilibria Copies from Baser/Olsder Material 2
3 Difference Games State equation u i determined by player i x k+1 = f(x k,u 1,u 2,...,u N ), x 1 given Observations y i k = hi k (x k), i = 1,...,N Information structure, for each player a subset η i k of {y 1 1,...,y 1 k;...;y N 1,...,y N k ;u 1 1,...,u 1 k 1;...;u N 1,...,u N k 1} Goal: Player i wants to minimize some functional, L i, of x and u given his available information. Solution concepts: As before, Nash, Stackelberg, etc 3
4 Differential Games d dt x(t) = f(x(t),u 1(t),u 2 (t),...,u N (t)), x(t 0 ) given Goal: to minimize L i = T 0 g i (x,u 1,...,u N ) dt T either fixed, or given by an implicit equation T = min t {t;l(x(t)) 0} 4
5 Example, Robust Control H control Assume u 1,u 2 and y are related via ẋ = Ax+B 1 u 1 +B 2 u 2 z = C 1 x y = C 2 x+du 2 Typical H question: Is z 2 (t)+u 2 1(t)dt γ 2 u 2 2(t)dt? Differential game between controller, u 1 (t) function of y([0,t]) disturbance, u 2 (t) function of x([0,t]) and u 1 ([0,t]) Performance criterium min u 1 max u 2 0 (z 2 +u 2 1 γ 2 u 2 2)dt 5
6 Robust/H Control The interpretation is that u 1 is the control signal and u 2 is a worstcase disturbance signal. Introduce the L 2 (or energy) norm of the signal w w 2 = ( 0 ) 1/2 w(t) 2 dt The H norm of a linear system G(s) is defined by G = sup w 0 Gw w 6
7 Robustness The H norm measures the largest amplification of energy by the system. Minimization of the norm is clearly interesting if w is a disturbance signal. Another motivation is given by the so called small-gain theorem Theorem A closed loop system is stable for all perturbations with norm < γ if and only if G K 1/γ G Κ 7
8 H control G K < γ K : z 2 γ 2 w 2 < 0 min u max w ( z 2 γ 2 w 2 ) < 0 This is exactly (the upper value of) an affine quadratic game The relation between H control and game theory was noted rather late (end of 80s). For details see Section 6.6 8
9 Pursuit Evasion Games 9
10 Tools from One-Person Optimization Dynamic programming The (maximum) minimum principle Baser/Olsder Ch
11 Dynamic Programming, discrete time x k+1 = f k (x k,u k ), u k U k K L(u) = g k (x k+1,u k,x k ) k=1 f k,g k,u k,k,x 1 given. Want u k = γ k (x k ) that minimizes L Idea: Generalize the problem; Calculate the value function V(k,x) = min γ k,...,γ K for all initial conditions x k = x. K i=k g i (x i+1,u i,x k ) 11
12 Principle of Optimality An optimal control sequence u 1,...,u K should be sequentially optimal. The only coupling between the optimization problems on time horizons [1,k 1] and [k,k] is via the state x k. This leads to V(k,x) = min u k U k [g k (f k (x,u k ),u k,x)+v(k +1,f k (x,u k ))] with the final condition V(K,x) = min u K g k (x K+1,u K,x K ). 12
13 Example Affine quadratic problems has the solution x k+1 = A k x k +B k u k +c k L(u) = 1 K (x 2 k+1qx k+1 +u kru k ) k=1 V(k,x) = 1 2 x S k x+x s k +q k u k = P k S k+1 A k x k P k (s k+1 +S k+1 c k ) Formulas for P k,s k,s k are given on p
14 Dynamic Programming, continuous-time The same reasoning leads to a PDE equation called the Hamilton- Jacobi-Bellman (HJB) equation ẋ(t) = f(x(t),u(t)) u(t) = γ(x(t)) L(u) = T 0 g(x(t),u(t)dt+q(t,x(t)) The final time T can be either fixed known, or given implicitly by T = min{t : l(x(t)) = 0} t 0 14
15 HJB equation T V(t,x) = min g(x(s),u(s))ds+q(t,x(t)) {u(s),s [t,t]} t V(T,x) = q(t,x) along l(x) = 0 Principle of optimality shows that if V is C 1 then V(t,x) [ ] V(t,x) = min f(x,u)+g(x,u) t u U x (+ final condition when t = T) 15
16 Theorem of Sufficiency Theorem 5.3 p 237 If a C 1 function V can be found that satisfies the HJB equation and boundary conditions above then it generates the optimal strategy u through the pointwise optimization problem defined by the right hand side. Proof see p. 237 Example: Affine-quadratic problems has a quadratic function V(t,x) = 1 2 x S(t)x+k (t)x+m(t), see p
17 The Minimum Principle Introduce the costate vector p (t) = V(t,x (t))/ x where x denotes the optimal trajectory corresponding to u, i.e. ẋ (t) = f(x (t),u (t)). Also define (the Hamiltonian) Theorem 5.4 H(t,p,x,u) = g(x,u)+p (t)f(x,u) p (T) = q(t(x ),x ) along l(t,x) = 0 x ṗ (t) = H(t,p,x,u ) x u = arg min u U H(t,p,x,u) See also the discrete-time counterpart Theorem
18 N person difference games How does this translate to difference games with N > 1 players? x k+1 = f k (x k,u 1 k,...,u N k ) K J i (u 1,...,u N ) = gk(x i k+1,u 1 k,...,u N k,x k ) k=1 The control laws u i = γ i constitute a Nash equilibrium if Open-loop information: u i k J i (γ ) J i ({γ i,γ i }) i is an open-loop function of k Closed-loop information: u i k (x k) is allowed to be a function of x k. Hence changing u i will result in changes in u j 18
19 See Section Open Loop Dynamic Games Idea: Player i solves a one-player problem. His choice u i will not influence the other u j, so these can be treated as given functions of time. Then the results from the N = 1 apply directly 19
20 Open Loop Theorems Theorem 6.1 If u = γ provides an open-loop Nash equilibrium, then there exists a sequence of costate vectors p i such that x k+1 = f k (x k,u k) γ i k = arg min u i k p i k = x k f k H i k(p i k+1,{γ i k,u i k},x k) [ p i k+1 + ( ) ] gk i x k+1 where H i = g i +p i f. For details see p ( ) gk i x k The affine case results in coupled Riccati equations. The special zero-sum case for N = 2 is given in Theorem 6.3, and 6.4. Only one Riccati equation must be solved. The condition for existence of Nash equilibrium has the form I B 2 k S k+1b 2 k > 0. 20
21 Feedback Solutions Theorem 6.6 The set of strategies γ provides a feedback Nash equilibrium if and only if there exists functions V i (k,x) such that V i (k,x) = arg min[gk( i,u i i k)+v i (k +1, u i k = g i k( i,γ i k )+V i (k +1, f i k (x,γ i k )) f i k (x,u i k))] where f i k (x,ui k ) = f k(x,{γ i (x),u i k }) 21
22 Feedback Solutions, Zero-sum case N = 2 See Corollary 6.2 p. 282 The set of strategies γ 1,γ 2 provides a feedback saddle-point solution if and only if there exists functions V(k,x) such that for all k V(k,x) = min u 1 k = max u 2 k max u 2 k min u 1 k [ gk (f k (x,u 1 k,u 2 k),u 1 k,u 2 k,x)+v(k +1,f k (x,u 1 k,u 2 k) ] [ gk (f k (x,u 1 k,u 2 k),u 1 k,u 2 k,x)+v(k +1,f k (x,u 1 k,u 2 k) ] = g k (f k (x,γ 1 k,γ 2 k ),γ 1 k (x),γ 2 k (x),x)+v(k +1,f k (x,γ 1 k (x),γ 2 k (x))) 22
23 Differential Games Open-loop Nash Equilibria Want to present counterparts for continuous time. Existence of smooth cost function V is more problematic. L i (u 1,...,u N ) = = ẋ(t) = f(x(t),u 1 (t),...,u N (t)) T 0 g i (x(t),u 1 (t),...,u N (t))dt+q i (x(t)) 23
24 Open-loop Nash Equilibria Theorem 6.11 (Some smoothness assumptions) If u i = γ i (t,x 0 ) provides an open-loop Nash equilibrium (with corresponding x ) then there exist N costate functions p i (t) such that ẋ (t) = f(x (t),u 1 (t),...,u N (t)) γ i = arg min u i U i H i (p i (t),x (t),{u i,u i }) ṗ i (t) = x Hi (p i (t),x,u (t)) ṗ i (T) = x qi (x (T)) where H i (p i,x,u) = g i (x,u)+p i f(x,u). 24
25 The Affine-Quadratic Case f(x,u) = A(t)x+ B i (t)u i +c(t) g i (x,u) = 1 2 x Q i (t)x+ u j R ij u j q i (x(t)) = 1 2 x (T)Q i fx(t) with Q i 0,Q i f 0,Rii > 0. 25
26 The Affine-Quadratic Case Theorem 6.12+p.316 There is an open-loop Nash equilibrium solution to the affine-quadratic game for any T [0,T f ] if and only if the following coupled matrix Riccati differential equations have solutions for any T [0,T f ] M i +M i A+A M i +Q i M i j B j R jj (t) 1 B j M j = 0 M i (T) = Q i f The NE is given by u i = R ii (t) 1 B i [M i x (t) + m i ] where the feedforward signal m i is given by (6.51) 26
27 P1 minimizes, P2 maximizes. L(u 1,u 2 ) = 1 2 N = 2 Zero-Sum Case T with Q 0, Q f 0. 0 x Qx+u 1 u 1 u 2 u 2 dt+ 1 2 x (T)Q f x(t) Assume there exists a unique bounded symmetric solution S( ) to the matrix equation Ṡ +A S +SA+Q+SB 2 B 2 S = 0, S(T) = Q f on the interval [0,T]. 27
28 Then there exists a solution to Ṁ +A M +MA+Q M(B 1 N 1 N 2 B 2 )M = 0;M(T) = Q f and the game admits a unique open-loop saddle-point given by γ 1 (t,x 0 ) = B 1 (t) [M(t)x (t)+m(t)] γ 2 (t,x 0 ) = B 2 (t) [M(t)x (t)+m(t)] 28
29 Closed-loop Feedback Nash Equilibria Will only give the result for N = 2 and zero-sum situation Corollary 6.6, p326 A pair of strategies {γ i } provides a feedback saddle-point solution if there exists a function V satisfying the PDE V(t,x) [ ] V(t,x) = minmax f(x,u 1,u 2 )+g(x,u 1,u 2 ) t u 1 u 2 x [ ] V(t,x) = maxmin f(x,u 1,u 2 )+g(x,u 1,u 2 ) u 2 u 1 x [ ] V(t,x) = f(x,γ 1,γ 2 )+g(x,γ 1,γ 2 ) x The value of the game is V(0,x 0 ) This is the famous Hamilton-Jacobi-Bellman-Isaacs equation obtained by Isaacs in 1950s. 29
30 N = 2, Zero-sum Affine Case ẋ = Ax+B 1 u 1 +B 2 u 2 +c L = 1 2 T 0 x Qx+u 1 u 1 u 2 u 2 dt Theorem 6.17 If there exists a unique symmetric bounded solution to the Riccati equation Ż +A Z +ZA+Q Z(B 1 B 1 B 2 B 2 )Z = 0, Z(T) = Q f then the two-player zero-sum game admits a unique feedback saddle-point given by (for details see p. 327) γ i (t,x) = ( 1) i B i (t) [Z(t)x(t)+ζ(t)] 30
OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationHamilton-Jacobi-Bellman Equation Feb 25, 2008
Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationFormula Sheet for Optimal Control
Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming
More informationStability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games
Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,
More informationEE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games
EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations
More informationLinear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,
More informationRisk-Sensitive and Robust Mean Field Games
Risk-Sensitive and Robust Mean Field Games Tamer Başar Coordinated Science Laboratory Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL - 6181 IPAM
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationThermodynamic limit and phase transitions in non-cooperative games: some mean-field examples
Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Conference on the occasion of Giovanni Colombo and Franco Ra... https://events.math.unipd.it/oscgc18/ Optimization,
More informationHomework Solution # 3
ECSE 644 Optimal Control Feb, 4 Due: Feb 17, 4 (Tuesday) Homework Solution # 3 1 (5%) Consider the discrete nonlinear control system in Homework # For the optimal control and trajectory that you have found
More informationNonlinear Control. Nonlinear Control Lecture # 25 State Feedback Stabilization
Nonlinear Control Lecture # 25 State Feedback Stabilization Backstepping η = f a (η)+g a (η)ξ ξ = f b (η,ξ)+g b (η,ξ)u, g b 0, η R n, ξ, u R Stabilize the origin using state feedback View ξ as virtual
More informationOPTIMAL CONTROL SYSTEMS
SYSTEMS MIN-MAX Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University OUTLINE MIN-MAX CONTROL Problem Definition HJB Equation Example GAME THEORY Differential Games Isaacs
More informationChap. 3. Controlled Systems, Controllability
Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :
More informationAn Introduction to Noncooperative Games
An Introduction to Noncooperative Games Alberto Bressan Department of Mathematics, Penn State University Alberto Bressan (Penn State) Noncooperative Games 1 / 29 introduction to non-cooperative games:
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-Time Nonlinear Quadratic Optimal Control V. Manousiouthakis D. Chmielewski Chemical Engineering Department UCLA 1998 AIChE Annual Meeting Outline Unconstrained Innite-Time Nonlinear
More informationLinear Quadratic Zero-Sum Two-Person Differential Games
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,
More informationRobust control and applications in economic theory
Robust control and applications in economic theory In honour of Professor Emeritus Grigoris Kalogeropoulos on the occasion of his retirement A. N. Yannacopoulos Department of Statistics AUEB 24 May 2013
More informationLecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster.
Lecture 8 Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationAn Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game
An Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game Marzieh Khakestari (Corresponding author) Institute For Mathematical Research, Universiti Putra Malaysia, 43400
More informationMulti-Pursuer Single-Evader Differential Games with Limited Observations*
American Control Conference (ACC) Washington DC USA June 7-9 Multi-Pursuer Single- Differential Games with Limited Observations* Wei Lin Student Member IEEE Zhihua Qu Fellow IEEE and Marwan A. Simaan Life
More informationComputational Issues in Nonlinear Dynamics and Control
Computational Issues in Nonlinear Dynamics and Control Arthur J. Krener ajkrener@ucdavis.edu Supported by AFOSR and NSF Typical Problems Numerical Computation of Invariant Manifolds Typical Problems Numerical
More informationStatic and Dynamic Optimization (42111)
Static and Dynamic Optimization (421) Niels Kjølstad Poulsen Build. 0b, room 01 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark Email:
More informationThe Linear Quadratic Regulator
10 The Linear Qadratic Reglator 10.1 Problem formlation This chapter concerns optimal control of dynamical systems. Most of this development concerns linear models with a particlarly simple notion of optimality.
More informationMin-Max Certainty Equivalence Principle and Differential Games
Min-Max Certainty Equivalence Principle and Differential Games Pierre Bernhard and Alain Rapaport INRIA Sophia-Antipolis August 1994 Abstract This paper presents a version of the Certainty Equivalence
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationAlberto Bressan. Department of Mathematics, Penn State University
Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i
More informationAdvanced Mechatronics Engineering
Advanced Mechatronics Engineering German University in Cairo 21 December, 2013 Outline Necessary conditions for optimal input Example Linear regulator problem Example Necessary conditions for optimal input
More informationNumerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini
Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33
More informationpacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre
HJ pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre Control of State Constrained Dynamical Systems 25-29 September 2017, Padova Main
More informationOn the Stability of the Best Reply Map for Noncooperative Differential Games
On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University
More informationModel Predictive Regulation
Preprints of the 19th World Congress The International Federation of Automatic Control Model Predictive Regulation Cesar O. Aguilar Arthur J. Krener California State University, Bakersfield, CA, 93311,
More informationNonlinear Control Systems
Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 5. Input-Output Stability DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Input-Output Stability y = Hu H denotes
More informationNonlinear Control Systems
Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f
More informationJosé C. Geromel. Australian National University Canberra, December 7-8, 2017
5 1 15 2 25 3 35 4 45 5 1 15 2 25 3 35 4 45 5 55 Differential LMI in Optimal Sampled-data Control José C. Geromel School of Electrical and Computer Engineering University of Campinas - Brazil Australian
More informationESC794: Special Topics: Model Predictive Control
ESC794: Special Topics: Model Predictive Control Nonlinear MPC Analysis : Part 1 Reference: Nonlinear Model Predictive Control (Ch.3), Grüne and Pannek Hanz Richter, Professor Mechanical Engineering Department
More informationLecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.
Lecture 4 Chapter 4: Lyapunov Stability Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 4 p. 1/86 Autonomous Systems Consider the autonomous system ẋ
More informationOptimal Control Theory
Optimal Control Theory The theory Optimal control theory is a mature mathematical discipline which provides algorithms to solve various control problems The elaborate mathematical machinery behind optimal
More informationGame Theoretic Continuous Time Differential Dynamic Programming
Game heoretic Continuous ime Differential Dynamic Programming Wei Sun, Evangelos A. heodorou and Panagiotis siotras 3 Abstract In this work, we derive a Game heoretic Differential Dynamic Programming G-DDP
More informationControl of Dynamical System
Control of Dynamical System Yuan Gao Applied Mathematics University of Washington yuangao@uw.edu Spring 2015 1 / 21 Simplified Model for a Car To start, let s consider the simple example of controlling
More informationLecture Note 7: Switching Stabilization via Control-Lyapunov Function
ECE7850: Hybrid Systems:Theory and Applications Lecture Note 7: Switching Stabilization via Control-Lyapunov Function Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio
More informationDifferential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011
Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.
More information1. Find the solution of the following uncontrolled linear system. 2 α 1 1
Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +
More informationInput to state Stability
Input to state Stability Mini course, Universität Stuttgart, November 2004 Lars Grüne, Mathematisches Institut, Universität Bayreuth Part IV: Applications ISS Consider with solutions ϕ(t, x, w) ẋ(t) =
More informationOptimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:
Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic
More informationOutput Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x)
Output Feedback and State Feedback EL2620 Nonlinear Control Lecture 10 Exact feedback linearization Input-output linearization Lyapunov-based control design methods ẋ = f(x,u) y = h(x) Output feedback:
More informationECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming
ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear
More informationSUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS
ICIC Express Letters ICIC International c 2008 ISSN 1881-803X Volume 2, Number 4, December 2008 pp. 371-377 SUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS Ma. Aracelia
More informationRiccati Equations and Inequalities in Robust Control
Riccati Equations and Inequalities in Robust Control Lianhao Yin Gabriel Ingesson Martin Karlsson Optimal Control LP4 2014 June 10, 2014 Lianhao Yin Gabriel Ingesson Martin Karlsson (LTH) H control problem
More informationConverse Lyapunov theorem and Input-to-State Stability
Converse Lyapunov theorem and Input-to-State Stability April 6, 2014 1 Converse Lyapunov theorem In the previous lecture, we have discussed few examples of nonlinear control systems and stability concepts
More informationState-Feedback Optimal Controllers for Deterministic Nonlinear Systems
State-Feedback Optimal Controllers for Deterministic Nonlinear Systems Chang-Hee Won*, Abstract A full-state feedback optimal control problem is solved for a general deterministic nonlinear system. The
More informationSuboptimality of minmax MPC. Seungho Lee. ẋ(t) = f(x(t), u(t)), x(0) = x 0, t 0 (1)
Suboptimality of minmax MPC Seungho Lee In this paper, we consider particular case of Model Predictive Control (MPC) when the problem that needs to be solved in each sample time is the form of min max
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationZ i Q ij Z j. J(x, φ; U) = X T φ(t ) 2 h + where h k k, H(t) k k and R(t) r r are nonnegative definite matrices (R(t) is uniformly in t nonsingular).
2. LINEAR QUADRATIC DETERMINISTIC PROBLEM Notations: For a vector Z, Z = Z, Z is the Euclidean norm here Z, Z = i Z2 i is the inner product; For a vector Z and nonnegative definite matrix Q, Z Q = Z, QZ
More informationLecture 15: H Control Synthesis
c A. Shiriaev/L. Freidovich. March 12, 2010. Optimal Control for Linear Systems: Lecture 15 p. 1/14 Lecture 15: H Control Synthesis Example c A. Shiriaev/L. Freidovich. March 12, 2010. Optimal Control
More informationProf. Krstic Nonlinear Systems MAE281A Homework set 1 Linearization & phase portrait
Prof. Krstic Nonlinear Systems MAE28A Homework set Linearization & phase portrait. For each of the following systems, find all equilibrium points and determine the type of each isolated equilibrium. Use
More informationIntroduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability Goal of This
More informationMean-Field Games with non-convex Hamiltonian
Mean-Field Games with non-convex Hamiltonian Martino Bardi Dipartimento di Matematica "Tullio Levi-Civita" Università di Padova Workshop "Optimal Control and MFGs" Pavia, September 19 21, 2018 Martino
More informationChain differentials with an application to the mathematical fear operator
Chain differentials with an application to the mathematical fear operator Pierre Bernhard I3S, University of Nice Sophia Antipolis and CNRS, ESSI, B.P. 145, 06903 Sophia Antipolis cedex, France January
More informationPontryagin s maximum principle
Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin
More informationChapter Introduction. A. Bensoussan
Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688
More informationA Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations
2005 American Control Conference June 8-10, 2005. Portland, OR, USA WeB10.3 A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations Arik Melikyan, Andrei Akhmetzhanov and Naira
More informationComputing Minmax; Dominance
Computing Minmax; Dominance CPSC 532A Lecture 5 Computing Minmax; Dominance CPSC 532A Lecture 5, Slide 1 Lecture Overview 1 Recap 2 Linear Programming 3 Computational Problems Involving Maxmin 4 Domination
More informationECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko
ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko Continuous-time optimization involves maximization wrt to an innite dimensional object, an entire
More informationLocally optimal controllers and application to orbital transfer (long version)
9th IFAC Symposium on Nonlinear Control Systems Toulouse, France, September 4-6, 13 FrA1.4 Locally optimal controllers and application to orbital transfer (long version) S. Benachour V. Andrieu Université
More informationOptimal Control. Lecture 20. Control Lyapunov Function, Optimal Estimation. John T. Wen. April 5. Ref: Papers by R. Freeman (on-line), B&H Ch.
Optimal Control Lecture 20 Control Lyapunov Function, Optimal Estimation John T. Wen April 5 Ref: Papers by R. Freeman (on-line), B&H Ch. 12 Outline Summary of Control Lyapunov Function and example Introduction
More informationModeling and Analysis of Dynamic Systems
Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability
More informationQuadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University
.. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if
More informationMax Plus Eigenvector Methods for Nonlinear H Problems: Error Analysis
Max Plus Eigenvector Methods for Nonlinear H Problems: Error Analysis William M. McEneaney August 28, 2003 Abstract The H problem for a nonlinear system is considered. The corresponding dynamic programming
More informationDynamic Consistency for Stochastic Optimal Control Problems
Dynamic Consistency for Stochastic Optimal Control Problems Cadarache Summer School CEA/EDF/INRIA 2012 Pierre Carpentier Jean-Philippe Chancelier Michel De Lara SOWG June 2012 Lecture outline Introduction
More informationLMI Methods in Optimal and Robust Control
LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 15: Nonlinear Systems and Lyapunov Functions Overview Our next goal is to extend LMI s and optimization to nonlinear
More informationPOD for Parametric PDEs and for Optimality Systems
POD for Parametric PDEs and for Optimality Systems M. Kahlbacher, K. Kunisch, H. Müller and S. Volkwein Institute for Mathematics and Scientific Computing University of Graz, Austria DMV-Jahrestagung 26,
More informationDirect Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.
Direct Methods Moritz Diehl Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Leuven Belgium Overview Direct Single Shooting Direct Collocation Direct Multiple
More informationMath-Net.Ru All Russian mathematical portal
Math-Net.Ru All Russian mathematical portal Anna V. Tur, A time-consistent solution formula for linear-quadratic discrete-time dynamic games with nontransferable payoffs, Contributions to Game Theory and
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-time Optimal Control Donald J. Chmielewski Chemical Engineering Department University of California Los Angeles February 23, 2000 Stochastic Formulation - Min Max Formulation - UCLA
More informationarxiv: v1 [math.pr] 18 Oct 2016
An Alternative Approach to Mean Field Game with Major and Minor Players, and Applications to Herders Impacts arxiv:161.544v1 [math.pr 18 Oct 216 Rene Carmona August 7, 218 Abstract Peiqi Wang The goal
More informationTilburg University. Linear Quadratic Games Engwerda, Jacob. Publication date: Link to publication
Tilburg University Linear Quadratic Games Engwerda, Jacob Publication date: 26 Link to publication Citation for published version (APA): Engwerda, J. C. (26). Linear Quadratic Games: An Overview. (CentER
More informationA Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices. Abstract
A Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices Kenji Fujiwara School of Economics, Kwansei Gakuin University Abstract We develop the following Stackelberg game model of
More informationThe Important State Coordinates of a Nonlinear System
The Important State Coordinates of a Nonlinear System Arthur J. Krener 1 University of California, Davis, CA and Naval Postgraduate School, Monterey, CA ajkrener@ucdavis.edu Summary. We offer an alternative
More informationDECENTRALIZED CONTROL OF INTERCONNECTED SINGULARLY PERTURBED SYSTEMS. A Dissertation by. Walid Sulaiman Alfuhaid
DECENTRALIZED CONTROL OF INTERCONNECTED SINGULARLY PERTURBED SYSTEMS A Dissertation by Walid Sulaiman Alfuhaid Master of Engineering, Arkansas State University, 2011 Bachelor of Engineering, King Fahd
More informationLecture 2: Discrete-time Linear Quadratic Optimal Control
ME 33, U Berkeley, Spring 04 Xu hen Lecture : Discrete-time Linear Quadratic Optimal ontrol Big picture Example onvergence of finite-time LQ solutions Big picture previously: dynamic programming and finite-horizon
More information2.152 Course Notes Contraction Analysis MIT, 2005
2.152 Course Notes Contraction Analysis MIT, 2005 Jean-Jacques Slotine Contraction Theory ẋ = f(x, t) If Θ(x, t) such that, uniformly x, t 0, F = ( Θ + Θ f x )Θ 1 < 0 Θ(x, t) T Θ(x, t) > 0 then all solutions
More informationA DIFFERENTIAL GAMES APPROACH FOR ANALYSIS OF SPACECRAFT POST-DOCKING OPERATIONS
A DIFFERENTIAL GAMES APPROACH FOR ANALYSIS OF SPACECRAFT POST-DOCKING OPERATIONS By TAKASHI HIRAMATSU A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT
More informationPOSITIVE CONTROLS. P.O. Box 513. Fax: , Regulator problem."
LINEAR QUADRATIC REGULATOR PROBLEM WITH POSITIVE CONTROLS W.P.M.H. Heemels, S.J.L. van Eijndhoven y, A.A. Stoorvogel y Technical University of Eindhoven y Dept. of Electrical Engineering Dept. of Mathematics
More informationOptimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries!
Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, 2017 Copyright 2017 by Robert Stengel. All rights reserved. For educational use only. http://www.princeton.edu/~stengel/mae546.html
More informationComputing Minmax; Dominance
Computing Minmax; Dominance CPSC 532A Lecture 5 Computing Minmax; Dominance CPSC 532A Lecture 5, Slide 1 Lecture Overview 1 Recap 2 Linear Programming 3 Computational Problems Involving Maxmin 4 Domination
More informationThe first order quasi-linear PDEs
Chapter 2 The first order quasi-linear PDEs The first order quasi-linear PDEs have the following general form: F (x, u, Du) = 0, (2.1) where x = (x 1, x 2,, x 3 ) R n, u = u(x), Du is the gradient of u.
More informationAnswers to exercises in Linear Quadratic Dynamic Optimization and Differential Games.
Answers to exercises in Linear Quadratic Dynamic Optimization and Differential Games. Jacob Engwerda Tilburg University Dept. of Econometrics and O.R. P.O. Box:95, 5 LE Tilburg, The Netherlands e-mail:engwerda@uvt.nl
More informationOptimal Control and Estimation MAE 546, Princeton University Robert Stengel, Preliminaries!
Optimal Control and Estimation MAE 546, Princeton University Robert Stengel, 2018 Copyright 2018 by Robert Stengel. All rights reserved. For educational use only. http://www.princeton.edu/~stengel/mae546.html
More informationRegion of attraction approximations for polynomial dynamical systems
Region of attraction approximations for polynomial dynamical systems Milan Korda EPFL Lausanne Didier Henrion LAAS-CNRS Toulouse & CTU Prague Colin N. Jones EPFL Lausanne Region of Attraction (ROA) ẋ(t)
More informationValue Function and Optimal Trajectories for some State Constrained Control Problems
Value Function and Optimal Trajectories for some State Constrained Control Problems Hasnaa Zidani ENSTA ParisTech, Univ. of Paris-saclay "Control of State Constrained Dynamical Systems" Università di Padova,
More informationMEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS
MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS René Carmona Department of Operations Research & Financial Engineering PACM Princeton University CEMRACS - Luminy, July 17, 217 MFG WITH MAJOR AND MINOR PLAYERS
More informationGame of Defending a Target: A Linear Quadratic Differential Game Approach
Game of Defending a Target: A Linear Quadratic Differential Game Approach Dongxu Li Jose B. Cruz, Jr. Department of Electrical and Computer Engineering, The Ohio State University, 5 Dreese Lab, 5 Neil
More informationComputational Issues in Nonlinear Control and Estimation. Arthur J Krener Naval Postgraduate School Monterey, CA 93943
Computational Issues in Nonlinear Control and Estimation Arthur J Krener Naval Postgraduate School Monterey, CA 93943 Modern Control Theory Modern Control Theory dates back to first International Federation
More informationIterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem
Iterative methods to compute center and center-stable manifolds with application to the optimal output regulation problem Noboru Sakamoto, Branislav Rehak N.S.: Nagoya University, Department of Aerospace
More informationA Note on Open Loop Nash Equilibrium in Linear-State Differential Games
Applied Mathematical Sciences, vol. 8, 2014, no. 145, 7239-7248 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.49746 A Note on Open Loop Nash Equilibrium in Linear-State Differential
More informationStabilization and Passivity-Based Control
DISC Systems and Control Theory of Nonlinear Systems, 2010 1 Stabilization and Passivity-Based Control Lecture 8 Nonlinear Dynamical Control Systems, Chapter 10, plus handout from R. Sepulchre, Constructive
More informationRobotics. Control Theory. Marc Toussaint U Stuttgart
Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,
More informationOn the Bellman equation for control problems with exit times and unbounded cost functionals 1
On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road
More information