Static and Dynamic Optimization (42111)
|
|
- Ira Blake
- 5 years ago
- Views:
Transcription
1 Static and Dynamic Optimization (421) Niels Kjølstad Poulsen Build. 0b, room 01 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark phone: mobile: :50 Lecture : Dynamic Programming 1/
2 Outline of lecture Dynamic Programming Recap F Dynamic Programming (D) Unconstrained Constrained Reading guidance (DO: chapter, page -) 2/
3 Dynamic Optimization (D) Find a sequence u i U i, i = 0,..., N 1 which takes the system x i+1 = f i (x i,u i ) x 0 = x 0 ψ(x N ) = 0 from its initial state x 0 along a trajectory such that the performance index N 1 J = φ N [x N ]+ i=0 L i (x i,u i ) is minimized. Define the Hamiltonian function as: H i = L i (x i,u i )+λ T i+1 f i(x i,u i ) The necessary conditions are: x i+1 = f i (x i,u i ) λ T i = x i H i ( u i = arg min H i 0 T = ) H i u i U i u i with boundary conditions: x 0 = x 0 λ T N = νt x N ψ + x N φ Dynamic Optimization (C) Find a function u t U t t [0; T] R which takes the system system ẋ t = f t(x t,u t) x 0 = x 0 ψ(x T ) = 0 from its initial state x 0 along a trajectory such that the performance index T J = φ T [x T ]+ L t(x t,u t) dt 0 is minimized. Define the Hamilton function as: H t = L t(x t,u t)+λ T t f t(x t,u t) The necessary conditions are: ẋ t = f t(x t,u t) λ T t = x t H t ( u t = arg min H t 0 T = ) H t u t U t u t with boundary conditions: x 0 = x 0 λ T T = νt x T ψ + x T φ /
4 Dynamic Programming Found in Bertsekas. Kierkegaard: Life can only be understood going backwards, but it must be lived going forwards. Let us (for a start) now focus on: Discrete independent variable (normally the time). Discrete and finite state space and decision space ( X i and U i ) 4/
5 Stagecoach problem Example (Hans Ravn: Statisk og Dynamisk Optimering) Go from A to B with minimal cost. 4 x 14 9 A: London B: Esbjerg i 5/
6 Dynamic Programming 4 x 14 9 A: London B: Esbjerg N 1 J = φ N (x N )+ L i (x i,u i ) i= i Objective x i+1 = f i (x i,u i ) x 0 = x 0 Dynamics (x i,u i ) V i Constraints x N V N /
7 Dynamic Programming Let us start with the easy part - at the end (i.e. i = ) A: London B: Esbjerg Here we easily find the optimal decisions: 1 for x = u (x ) = 0 for x = 2 1 for x = 1 and the optimal costs: for x = V (x ) = for x = 2 for x = /
8 Dynamic Programming For i = 2 we have: A: London 1 B: Esbjerg the optimal decisions: 1 for x 2 = u 2 (x 2) = 1 for x 2 = 2 1 for x 2 = 1 and the optimal costs: for x 2 = V 2 (x 2 ) = 1 for x 2 = 2 14 for x 2 = /
9 Dynamic Programming For i = 1 we have: A: London 22 1 B: Esbjerg the optimal decisions: 0 for x 1 = u 1 (x 1) = 1 for x 1 = 2 0 for x 1 = 1 and the optimal costs: 21 for x 1 = V 1 (x 1 ) = 22 for x 1 = 2 2 for x 1 = 1 9/
10 Dynamic Programming Finally, for i = 0 we have: A: London B: Esbjerg the optimal decisions: u 0(x 0 ) = 0 since x 0 = 2 and the optimal cost: V 0 (x 0 ) = 2 since x 0 = 2 /
11 Dynamic Programming Let us return to i = 1 and see what we actually did A: London 22 1 B: Esbjerg We had the optimal cost to go, V 2 (x 2 ), for the next stage stored ie. for x 2 = V 2 (x 2 ) = 1 for x 2 = 2 14 for x 2 = 1 as well as the optimal decision sequence. For each possible value of x 1 (x 1 = 1,2,), we evaluated the loss for each possible decision (e.g. u 1 = 1,0,1 for x 1 = 2), and minimized wrt. u 1 : L 1 (x 1,u 1 )+V 2 (f 1 (x 1,u 1 )) because x 2 = f 1 (x 1,u 1 ) That minimization results in V 1 (x 1 ). /
12 Dynamic Programming A: London 22 1 B: Esbjerg For each possible value of x 1 (1,2,): { } V 1 (x 1 ) = min L 1 (x 1,u 1 )+V 2 (f 1 (x 1,u 1 )) u 1 (x 1,u 1 ) V 1 Or in general (for i =,2,1,0): { } V i (x i ) = min L(x i,u i )+V i+1 (f i (x i,u i )) u i (x i,u i ) V i 12/
13 Free Dynamic Programming Let us now focus on finding a sequence of decisions u i i = 0,,1,... N which takes the system x i+1 = f i (x i,u i ) x 0 = x 0 along a trajectory, such that the cost function N 1 J = φ(x N )+ L i (x i,u i )= J 0 (x 0,u N 1 0 ) i=0 is minimized i i+1 N Def.: u k i the sequence of decision from i to k. The truncated performance index (the cost to go) N 1 J i (x i,u N 1 i ) = φ(x N )+ L k (x k,u k ) k=i It is quite easy to see that J i (x i,u N 1 i ) = L i (x i,u i )+J i+1 (x i+1,u N 1 i+1 ) and that in particular J = J 0 (x 0,u N 1 0 ) J N = φ(x N ) 1/
14 Free Dynamic Programming The Bellman function (the optimal cost to go) is defined as: V i (x i ) = min u N 1 i J i (x i,u N 1 i ) and is a function of the present state, x i, and index, i. In particular V N (x N ) = φ N (x N ) Theorem The Bellman function V i, is given by the backwards recursion [ V i (x i ) = min u i L i (x i,u i )+V i+1 (x i+1 ) ] with the boundary condition V N (x N ) = φ N (x N ) Bellman equation is a functional equation, gives a sufficient condition and the overall optimimum is given as V 0 (x 0 ) = J. 14/
15 Free Dynamic Programming What about the decisions? [ u i = arg min u i L i (x i,u i )+V i+1 ( f i (x i,u i ) ) }{{} x i+1 ] = u i (x i ) A maximization problem: min max. 15/
16 Proof. By definition we have: V i (x i ) = min u N 1 i = min u N 1 i J i (x i,u N 1 i ) Since u N 1 i+1 do not affect L i we can write V i (x i ) = min u i [ ] L i (x i,u i )+J i+1 (x i+1,u N 1 i+1 ) L i (x i,u i )+ min u N 1 i+1 J i+1 (x i+1,u N 1 i+1 ) The last term is nothing but V i+1, due to the definition of the Bellman function. The boundary condition is also given by the definition of the Bellman function. 1/
17 Pause 1/
18 LQ problem Simple LQ problem: The problem is bring the system x i+1 = ax i +bu i x 0 = x 0 from the initial state along a trajectory such the performance index N 1 J = px 2 N + is minimized. i=0 qx 2 i +ru2 i In the boundary we have V N (x N ) = φ(x N ) = px 2 N Inspired of this, we will try the candidate function V i = s i x 2 i 1/
19 The Bellman equation [ V i (x i ) = min L i (x i,u i )+V i+1 (x i+1 ) u i gives in this special (scalar LQ) case: s i x 2 i = min u i [qx 2 i +ru2 i }{{} L i (x i,u i ) + s i+1 x 2 i+1 }{{} V i+1 (x i+1 ) ] ] (cut an paste) or with the state equation inserted [ s i x 2 i = min qx 2 u i +ru2 i +s i+1(ax i +bu i ) 2] i }{{} f i (x i,u i ) The minimum is obtained for u i = abs i+1 r +b 2 s i+1 x i which inserted in the recursion results in: [ s i x 2 i = q +a 2 s i+1 a2 b 2 s 2 ] i+1 r +b 2 x 2 i s i+1 19/
20 V The candidate function satisfies the Bellman equation if s i = q +a 2 s i+1 a2 b 2 s 2 i+1 r +b 2 s i+1 s N = p which is the (scalar version of the) Riccati equation. The solution: s i+1 and u i = abs i+1 r +b 2 s i+1 x i 25 V t (x) and 20 V i (x) = s i x 2 15 x t time (i) x Method: Guess the type of functionallity in V i (x) i.e. up to a number of parameter. Check if it satisfy the Bellman equation. This results in a (number of) recursion(s) for the parameter(s). 20/
21 Euler-Lagrange and Bellman [ V i (x i ) = min u i L i (x i,u i )+V i+1 (x i+1 ) ] Necessary condition (stationarity): 0 T = L i u i + V i+1 x i+1 x i+1 u i x i+1 = f i (x i,u i ) If the costate (or adjoint state) is defined as the sensitivity, i.e. as: then λ T i = V i(x i ) x i or: 0 T = L i +λ T f i i+1 u i u i 0 T = u i H i H i = L i (x i,u i )+λ T i+1 f i(x i,u i ) i.e. the stationarity condition in the the Euler-Lagrange equation. 21/
22 Euler-Lagrange and Bellman On the optimal trajectory V i (x i ) = L i (x i,u i )+V i+1(f i (x i,u i )) or if we apply the chain rule or λ T i = V i(x i ) = x i L i x i + V i+1 x i+1 f i x i = L i f i x +λt i+1 x i λ T i = x i H i λ T N = x φ N(x N ) i.e. the costate equation. 22/
23 Pause 2/
24 Constrained Dynamic Programming Constraints: u i U i x i X i System dynamics: x i+1 = f i (x i,u i ) x 0 = x 0 Performance index N 1 J = φ N (x N )+ L i (x i,u i ) Feasible state set: i=0 D i = { x i X i u i U i : f i (x i,u i ) D i+1 } Feasible decision set: D N = X N U i (x i) = { u i U i : f i (x i,u i ) D i+1 } 24/
25 Constrained Dynamic Programming Theorem Bellman equation: [ V i (x i ) = min u i U i L i (x i,u i )+V i+1 (x i+1 ) ] V N (x N ) = φ N (x N ) Feasible state set: D i = { x i X i u i U i : f i (x i,u i ) D i+1 } D N = X N Feasible decision set: Ui (x i) = { u i U i : f i (x i,u i ) D i+1 } 25/
26 Stagecoach problem II K A 2 4 B C D E 1 4 F G H I 4 J It is easy to realize that X 4 = { J } X = { K,H,I } X 2 = { E,F,G } X 1 = { B,C,D } X 0 = { A } However, since there is no path from K to J D 4 = { J } D = { H,I } D 2 = { E,F,G } D 1 = { B,C,D } D 0 = { A } 2/
27 Optimal stepping (DD). Consider the system x i+1 = x i +u i x 0 = 2 the performance index N 1 J = x 2 N + i=0 and the constraints x 2 i +u2 i with N = 4 u i { 1, 0, 1} x i { 2, 1, 0, 1, 2} 2/
28 Firstly, we establish V 4 (x 4 ) as in the following table x 4 V The combination of x and u determines the following state x 4 and consequently the V 4 (x 4 ) contribution. x 4 u x V 4 (x 4 ) u x
29 The combination of x and u also determines the instantaneous loss L = x 2 +u2. L u x If we add up the instantaneous loss and V 4 we have a tableau in which we for each possible value of x can perform the minimization in V (x ) = min u [L (x,u )+V 4 (x +u )] and determine the optimal value for the decision and the Bellman function (as function of x ). L +V 4 u x V u , ,0 2-1 Knowing V (x ) we have one of the components for i = 2. In this manner we can iterate backwards and finds:
30 L 2 +V u 2 x V 2 u L 1 +V 2 u 1 x V 1 u L 0 +V 1 u 0 x V 0 u With x 0 = 2 we can trace forward and find the input sequence 1, 1, 0, 0 which give (an optimal) performance equal. Sensitivity. Steering vs. feedback.
31 Optimal stepping (DD) II - EPC Consider the system from previous example, but with the constraints that x 4 = 1. System x i+1 = x i +u i x 0 = 2 x 4 = 1 Performance index J = x x 2 i +u2 i i=0 Constraints: u i { 1, 0, 1} x i { 2, 1, 0, 1, 2} x 4 V L +V 4 u x V u /
32 L 2 +V u 2 x V 2 u L 1 +V 2 u 1 x V 1 u L 0 +V 1 u 0 x V 0 u With x 0 = 2 we can iterate forward and find the optimal input sequence 1, 1, 0, 1 which is connected to a performance index equal 9. 2/
33 Reading guidance DO: chapter, page - /
Static and Dynamic Optimization (42111)
Static and Dynamic Optimization (42111) Niels Kjølstad Poulsen Build. 303b, room 016 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark
More informationDynamic Optimization
Dynamic Optimization Optimal Control Niels Kjølstad Poulsen Department of Informatics and Matematical Modelleling The Technical University of Denmark Version: 15 Januar 212 (B5) 216-11-14 11.36 2 Preface
More informationHamilton-Jacobi-Bellman Equation Feb 25, 2008
Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS
More informationStatic and Dynamic Optimization (42111)
Static and Dynamic Optimization (42111) Niels Kjølstad Poulsen Build. 303b, room 016 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark
More informationStatic and Dynamic Optimization
Static and Dynamic Optimization Course Introduction Niels Kjølstad Poulsen Informatics and Mathematical Modelling build. 321, room 16 The Technical University of Denmark email: nkp@imm.dtu.dk phone: +45
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationOptimal Control. Lecture 3. Optimal Control of Discrete Time Dynamical Systems. John T. Wen. January 22, 2004
Optimal Control Lecture 3 Optimal Control of Discrete Time Dynamical Systems John T. Wen January, 004 Outline optimization of a general multi-stage discrete time dynamical systems special case: discrete
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationPontryagin s maximum principle
Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin
More informationLecture 9: Discrete-Time Linear Quadratic Regulator Finite-Horizon Case
Lecture 9: Discrete-Time Linear Quadratic Regulator Finite-Horizon Case Dr. Burak Demirel Faculty of Electrical Engineering and Information Technology, University of Paderborn December 15, 2015 2 Previous
More informationFormula Sheet for Optimal Control
Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming
More informationESC794: Special Topics: Model Predictive Control
ESC794: Special Topics: Model Predictive Control Nonlinear MPC Analysis : Part 1 Reference: Nonlinear Model Predictive Control (Ch.3), Grüne and Pannek Hanz Richter, Professor Mechanical Engineering Department
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More informationQuadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University
.. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if
More informationSuppose that we have a specific single stage dynamic system governed by the following equation:
Dynamic Optimisation Discrete Dynamic Systems A single stage example Suppose that we have a specific single stage dynamic system governed by the following equation: x 1 = ax 0 + bu 0, x 0 = x i (1) where
More informationLecture 18: Optimization Programming
Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming
More informationRegional Solution of Constrained LQ Optimal Control
Regional Solution of Constrained LQ Optimal Control José DeDoná September 2004 Outline 1 Recap on the Solution for N = 2 2 Regional Explicit Solution Comparison with the Maximal Output Admissible Set 3
More informationRobotics. Control Theory. Marc Toussaint U Stuttgart
Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,
More informationOptimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:
Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationCHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67
CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Chapter2 p. 1/67 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Main Purpose: Introduce the maximum principle as a necessary condition to be satisfied by any optimal
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationChapter 5. Pontryagin s Minimum Principle (Constrained OCP)
Chapter 5 Pontryagin s Minimum Principle (Constrained OCP) 1 Pontryagin s Minimum Principle Plant: (5-1) u () t U PI: (5-2) Boundary condition: The goal is to find Optimal Control. 2 Pontryagin s Minimum
More informationLecture 5 Linear Quadratic Stochastic Control
EE363 Winter 2008-09 Lecture 5 Linear Quadratic Stochastic Control linear-quadratic stochastic control problem solution via dynamic programming 5 1 Linear stochastic system linear dynamical system, over
More informationLecture 4: State Estimation in Hidden Markov Models (cont.)
EE378A Statistical Signal Processing Lecture 4-04/13/2017 Lecture 4: State Estimation in Hidden Markov Models (cont.) Lecturer: Tsachy Weissman Scribe: David Wugofski In this lecture we build on previous
More informationNumerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini
Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33
More informationStochastic and Adaptive Optimal Control
Stochastic and Adaptive Optimal Control Robert Stengel Optimal Control and Estimation, MAE 546 Princeton University, 2018! Nonlinear systems with random inputs and perfect measurements! Stochastic neighboring-optimal
More informationSteady State Kalman Filter
Steady State Kalman Filter Infinite Horizon LQ Control: ẋ = Ax + Bu R positive definite, Q = Q T 2Q 1 2. (A, B) stabilizable, (A, Q 1 2) detectable. Solve for the positive (semi-) definite P in the ARE:
More informationHomework Solution # 3
ECSE 644 Optimal Control Feb, 4 Due: Feb 17, 4 (Tuesday) Homework Solution # 3 1 (5%) Consider the discrete nonlinear control system in Homework # For the optimal control and trajectory that you have found
More informationGame Theory Extra Lecture 1 (BoB)
Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,
More informationEE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games
EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations
More informationNumerical Methods for Model Predictive Control. Jing Yang
Numerical Methods for Model Predictive Control Jing Yang Kongens Lyngby February 26, 2008 Technical University of Denmark Informatics and Mathematical Modelling Building 321, DK-2800 Kongens Lyngby, Denmark
More informationTrajectory-based optimization
Trajectory-based optimization Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 6 1 / 13 Using
More informationOPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationLinear Quadratic Regulator (LQR) I
Optimal Control, Guidance and Estimation Lecture Linear Quadratic Regulator (LQR) I Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore Generic Optimal Control Problem
More informationThe Implicit Function Theorem with Applications in Dynamics and Control
48th AIAA Aerospace Sciences Meeting Including the New Horizons Forum and Aerospace Exposition 4-7 January 2010, Orlando, Florida AIAA 2010-174 The Implicit Function Theorem with Applications in Dynamics
More informationModule 05 Introduction to Optimal Control
Module 05 Introduction to Optimal Control Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html October 8, 2015
More informationProblem 1 Cost of an Infinite Horizon LQR
THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 5 Ahmad F. Taha October 12, 215 Homework Instructions: 1. Type your solutions in the LATEX homework
More informationContinuous Time Finance
Continuous Time Finance Lisbon 2013 Tomas Björk Stockholm School of Economics Tomas Björk, 2013 Contents Stochastic Calculus (Ch 4-5). Black-Scholes (Ch 6-7. Completeness and hedging (Ch 8-9. The martingale
More informationDissipativity. Outline. Motivation. Dissipative Systems. M. Sami Fadali EBME Dept., UNR
Dissipativity M. Sami Fadali EBME Dept., UNR 1 Outline Differential storage functions. QSR Dissipativity. Algebraic conditions for dissipativity. Stability of dissipative systems. Feedback Interconnections
More informationChapter 2 Optimal Control Problem
Chapter 2 Optimal Control Problem Optimal control of any process can be achieved either in open or closed loop. In the following two chapters we concentrate mainly on the first class. The first chapter
More informationLecture 4 Continuous time linear quadratic regulator
EE363 Winter 2008-09 Lecture 4 Continuous time linear quadratic regulator continuous-time LQR problem dynamic programming solution Hamiltonian system and two point boundary value problem infinite horizon
More informationLinear Quadratic Regulator (LQR) Design II
Lecture 8 Linear Quadratic Regulator LQR) Design II Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore Outline Stability and Robustness properties
More informationDynamical Systems & Lyapunov Stability
Dynamical Systems & Lyapunov Stability Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University Outline Ordinary Differential Equations Existence & uniqueness Continuous dependence
More informationIntroduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability Goal of This
More informationOptimal Control Theory
Optimal Control Theory The theory Optimal control theory is a mature mathematical discipline which provides algorithms to solve various control problems The elaborate mathematical machinery behind optimal
More informationMinimization with Equality Constraints!
Path Constraints and Numerical Optimization Robert Stengel Optimal Control and Estimation, MAE 546, Princeton University, 2017 State and Control Equality Constraints Pseudoinverse State and Control Inequality
More informationMinimization with Equality Constraints
Path Constraints and Numerical Optimization Robert Stengel Optimal Control and Estimation, MAE 546, Princeton University, 2015 State and Control Equality Constraints Pseudoinverse State and Control Inequality
More informationAn Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game
An Application of Pontryagin s Maximum Principle in a Linear Quadratic Differential Game Marzieh Khakestari (Corresponding author) Institute For Mathematical Research, Universiti Putra Malaysia, 43400
More informationA Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley
A Tour of Reinforcement Learning The View from Continuous Control Benjamin Recht University of California, Berkeley trustable, scalable, predictable Control Theory! Reinforcement Learning is the study
More informationTo keep things simple, let us just work with one pattern. In that case the objective function is defined to be. E = 1 2 xk d 2 (1)
Backpropagation To keep things simple, let us just work with one pattern. In that case the objective function is defined to be E = 1 2 xk d 2 (1) where K is an index denoting the last layer in the network
More informationOptimization-Based Control. Richard M. Murray Control and Dynamical Systems California Institute of Technology
Optimization-Based Control Richard M. Murray Control and Dynamical Systems California Institute of Technology Version v2.1b (2 Oct 218) c California Institute of Technology All rights reserved. This manuscript
More informationSecond Order Optimality Conditions for Constrained Nonlinear Programming
Second Order Optimality Conditions for Constrained Nonlinear Programming Lecture 10, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk)
More informationConstrained Optimal Control I
Optimal Control, Guidance and Estimation Lecture 34 Constrained Optimal Control I Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore opics Motivation Brie Summary
More informationEE C128 / ME C134 Feedback Control Systems
EE C128 / ME C134 Feedback Control Systems Lecture Additional Material Introduction to Model Predictive Control Maximilian Balandat Department of Electrical Engineering & Computer Science University of
More informationMATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem
MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear
More informationAlberto Bressan. Department of Mathematics, Penn State University
Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-time Optimal Control Donald J. Chmielewski Chemical Engineering Department University of California Los Angeles February 23, 2000 Stochastic Formulation - Min Max Formulation - UCLA
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationConvex Optimization M2
Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization
More informationME 233, UC Berkeley, Spring Background Parseval s Theorem Frequency-shaped LQ cost function Transformation to a standard LQ
ME 233, UC Berkeley, Spring 214 Xu Chen Lecture 1: LQ with Frequency Shaped Cost Function FSLQ Background Parseval s Theorem Frequency-shaped LQ cost function Transformation to a standard LQ Big picture
More informationSolution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark
Solution Methods Richard Lusby Department of Management Engineering Technical University of Denmark Lecture Overview (jg Unconstrained Several Variables Quadratic Programming Separable Programming SUMT
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationECE7850 Lecture 8. Nonlinear Model Predictive Control: Theoretical Aspects
ECE7850 Lecture 8 Nonlinear Model Predictive Control: Theoretical Aspects Model Predictive control (MPC) is a powerful control design method for constrained dynamical systems. The basic principles and
More informationFeedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field
Vol. 4, No. 4, 23 Feedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field Ashraf H. Owis Department of Astronomy, Space and Meteorology, Faculty of Science, Cairo University Department
More informationAdvanced Mechatronics Engineering
Advanced Mechatronics Engineering German University in Cairo 21 December, 2013 Outline Necessary conditions for optimal input Example Linear regulator problem Example Necessary conditions for optimal input
More informationEE363 homework 2 solutions
EE363 Prof. S. Boyd EE363 homework 2 solutions. Derivative of matrix inverse. Suppose that X : R R n n, and that X(t is invertible. Show that ( d d dt X(t = X(t dt X(t X(t. Hint: differentiate X(tX(t =
More informationMathematical Economics. Lecture Notes (in extracts)
Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter
More informationLinear Quadratic Regulator (LQR) Design I
Lecture 7 Linear Quadratic Regulator LQR) Design I Dr. Radhakant Padhi Asst. Proessor Dept. o Aerospace Engineering Indian Institute o Science - Bangalore LQR Design: Problem Objective o drive the state
More information4F3 - Predictive Control
4F3 Predictive Control - Lecture 2 p 1/23 4F3 - Predictive Control Lecture 2 - Unconstrained Predictive Control Jan Maciejowski jmm@engcamacuk 4F3 Predictive Control - Lecture 2 p 2/23 References Predictive
More informationCHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73
CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS p. 1/73 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS Mixed Inequality Constraints: Inequality constraints involving control and possibly
More informationOptimal Control, TSRT08: Exercises. This version: October 2015
Optimal Control, TSRT8: Exercises This version: October 215 Preface Acknowledgment Abbreviations The exercises are modified problems from different sources: Exercises 2.1, 2.2,3.5,4.1,4.2,4.3,4.5, 5.1a)
More informationCE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review
CE 191: Civil & Environmental Engineering Systems Analysis LEC 17 : Final Review Professor Scott Moura Civil & Environmental Engineering University of California, Berkeley Fall 2014 Prof. Moura UC Berkeley
More informationLinear-Quadratic Optimal Control: Full-State Feedback
Chapter 4 Linear-Quadratic Optimal Control: Full-State Feedback 1 Linear quadratic optimization is a basic method for designing controllers for linear (and often nonlinear) dynamical systems and is actually
More informationLecture 6: Conic Optimization September 8
IE 598: Big Data Optimization Fall 2016 Lecture 6: Conic Optimization September 8 Lecturer: Niao He Scriber: Juan Xu Overview In this lecture, we finish up our previous discussion on optimality conditions
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-Time Nonlinear Quadratic Optimal Control V. Manousiouthakis D. Chmielewski Chemical Engineering Department UCLA 1998 AIChE Annual Meeting Outline Unconstrained Innite-Time Nonlinear
More informationTheoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems.
ES22 Lecture Notes #11 Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems. J = φ(x( ) + L(x,u,t)dt ; x= f(x,u,t)
More informationIEOR 265 Lecture 14 (Robust) Linear Tube MPC
IEOR 265 Lecture 14 (Robust) Linear Tube MPC 1 LTI System with Uncertainty Suppose we have an LTI system in discrete time with disturbance: x n+1 = Ax n + Bu n + d n, where d n W for a bounded polytope
More informationOptimal Control Theory - Module 3 - Maximum Principle
Optimal Control Theory - Module 3 - Maximum Principle Fall, 215 - University of Notre Dame 7.1 - Statement of Maximum Principle Consider the problem of minimizing J(u, t f ) = f t L(x, u)dt subject to
More informationOutline. 1 Linear Quadratic Problem. 2 Constraints. 3 Dynamic Programming Solution. 4 The Infinite Horizon LQ Problem.
Model Predictive Control Short Course Regulation James B. Rawlings Michael J. Risbeck Nishith R. Patel Department of Chemical and Biological Engineering Copyright c 217 by James B. Rawlings Outline 1 Linear
More informationLesson 27 Linear Programming; The Simplex Method
Lesson Linear Programming; The Simplex Method Math 0 April 9, 006 Setup A standard linear programming problem is to maximize the quantity c x + c x +... c n x n = c T x subject to constraints a x + a x
More informationDynamic Consistency for Stochastic Optimal Control Problems
Dynamic Consistency for Stochastic Optimal Control Problems Cadarache Summer School CEA/EDF/INRIA 2012 Pierre Carpentier Jean-Philippe Chancelier Michel De Lara SOWG June 2012 Lecture outline Introduction
More informationDELFT UNIVERSITY OF TECHNOLOGY Faculty of Electrical Engineering, Mathematics and Computer Science
DELFT UNIVERSITY OF TECHNOLOGY Faculty of Electrical Engineering, Matematics and Computer Science. ANSWERS OF THE TEST NUMERICAL METHODS FOR DIFFERENTIAL EQUATIONS (WI3097 TU) Tuesday January 9 008, 9:00-:00
More informationDeterministic Optimal Control
page A1 Online Appendix A Deterministic Optimal Control As far as the laws of mathematics refer to reality, they are not certain; and as far as they are certain, they do not refer to reality. Albert Einstein
More informationLinear and Loglinear Approximations (Started: July 7, 2005; Revised: February 6, 2009)
Dave Backus / NYU Linear and Loglinear Approximations (Started: July 7, 2005; Revised: February 6, 2009) Campbell s Inspecting the mechanism (JME, 1994) added some useful intuition to the stochastic growth
More informationOptimal Control with Learned Forward Models
Optimal Control with Learned Forward Models Pieter Abbeel UC Berkeley Jan Peters TU Darmstadt 1 Where we are? Reinforcement Learning Data = {(x i, u i, x i+1, r i )}} x u xx r u xx V (x) π (u x) Now V
More informationValue Function and Optimal Trajectories for some State Constrained Control Problems
Value Function and Optimal Trajectories for some State Constrained Control Problems Hasnaa Zidani ENSTA ParisTech, Univ. of Paris-saclay "Control of State Constrained Dynamical Systems" Università di Padova,
More informationOptimal Stopping Problems
2.997 Decision Making in Large Scale Systems March 3 MIT, Spring 2004 Handout #9 Lecture Note 5 Optimal Stopping Problems In the last lecture, we have analyzed the behavior of T D(λ) for approximating
More information6.231 DYNAMIC PROGRAMMING LECTURE 6 LECTURE OUTLINE
6.231 DYNAMIC PROGRAMMING LECTURE 6 LECTURE OUTLINE Review of Q-factors and Bellman equations for Q-factors VI and PI for Q-factors Q-learning - Combination of VI and sampling Q-learning and cost function
More informationMATH 4211/6211 Optimization Constrained Optimization
MATH 4211/6211 Optimization Constrained Optimization Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Constrained optimization
More informationDetermination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study
International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:
More informationNonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1
Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems p. 1/1 p. 2/1 Converse Lyapunov Theorem Exponential Stability Let x = 0 be an exponentially stable equilibrium
More informationMarkov Chain Monte Carlo Methods for Stochastic
Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013
More informationSolution of Stochastic Optimal Control Problems and Financial Applications
Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty
More informationSuboptimality of minmax MPC. Seungho Lee. ẋ(t) = f(x(t), u(t)), x(0) = x 0, t 0 (1)
Suboptimality of minmax MPC Seungho Lee In this paper, we consider particular case of Model Predictive Control (MPC) when the problem that needs to be solved in each sample time is the form of min max
More informationOutput Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x)
Output Feedback and State Feedback EL2620 Nonlinear Control Lecture 10 Exact feedback linearization Input-output linearization Lyapunov-based control design methods ẋ = f(x,u) y = h(x) Output feedback:
More informationOptimal Control. McGill COMP 765 Oct 3 rd, 2017
Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps
More informationStochastic optimal control theory
Stochastic optimal control theory Bert Kappen SNN Radboud University Nijmegen the Netherlands July 5, 2008 Bert Kappen Introduction Optimal control theory: Optimize sum of a path cost and end cost. Result
More informationConstrained Optimization
1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange
More information