Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
|
|
- Helen Bell
- 5 years ago
- Views:
Transcription
1 Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
2 Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability
3 Goal of This Presentation Reachability is nothing but an optimal control/differential game problem Everything in reachability ultimately amounts to solving a PDE. Any (small enough) optimal control problem (including reachability problems) can be solved using the Level Set Toolbox.
4 Optimal Control Optimize a cost function subject to system dynamics. Discrete-time systems: Continuous-time systems:
5 A Quick Example: Dubins Car Terminal/Final Time Running Cost Initial Time Terminal Cost Dynamics Initial State Quick question: what are we trying to do to the system here?
6 Solving Optimal Control Problems Approach1: Calculus of Variations (CoV) Takes an optimization perspective Uses Lagrange multipliers to eliminate constraints Derive first-order optimality conditions Globally optimal solution is not guaranteed
7 Solving Optimal Control Problems Approach2: Principle of Dynamic Programming Gives the globally optimal solution. Principle: The optimal state trajectory remains optimal at intermediate points in time.
8 Dynamic Programming Example* n Step T: V T (x = 4) = 0 m 1(5) 2(1) Step T-1: V T 1 (2) = w down (2) + V T (4) V T 1 (2) = = 1 3(4) 4(0) V T 1 (3) = w right (3) + V T (4) V T 1 (3) = = 4 Intersection w right w down Step T-2: V T 2 (1) = min [(w right (1) + V T 1 (2)), (w down (1) + V T 1 (3))] V T 2 (1) = min [(5 + 1), (1 + 4)] = *Shamelessly copied from Sylvia Herbert s presentation 8
9 Magic of Dynamic Programming Target set Compute the set of states that can reach the target set within 4s? Assume that we can compute the set of states that can reach any other given set of states within 1s. t = -1 t = -2 t = -3 t = -4
10 Principle of Dynamic Programming Recall our optimal control problem: + Minimize J x, t = C x t, u t dt, Subject to x = f x, u, t + l x T Define cost from t, x V x t, t = min C x t, u t dt, + l x T We are interested in finding cost from state x and time 0 V x 0, 0 = min C x t, u t dt : + l x T
11 Principle of Dynamic Programming Dynamic programming principle implies that:,<= V x t, t = min 6 V x t + δ, t + δ + 8 C x, u dt, 8 C x, u dt,<=,
12 Principle of Dynamic Programming,<= V x t, t = min 6 V x t + δ, t + δ + 8 C x, u dt, V x t, t + dv dt x t, t δ + V x t, t dx dt δ Taylor Expansion C x, u δ Approximation V x t, t = min 6, V x t, t + dv dt x t, t δ + V x t, t dx dt δ + C x, u δ
13 Principle of Dynamic Programming V x t, t = min 6, V x t, t + dv dt x t, t δ + V x t, t dx dt δ + C x, u δ V x t, t = V x t, t + min 6, dv dt x t, t δ + V x t, t dx dt δ + C x, u δ V x t, t = V x t, t + min 6, dv dt x t, t δ + V x t, t dx dt δ + C x, u δ 0 = min 6, dv dt x t, t δ + V x t, t dx dt δ + C x, u δ 0 = min 6, dv dt x t, t + V x t, t dx dt + C x, u
14 Principle of Dynamic Programming 0 = min 6, dv dt x t, t + V x t, t dx dt + C x, u dv dt + min 6 V x t, t f x, u + C x, u = 0 Hamilton-Jacobi Bellman PDE V x T, T = l x T Final Value of PDE
15 Hamilton-Jacobi Bellman(HJB) PDE Problem: + Minimize J x, t = C x t, u t dt + l x T, Subject to x = f x, u, t Solution A final-value PDE dv dt + min 6 V x t, t f x, u + C x, u = 0 V x T, T = l x T What is V x t, t? Where is my optimal control?
16 Curse of Dimensionality Wow! One single PDE for any optimal control problem. What is the problem here?
17 Level Set Toolbox Developed by Prof. Ian Mitchell during his PhD Can solve any initial-value PDE of the form: dv dt + H (x, V x t, t), t = 0 V x 0, 0 = l x 0 But optimal control problem is a final value PDE.L dv dt + min 6 V x t, t f x, u + C x, u = 0 V x T, T = l x T
18 Level Set Toolbox Good news: They are interchangeable! dv dt + H (x, V x t, t), t = 0 V x 0, 0 = l x 0 dw dt H (x, W x t, t), t = 0 W x T, T = l x T W x, T t = V(x, t) So we can solve any optimal control problem with the toolbox!
19 Optimal Control: Quick Recap Optimal control problem: Optimize a cost function subject to system dynamics. Two approaches: Calculus of Variations: Gives local solutions, but faster to compute. Dynamic Programming: Gives global solution A Final-value PDE needs to be solved May not even have a classical solution! Curse of dimensionality! Can be solved using Level Set Toolbox for low dimensional problems.
20
21 Reachability Problem: Find the set of all states that can reach a given set of states L within a time duration of T. Backward Reachable Tube R T Target Set (L) R T = {x : : u, s.t. x K satisfies x = f x, u, x 0 = x : ; t 0, T, s.t. x(t) L} Any thoughts?
22 Reachability Hint 1: Define a function l x such that, L = x: l x 0 Now consider the problem, V x t, t = min 6 l(x T ) Subject to x = f x, u, t What does V x t, t represent?
23 Reachability What are V x t, t for each of the following trajectories? x 1 (t) x 2 (t) x 1 (T) x 2 (T) l x = 1 l x = 0 l x = 1 x 3 (T) x 3 (t) V x t, t = min 6 l(x T )
24 Reachability So what does V x t, t represent? The value of l x that we will reach at time T How do I answer my original question? x 0 R T V x 0, 0 0 x 0 R T V x 0, 0 > 0
25 Reachability So reachability is nothing but an optimal control problem. min 6 l(x T ) Subject to x = f x, u, t L = x: l x 0 What is the corresponding PDE? Co-state Value Function dv dt + H (x, V x t, t), t = 0 V x T, T = l x T Hamiltonian H = min 6 V x t, t f x, u, t How to get "target-reaching control? u = argmin 6 V x t, t f x, u, t
26 But hold on What are V x t, t for each of the following trajectories? l x = 1 l x = 0 L x 2 (T) x 3 (T) x 2 (t) x 3 (t) V x t, t = min 6 l(x T ) What they should be? R T = {x : : u, s.t. x K satisfies x = f x, u, x 0 = x : ; t 0, T, s.t. x(t) L} x 0 R T V x 0, 0 0
27 What s going on? Need to account for the fact that trajectories can reach the target but then leave it. min ( min l(x t )) 6, :,+ Subject to x = f x, u, t L = x: l x 0 Does this fix the issue? L l x = 1 l x = 0 x 2 (T) x 3 (T) x 2 (t) x 3 (t) V x t, t = min 6 l(x T )
28 Freezing the Trajectories in the Target Need to account for the fact that trajectories can reach the target but then escape it. min ( min l(x t )) 6, :,+ Subject to x = f x, u, t L = x: l x 0 What is the corresponding PDE? dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = min 6 V x t, t f x, u, t
29 Reachability: Reachable Sets vs Tubes Backward Reachable Set (BRS): the set of all states that can reach a target set of states L exactly at time T. R T = {x : : u, s.t. x K satisfies x = f x, u, x 0 = x : ; x(t) L} min 6 l(x T ) Subject to x = f x, u, t L = x: l x 0 dv dt + H (x, V x t, t), t = 0 V x T, T = l x T H = min 6 V x t, t f x, u, t Backward Reachable Tube (BRT): the set of all states that can reach a target set of states L within a duration of time T. R T = {x : : u, s.t. x K satisfies x = f x, u, x 0 = x : ; t 0, T, s.t. x(t) L} min ( min l(x t )) 6, :,+ Subject to x = f x, u, t L = x: l x 0 dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = min 6 V x t, t f x, u, t
30 Backward Reachable Tube: Example Target set Backward reachable tubes t = -1 t = -2 t = -3 t = -4
31 Backward Reachable Set: Example Target set t = -1 Backward reachable sets t = -2 t = -3 t = -4
32 Computing Backward Reachable Tube 1. Define target set L for the system to reach within a given time horizon: R T 2. Define implicit level set function for final time: l z, L = z: l z 0 3. Find an appropriate value function V z(t), t dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = min 6 V x t, t f x, u, t 4. Retrieve zero sub-level of level set function at initial time
33 Reachability: Key Takeaways Reachability is just an optimal control problem. Backward Reachable Set (BRS) vs Backward Reachable Tube (BRT) Both can be computed using the Level Set Toolbox. Suffers from the curse of dimensionality
34
35 Introducing the Disturbance Suppose our dynamics were: x = f x, u, d u control, d disturbance + And cost were: J x, t = C x t, u t, d t dt, + l x T Now, we want to solve the following differential game: V x t, t = min 6 max a + 8 C x t, u t, d t dt, + l x T A similar PDE can be derived in this case (called HJI PDE)
36 Optimal Control vs Differential Game Static Evolving Over Time One agent (input) Optimization Optimal Control Multiple agents (input and disturbance) Game Theory Differential Games
37 Hamilton-Jacobi Isaacs(HJI) PDE Problem: + Minimize J x, t = C x t, u t, d(t) dt + l x T, Subject to x = f x, u, d, t Solution A final-value PDE dv dt + min max 6 a V x t, t f x, u, d + C x, u, d = 0 V x T, T = l x T What is V x t, t? Where is my optimal control?
38 Reachability With Disturbance Problem: Find the set of all states that can reach a given set of states L despite the disturbance within a time duration of T. Backward Reachable Tube R T Target Set (L) R T = x : : u, s.t. d, x K satisfies x = f x, u, d, x 0 = x : ; t 0, T, s.t. x t L What does this definition mean? How to solve this problem?
39 Reachability With Disturbance We can again formulate it as a differential game min max min l(x t ) 6 a, :,+ Subject to x = f x, u, d, t L = x: l x 0 What is the corresponding PDE? dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = min 6 max a V x t, t f x, u, d, t How to get "target-reaching control? u = argmin 6 max a V x t, t f x, u, d, t
40 Reachability With Disturbance: Key Takeaways Very similar to classic reachability; just an extra max Backward Reachable Set (BRS) vs Backward Reachable Tube (BRT) Both can be computed using the Level Set Toolbox. Suffers from the curse of dimensionality
41 Reachability With Disturbance: Quick Trivia Problem: Find the set of all states that can reach a given set of states L for some disturbance within a time duration of T. R T = x : : u, d, s.t. x K satisfies x = f x, u, d, x 0 = x : ; t 0, T, s.t. x t L What does this definition mean? How to solve this problem?
42 Reachability With Disturbance: Trivia Solution Disturbance is like control here min min min l(x t ) 6 a, :,+ Subject to x = f x, u, d, t L = x: l x 0 What is the corresponding PDE? dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = min min 6 a V x t, t f x, u, d, t
43 Shades of Reachability: Avoid Set Problem: Find the set of all states that can avoid a given set of states L despite the disturbance for a time duration of T. Backward Reachable Tube R T Avoid Set (L) R T = x : : d, s.t. u, x K satisfies x = f x, u, d, x 0 = x : ; t 0, T, s.t. x t L What does this definition mean? How to solve this problem?
44 Reachability: Avoid Set Simply exchange the role of input and disturbance max min min l(x t ) 6 a, :,+ Subject to x = f x, u, d, t L = x: l x 0 What is the corresponding PDE? dv dt + min {0, H (x, V x t, t), t } = 0 V x T, T = l x T H = max min 6 a V x t, t f x, u, d, t How to get "target-avoiding control? u = argmax 6 min a V x t, t f x, u, d, t
45 Avoid Set: Example Target set What do the green sets represent here? Backward reachable tubes t = -1 t = -2 t = -3 t = -4
46 Shades of Reachability: Forward Reachable Set Problem: Find the set of all states that I can reach from a given set of states L at time T. Forward Reachable Set R T Initial Set (L) F T = x + : u, s.t. x K satisfies x = f x, u, x 0 L; x T = x + What does this definition mean? How to solve this problem?
47 Forward Reachable Set Define a function l x such that, L = x: l x 0 Now consider the problem, V x t, t = max l(x T ) 6 Subject to x = f x, u, t Will this work?
48 Forward Reachable Set Again it is an optimal control problem. max l(x T ) 6 Subject to x = f x, u, t L = x: l x 0 What is the corresponding PDE? An Initial Value PDE dv dt + H (x, V x t, t), t = 0 V x 0, 0 = l x 0 H = max 6 V x t, t f x, u, t
49 Forward Reachable Set Trivia Problem: Find the set of all states that I can reach from a given set of states L despite the disturbance at time T. F T = x + : u, s.t. d, x K satisfies x = f x, u, d, x 0 L; x T = x + Solve the following PDE: dv dt + H (x, V x t, t), t = 0 V x 0, 0 = l x 0 H = max min 6 a V x t, t f x, u, t
50 Forward Reachable Set: Key Takeaways Gives an initial value PDE Forward Reachable Set (FRS) vs Forward Reachable Tube (FRT) Both can be computed using the Level Set Toolbox. Suffers from the curse of dimensionality
51 Shades of Reachability: Obstacles Problem: Find the set of all states that can reach a given set of states L without hitting the obstacle (G) despite the disturbance within a time duration of T. Backward Reachable Tube R T Target Set (L) Obstacle (G) R T = x :: u, s.t. d, x K satisfies x = f x, u, d, x 0 = x : ; t 0, T, x t G and t 0, T, s.t. x t L What does this definition mean?
52 Reachability With Obstacles Again can be formulated as a differential game Value function can be shown to satisfy the following equation: max { ae a, + min {0, H x, V x t, t), t, g x V x, t } = 0 V x T, T = max l x, T, g x, T H = min 6 max a V x t, t f x, u, d, t How to get "target-reaching control without hitting the obstacle? u = argmin 6 max a V x t, t f x, u, d, t
53 Reachability With Obstacles: Example Target set Fixed obstacle
54 Reachability With Obstacles: Example Reachable Tube t = -1 t = -2 t = -3 t = -4
55 Shades of Reachability: Key Takeaways Everything in reachability ultimately amounts to solving a PDE. Different min-max combinations appear in the PDE based on what control and disturbance are trying to do. Min with zero appears in the PDE depending on whether a set or a tube is being computed. Initial or final value PDE is solved based on whether a FRS or a BRS is being computed Obstacles can be considered Any combination above can be computed using the Level Set Toolbox.
56 Reachability: Final Remarks Reachability theory has much more rigorous mathematical foundation. Time-varying targets/obstacles can also be treated very easily.
57 Thank You!
Models for Control and Verification
Outline Models for Control and Verification Ian Mitchell Department of Computer Science The University of British Columbia Classes of models Well-posed models Difference Equations Nonlinear Ordinary Differential
More informationOptimal Control Theory
Optimal Control Theory The theory Optimal control theory is a mature mathematical discipline which provides algorithms to solve various control problems The elaborate mathematical machinery behind optimal
More informationHamilton-Jacobi-Bellman Equation Feb 25, 2008
Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationOn continuous time contract theory
Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem
More informationReach Sets and the Hamilton-Jacobi Equation
Reach Sets and the Hamilton-Jacobi Equation Ian Mitchell Department of Computer Science The University of British Columbia Joint work with Alex Bayen, Meeko Oishi & Claire Tomlin (Stanford) research supported
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationNumerical Optimization of Partial Differential Equations
Numerical Optimization of Partial Differential Equations Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada URL: http://www.math.mcmaster.ca/bprotas Rencontres
More informationControl of Dynamical System
Control of Dynamical System Yuan Gao Applied Mathematics University of Washington yuangao@uw.edu Spring 2015 1 / 21 Simplified Model for a Car To start, let s consider the simple example of controlling
More informationStatic and Dynamic Optimization (42111)
Static and Dynamic Optimization (421) Niels Kjølstad Poulsen Build. 0b, room 01 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark Email:
More informationOptimal Control. McGill COMP 765 Oct 3 rd, 2017
Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps
More informationOverapproximating Reachable Sets by Hamilton-Jacobi Projections
Overapproximating Reachable Sets by Hamilton-Jacobi Projections Ian M. Mitchell and Claire J. Tomlin September 23, 2002 Submitted to the Journal of Scientific Computing, Special Issue dedicated to Stanley
More informationOverapproximating Reachable Sets by Hamilton-Jacobi Projections
Journal of Scientific Computing, Vol. 19, Nos. 1 3, December 2003 ( 2003) Overapproximating Reachable Sets by Hamilton-Jacobi Projections Ian M. Mitchell 1 and Claire J. Tomlin 2 Received May 3, 2002;
More informationHJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011
Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance
More informationpacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre
HJ pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre Control of State Constrained Dynamical Systems 25-29 September 2017, Padova Main
More informationEE C128 / ME C134 Feedback Control Systems
EE C128 / ME C134 Feedback Control Systems Lecture Additional Material Introduction to Model Predictive Control Maximilian Balandat Department of Electrical Engineering & Computer Science University of
More informationOptimal Sojourn Time Control within an Interval 1
Optimal Sojourn Time Control within an Interval Jianghai Hu and Shankar Sastry Department of Electrical Engineering and Computer Sciences University of California at Berkeley Berkeley, CA 97-77 {jianghai,sastry}@eecs.berkeley.edu
More informationLecture 3: Hamilton-Jacobi-Bellman Equations. Distributional Macroeconomics. Benjamin Moll. Part II of ECON Harvard University, Spring
Lecture 3: Hamilton-Jacobi-Bellman Equations Distributional Macroeconomics Part II of ECON 2149 Benjamin Moll Harvard University, Spring 2018 1 Outline 1. Hamilton-Jacobi-Bellman equations in deterministic
More informationEE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games
EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations
More informationCHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67
CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Chapter2 p. 1/67 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Main Purpose: Introduce the maximum principle as a necessary condition to be satisfied by any optimal
More informationEllipsoidal Toolbox. TCC Workshop. Alex A. Kurzhanskiy and Pravin Varaiya (UC Berkeley)
Ellipsoidal Toolbox TCC Workshop Alex A. Kurzhanskiy and Pravin Varaiya (UC Berkeley) March 27, 2006 Outline Problem setting and basic definitions Overview of existing methods and tools Ellipsoidal approach
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationGame Theory Extra Lecture 1 (BoB)
Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationMean field games and related models
Mean field games and related models Fabio Camilli SBAI-Dipartimento di Scienze di Base e Applicate per l Ingegneria Facoltà di Ingegneria Civile ed Industriale Email: Camilli@dmmm.uniroma1.it Web page:
More informationAlberto Bressan. Department of Mathematics, Penn State University
Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i
More informationarxiv: v1 [cs.ma] 25 Nov 2016
Robust Sequential Path Planning Under Disturbances and Adversarial Intruder Mo Chen, Somil Bansal, Jaime F. Fisac, Claire J. Tomlin arxiv:6.8364v [cs.ma] 25 Nov 26 Abstract Provably safe and scalable multi-vehicle
More information25. Chain Rule. Now, f is a function of t only. Expand by multiplication:
25. Chain Rule The Chain Rule is present in all differentiation. If z = f(x, y) represents a two-variable function, then it is plausible to consider the cases when x and y may be functions of other variable(s).
More informationNonlinear representation, backward SDEs, and application to the Principal-Agent problem
Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem
More informationb) The system of ODE s d x = v(x) in U. (2) dt
How to solve linear and quasilinear first order partial differential equations This text contains a sketch about how to solve linear and quasilinear first order PDEs and should prepare you for the general
More informationNew Identities for Weak KAM Theory
New Identities for Weak KAM Theory Lawrence C. Evans Department of Mathematics University of California, Berkeley Abstract This paper records for the Hamiltonian H = p + W (x) some old and new identities
More informationMATH 312 Section 2.4: Exact Differential Equations
MATH 312 Section 2.4: Exact Differential Equations Prof. Jonathan Duncan Walla Walla College Spring Quarter, 2007 Outline 1 Exact Differential Equations 2 Solving an Exact DE 3 Making a DE Exact 4 Conclusion
More informationOptimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations
Martino Bardi Italo Capuzzo-Dolcetta Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Birkhauser Boston Basel Berlin Contents Preface Basic notations xi xv Chapter I. Outline
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationModeling and Analysis of Dynamic Systems
Modeling and Analysis of Dynamic Systems by Dr. Guillaume Ducard Fall 2016 Institute for Dynamic Systems and Control ETH Zurich, Switzerland based on script from: Prof. Dr. Lino Guzzella 1/33 Outline 1
More informationFourier Transform for Continuous Functions
Fourier Transform for Continuous Functions Central goal: representing a signal by a set of orthogonal bases that are corresponding to frequencies or spectrum. Fourier series allows to find the spectrum
More informationFINITE DIFFERENCES. Lecture 1: (a) Operators (b) Forward Differences and their calculations. (c) Backward Differences and their calculations.
FINITE DIFFERENCES Lecture 1: (a) Operators (b) Forward Differences and their calculations. (c) Backward Differences and their calculations. 1. Introduction When a function is known explicitly, it is easy
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationMulti-agent avoidance control using an M-matrix property
Electronic Journal of Linear Algebra Volume 12 Volume 12 (2004) Article 7 2004 Multi-agent avoidance control using an M-matrix property Dusan M. Stipanovic dusan@uiuc.edu Sriram Shankaran Claire J. Tomlin
More informationStability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games
Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationMA 201: Partial Differential Equations Lecture - 2
MA 201: Partial Differential Equations Lecture - 2 Linear First-Order PDEs For a PDE f(x,y,z,p,q) = 0, a solution of the type F(x,y,z,a,b) = 0 (1) which contains two arbitrary constants a and b is said
More informationDIFFERENTIAL EQUATIONS
DIFFERENTIAL EQUATIONS Basic Concepts Paul Dawkins Table of Contents Preface... Basic Concepts... 1 Introduction... 1 Definitions... Direction Fields... 8 Final Thoughts...19 007 Paul Dawkins i http://tutorial.math.lamar.edu/terms.aspx
More informationA Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes (bilmes@cs.berkeley.edu) International Computer Science Institute
More informationOptimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:
Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic
More informationSemi-decidable Synthesis for Triangular Hybrid Systems
Semi-decidable Synthesis for Triangular Hybrid Systems Omid Shakernia 1, George J. Pappas 2, and Shankar Sastry 1 1 Department of EECS, University of California at Berkeley, Berkeley, CA 94704 {omids,sastry}@eecs.berkeley.edu
More informationON CALCULATING THE VALUE OF A DIFFERENTIAL GAME IN THE CLASS OF COUNTER STRATEGIES 1,2
URAL MATHEMATICAL JOURNAL, Vol. 2, No. 1, 2016 ON CALCULATING THE VALUE OF A DIFFERENTIAL GAME IN THE CLASS OF COUNTER STRATEGIES 1,2 Mikhail I. Gomoyunov Krasovskii Institute of Mathematics and Mechanics,
More informationA Very General Hamilton Jacobi Theorem
University of Salerno 9th AIMS ICDSDEA Orlando (Florida) USA, July 1 5, 2012 Introduction: Generalized Hamilton-Jacobi Theorem Hamilton-Jacobi (HJ) Theorem: Consider a Hamiltonian system with Hamiltonian
More informationChapter 5. Pontryagin s Minimum Principle (Constrained OCP)
Chapter 5 Pontryagin s Minimum Principle (Constrained OCP) 1 Pontryagin s Minimum Principle Plant: (5-1) u () t U PI: (5-2) Boundary condition: The goal is to find Optimal Control. 2 Pontryagin s Minimum
More informationGame Theoretic Continuous Time Differential Dynamic Programming
Game heoretic Continuous ime Differential Dynamic Programming Wei Sun, Evangelos A. heodorou and Panagiotis siotras 3 Abstract In this work, we derive a Game heoretic Differential Dynamic Programming G-DDP
More informationHamilton-Jacobi formulation for Reach-Avoid Problems with an application to Air Traffic Management
2010 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 30-July 02, 2010 ThA20.2 Hamilton-Jacobi formulation for Reach-Avoid Problems with an application to Air Traffic Management
More informationReal Time Stochastic Control and Decision Making: From theory to algorithms and applications
Real Time Stochastic Control and Decision Making: From theory to algorithms and applications Evangelos A. Theodorou Autonomous Control and Decision Systems Lab Challenges in control Uncertainty Stochastic
More informationOptimal Control with Learned Forward Models
Optimal Control with Learned Forward Models Pieter Abbeel UC Berkeley Jan Peters TU Darmstadt 1 Where we are? Reinforcement Learning Data = {(x i, u i, x i+1, r i )}} x u xx r u xx V (x) π (u x) Now V
More informationDuality and dynamics in Hamilton-Jacobi theory for fully convex problems of control
Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory
More informationWaves in a Shock Tube
Waves in a Shock Tube Ivan Christov c February 5, 005 Abstract. This paper discusses linear-wave solutions and simple-wave solutions to the Navier Stokes equations for an inviscid and compressible fluid
More informationSection 1.8/1.9. Linear Transformations
Section 1.8/1.9 Linear Transformations Motivation Let A be a matrix, and consider the matrix equation b = Ax. If we vary x, we can think of this as a function of x. Many functions in real life the linear
More informationComputation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305
To appear in Dynamics of Continuous, Discrete and Impulsive Systems http:monotone.uwaterloo.ca/ journal Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationOPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationStochastic Modelling in Climate Science
Stochastic Modelling in Climate Science David Kelly Mathematics Department UNC Chapel Hill dtbkelly@gmail.com November 16, 2013 David Kelly (UNC) Stochastic Climate November 16, 2013 1 / 36 Why use stochastic
More informationBackward Stochastic Differential Equations with Infinite Time Horizon
Backward Stochastic Differential Equations with Infinite Time Horizon Holger Metzler PhD advisor: Prof. G. Tessitore Università di Milano-Bicocca Spring School Stochastic Control in Finance Roscoff, March
More informationNumerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini
Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33
More informationSafety Preserving Control Synthesis for Sampled Data Systems
Safety Preserving Control Synthesis for Sampled Data Systems Ian M. Mitchell Department of Computer Science, University of British Columbia Shahab Kaynama, Mo Chen Department of Electrical Engineering
More informationECE7850 Lecture 8. Nonlinear Model Predictive Control: Theoretical Aspects
ECE7850 Lecture 8 Nonlinear Model Predictive Control: Theoretical Aspects Model Predictive control (MPC) is a powerful control design method for constrained dynamical systems. The basic principles and
More informationMaximum Process Problems in Optimal Control Theory
J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard
More informationIntroduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim
Introduction - Motivation Many phenomena (physical, chemical, biological, etc.) are model by differential equations. Recall the definition of the derivative of f(x) f f(x + h) f(x) (x) = lim. h 0 h Its
More informationMatrix-Exponentials. September 7, dx dt = ax. x(t) = e at x(0)
Matrix-Exponentials September 7, 207 In [4]: using PyPlot INFO: Recompiling stale cache file /Users/stevenj/.julia/lib/v0.5/LaTeXStrings.ji for module LaTeXString Review: Solving ODEs via eigenvectors
More informationTHE NATIONAL UNIVERSITY OF IRELAND, CORK COLÁISTE NA hollscoile, CORCAIGH UNIVERSITY COLLEGE, CORK. Summer Examination 2009.
OLLSCOIL NA héireann, CORCAIGH THE NATIONAL UNIVERSITY OF IRELAND, CORK COLÁISTE NA hollscoile, CORCAIGH UNIVERSITY COLLEGE, CORK Summer Examination 2009 First Engineering MA008 Calculus and Linear Algebra
More informationPhysics 351 Friday, April 24, 2015
Physics 351 Friday, April 24, 2015 HW13 median report time = 5 hours. You ve solved 145 homework problems this term (not counting XC). Whew! This weekend, you ll read Feynman s two lectures (Feynman Lectures
More informationIntroduction to Mathematical Modeling
Introduction to Mathematical Modeling - Systems Theory in the Toolbox for Systems Biology The 5 th International Course in Yeast Systems Biology 2011 June 6, 2011, PhD, Assoc Prof Head of Department Systems
More informationThe Linear Quadratic Regulator
10 The Linear Qadratic Reglator 10.1 Problem formlation This chapter concerns optimal control of dynamical systems. Most of this development concerns linear models with a particlarly simple notion of optimality.
More informationMultibody simulation
Multibody simulation Dynamics of a multibody system (Newton-Euler formulation) Dimitar Dimitrov Örebro University June 8, 2012 Main points covered Newton-Euler formulation forward dynamics inverse dynamics
More informationCalculus I Review Solutions
Calculus I Review Solutions. Compare and contrast the three Value Theorems of the course. When you would typically use each. The three value theorems are the Intermediate, Mean and Extreme value theorems.
More informationArtificial Intelligence Bayes Nets: Independence
Artificial Intelligence Bayes Nets: Independence Instructors: David Suter and Qince Li Course Delivered @ Harbin Institute of Technology [Many slides adapted from those created by Dan Klein and Pieter
More informationTHE INVERSE FUNCTION THEOREM
THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)
More informationMon Jan Improved acceleration models: linear and quadratic drag forces. Announcements: Warm-up Exercise:
Math 2250-004 Week 4 notes We will not necessarily finish the material from a given day's notes on that day. We may also add or subtract some material as the week progresses, but these notes represent
More informationA Parcimonious Long Term Mean Field Model for Mining Industries
A Parcimonious Long Term Mean Field Model for Mining Industries Yves Achdou Laboratoire J-L. Lions, Université Paris Diderot (joint work with P-N. Giraud, J-M. Lasry and P-L. Lions ) Mining Industries
More information3 Applications of partial differentiation
Advanced Calculus Chapter 3 Applications of partial differentiation 37 3 Applications of partial differentiation 3.1 Stationary points Higher derivatives Let U R 2 and f : U R. The partial derivatives
More informationQuadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University
.. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if
More informationLagrangian for Central Potentials
Physics 411 Lecture 2 Lagrangian for Central Potentials Lecture 2 Physics 411 Classical Mechanics II August 29th 2007 Here we will review the Lagrange formulation in preparation for the study of the central
More informationTaylor Series and stationary points
Chapter 5 Taylor Series and stationary points 5.1 Taylor Series The surface z = f(x, y) and its derivatives can give a series approximation for f(x, y) about some point (x 0, y 0 ) as illustrated in Figure
More informationCSC321 Lecture 6: Backpropagation
CSC321 Lecture 6: Backpropagation Roger Grosse Roger Grosse CSC321 Lecture 6: Backpropagation 1 / 21 Overview We ve seen that multilayer neural networks are powerful. But how can we actually learn them?
More informationCompletion Date: Monday February 11, 2008
MATH 4 (R) Winter 8 Intermediate Calculus I Solutions to Problem Set #4 Completion Date: Monday February, 8 Department of Mathematical and Statistical Sciences University of Alberta Question. [Sec..9,
More informationLecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.
Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference
More informationBayes Nets: Independence
Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes
More informationCLEP Precalculus - Problem Drill 15: Systems of Equations and Inequalities
CLEP Precalculus - Problem Drill 15: Systems of Equations and Inequalities No. 1 of 10 1. What are the methods to solve a system of equations? (A) Graphing, replacing, substitution and matrix techniques.
More informationExperiment 2. F r e e F a l l
Suggested Reading for this Lab Experiment F r e e F a l l Taylor, Section.6, and standard deviation rule in Taylor handout. Review Chapters 3 & 4, Read Sections 8.1-8.6. You will also need some procedures
More information1.2 Derivation. d p f = d p f(x(p)) = x fd p x (= f x x p ). (1) Second, g x x p + g p = 0. d p f = f x g 1. The expression f x gx
PDE-constrained optimization and the adjoint method Andrew M. Bradley November 16, 21 PDE-constrained optimization and the adjoint method for solving these and related problems appear in a wide range of
More informationComputational Fluid Dynamics Prof. Dr. SumanChakraborty Department of Mechanical Engineering Indian Institute of Technology, Kharagpur
Computational Fluid Dynamics Prof. Dr. SumanChakraborty Department of Mechanical Engineering Indian Institute of Technology, Kharagpur Lecture No. #11 Fundamentals of Discretization: Finite Difference
More informationOptimization using Calculus. Optimization of Functions of Multiple Variables subject to Equality Constraints
Optimization using Calculus Optimization of Functions of Multiple Variables subject to Equality Constraints 1 Objectives Optimization of functions of multiple variables subjected to equality constraints
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationOrdinary Differential Equations
Ordinary Differential Equations In this lecture, we will look at different options for coding simple differential equations. Start by considering bicycle riding as an example. Why does a bicycle move forward?
More informationPhysics Fall Mechanics, Thermodynamics, Waves, Fluids. Recap: Position and displacement
Physics 5 Fall 28 Mechanics, Thermodynamics, Waves, Fluids Lecture 3: motion in a straight line II Slide 3- Recap: Position and displacement In one dimension, position can be described by a positive or
More informationEuler s Method applied to the control of switched systems
Euler s Method applied to the control of switched systems FORMATS 2017 - Berlin Laurent Fribourg 1 September 6, 2017 1 LSV - CNRS & ENS Cachan L. Fribourg Euler s method and switched systems September
More informationLecture : The Definite Integral & Fundamental Theorem of Calculus MTH 124. We begin with a theorem which is of fundamental importance.
We begin with a theorem which is of fundamental importance. The Fundamental Theorem of Calculus (FTC) If F (t) is continuous for a t b, then b a F (t) dt = F (b) F (a). Moreover the antiderivative F is
More informationMulti-Factor Finite Differences
February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit
More information2.20 Fall 2018 Math Review
2.20 Fall 2018 Math Review September 10, 2018 These notes are to help you through the math used in this class. This is just a refresher, so if you never learned one of these topics you should look more
More informationProduct rule. Chain rule
Probability Recap CS 188: Artificial Intelligence ayes Nets: Independence Conditional probability Product rule Chain rule, independent if and only if: and are conditionally independent given if and only
More informationStochastic Calculus Made Easy
Stochastic Calculus Made Easy Most of us know how standard Calculus works. We know how to differentiate, how to integrate etc. But stochastic calculus is a totally different beast to tackle; we are trying
More informationMATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem
MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear
More information