Pontryagin s Minimum Principle 1

Similar documents
Deterministic Dynamic Programming

Chapter 5. Pontryagin s Minimum Principle (Constrained OCP)

EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games

Constrained Optimal Control. Constrained Optimal Control II

Pontryagin s maximum principle

Principles of Optimal Control Spring 2008

Homework Solution # 3

Functional differentiation

CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73

Principles of Optimal Control Spring 2008

arxiv: v1 [q-fin.mf] 5 Jul 2016

Formula Sheet for Optimal Control

OPTIMAL CONTROL CHAPTER INTRODUCTION

Optimization via the Hamilton-Jacobi-Bellman Method: Theory and Applications

ECE 680 Fall Test #2 Solutions. 1. Use Dynamic Programming to find u(0) and u(1) that minimize. J = (x(2) 1) u 2 (k) x(k + 1) = bu(k),

Principles of Optimal Control Spring 2008

CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

Optimization-Based Control. Richard M. Murray Control and Dynamical Systems California Institute of Technology

The Rocket Car. UBC Math 403 Lecture Notes by Philip D. Loewen

14 F Time Optimal Control

Time-optimal control of a 3-level quantum system and its generalization to an n-level system

Minimum Fuel Optimal Control Example For A Scalar System

Principles of Optimal Control Spring 2008

Optimal Control and Applications

Optimal Control Theory - Module 3 - Maximum Principle

Chapter 2 Optimal Control Problem

Spacecraft Trajectory Optimization. Edited. Bruce A. Conway. Published by Cambridge University Press 2010

Feedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field

7.3 State Space Averaging!

A Tutorial on RIOTS 95 A Matlab Toolbox for Solving General Optimal Control Problems

7 OPTIMAL CONTROL 7.1 EXERCISE 1. Solve the following optimal control problem. max. (x u 2 )dt x = u x(0) = Solution with the first variation

Optimal Control. with. Aerospace Applications. James M. Longuski. Jose J. Guzman. John E. Prussing

Ostrogradsky theorem (from 1850)

Advanced Mechatronics Engineering

Dynamics and Control of Rotorcraft

Numerical Synthesis of Pontryagin Optimal Control Minimizers Using Sampling-Based Methods

Control, Stabilization and Numerics for Partial Differential Equations

A Study of Variable Thrust, Variable I sp Trajectories for Solar System Exploration. Tadashi Sakai

Topic # Feedback Control Systems

Solution of Stochastic Optimal Control Problems and Financial Applications

Robotics. Islam S. M. Khalil. November 15, German University in Cairo

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Theory and Applications of Constrained Optimal Control Proble

Time-optimal control of flexible systems subject to friction

Geometric Optimal Control with Applications

Legendre Transforms, Calculus of Varations, and Mechanics Principles

Numerical Methods for Embedded Optimization and Optimal Control. Exercises

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers

Optimal Control of Spacecraft Orbital Maneuvers by the Hamilton-Jacobi Theory

Classical Mechanics in Hamiltonian Form

Optimal Control. Historical Overview and Theoretical Development. Nicolai Christov

Problem 1 Cost of an Infinite Horizon LQR

Static and Dynamic Optimization (42111)

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

Lecture 11 Optimal Control

ECE 602 Exam 2 Solutions, 3/23/2011.

Necessary Optimality Conditions for Problem With Integral Boundary and Distributed Conditions

2.3 Calculus of variations

Objective. 1 Specification/modeling of the controlled system. 2 Specification of a performance criterion

Energy-based Swing-up of the Acrobot and Time-optimal Motion

Differential Games, Distributed Systems, and Impulse Control

Game Theory Extra Lecture 1 (BoB)

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle

9 Controller Discretization

Deterministic Optimal Control

Sketchy Notes on Lagrangian and Hamiltonian Mechanics

Asteroid Rendezvous Missions

Continuation from a flat to a round Earth model in the coplanar orbit transfer problem

Suppose that we have a specific single stage dynamic system governed by the following equation:

ON COLISSIONS IN NONHOLONOMIC SYSTEMS

Generalized Coordinates, Lagrangians

Real-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter

Physics 115/242 Comparison of methods for integrating the simple harmonic oscillator.

LECTURE NOTES IN CALCULUS OF VARIATIONS AND OPTIMAL CONTROL MSc in Systems and Control. Dr George Halikias

Topic # Feedback Control

José C. Geromel. Australian National University Canberra, December 7-8, 2017

2 Lyapunov Stability. x(0) x 0 < δ x(t) x 0 < ɛ

arxiv: v1 [cs.sy] 3 Sep 2015

Optimal Control Strategies in a Two Dimensional Differential Game Using Linear Equation under a Perturbed System

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Numerical Methods for Constrained Optimal Control Problems

ISTS-2017-d-114 ISSFD

Logistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations

QUALIFYING EXAMINATION, Part 1. Solutions. Problem 1: Mathematical Methods. x 2. 1 (1 + 2x 2 /3!) ] x 2 1 2x 2 /3

221A Lecture Notes Path Integral

LECTURES ON OPTIMAL CONTROL THEORY

MCE693/793: Analysis and Control of Nonlinear Systems

Numerical Optimal Control Overview. Moritz Diehl

Deterministic Optimal Control

Tutorial on Control and State Constrained Optimal Control Pro. Control Problems and Applications Part 3 : Pure State Constraints

Two-Point Boundary Value Problem and Optimal Feedback Control based on Differential Algebra

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri

Static and Dynamic Optimization (42111)

Statistical Mechanics Solution Set #1 Instructor: Rigoberto Hernandez MoSE 2100L, , (Dated: September 4, 2014)

Time-Optimal Output Transition for Minimum Phase Systems

Minimizing Mobility and Communication Energy in Robotic Networks: an Optimal Control Approach

Discrete abstractions of hybrid systems for verification

Numerical method for approximating the solution of an IVP. Euler Algorithm (the simplest approximation method)

Transcription:

ECE 680 Fall 2013 Pontryagin s Minimum Principle 1 In this handout, we provide a derivation of the minimum principle of Pontryagin, which is a generalization of the Euler-Lagrange equations that also includes problems with constraints on the control inputs. Only a special case of the minimum principle is stated. However, this special case covers a large class of control problems. We consider the problem of minimizing the performance index given by J = Φ(x(t f )) + F(x(t),u(t))dt (1) subject to ẋ = f(x,u), x( ) = x 0, and x(t f ) = x f. (2) We consider two cases: fixed final state and free final state. To proceed we define the Hamiltonian function H(x,u,p) as H(x,u,p) = H = F + p f, where the costate vector p will be determined in the analysis to follow. We discuss the case when the final time t f is fixed. We follow the development of Luenberger [1, Section 11.1]. We adjoin to J additional terms that sum up to zero. We note that because the state trajectories must satisfy the equation, ẋ = f(x,u), we have f(x,u) ẋ = 0. (3) We introduce the modified objective performance index, J = J + 1 handout prepared by S. H. Żak p(t) (f(x,u) ẋ)dt. (4) 1

Because for any state trajectory, (3) is satisfied, for any choice of p(t) the value of J is the same as that of J. Therefore, we can express J as J = Φ(x(t f )) + = Φ(x(t f )) + F(x(t),u(t))dt + p(t) (f(x,u) ẋ)dt (H(x,u,p) p ẋ)dt. (5) Let u(t) be a nominal control strategy; it determines a corresponding state trajectory x(t). If we apply another control strategy, say v(t), that is close to u(t), then v(t) will produce a state trajectory close to the nominal trajectory. This new state trajectory is just a perturbed version of x(t) and it can be represented as x(t) + δx(t). The change in the state trajectory yields a corresponding change in the modified performance index. We represent this change as δ J; it has the form, δ J = Φ(x(t f ) + δx(t f )) Φ(x(t f )) + We evaluate the integral above. Integrating by parts gives p δẋdt = p(t f ) δx(t f ) p( ) δx( ) (H(x + δx,v,p) H(x,u,p) p δẋ)dt. (6) ṗ δxdt. (7) Note that δx( ) = 0 because a change in the control strategy does not change the initial state. Taking into account the above, we represent (6) as δ J = Φ(x(t f ) + δx(t f )) Φ(x(t f )) p(t f ) δx(t f ) + (H(x + δx,v,p) H(x,u,p) + ṗ δx)dt. We next replace (Φ(x(t f ) + δx(t f )) Φ(x(t f ))) with its first-order approximation, and add and subtract the term H(x,v,p) under the integral to obtain δ J = ( xφ t=tf p(t f )) δx(tf ) + (H(x + δx,v,p) H(x,v,p) + H(x,v,p) H(x,u,p) + ṗ δx)dt. 2

Replacing (H(x + δx, v, p) H(x, v, p)) with its first-order approximation, H(x + δx,v,p) H(x,v,p) = H x δx, gives δ J = ( xφ t=tf p(t f )) δx(tf ) (( ) ) H + x + ṗ δx + H(x,v,p) H(x,u,p) dt. Selecting in the above p as the solution to the differential equation H x + ṗ = 0 with the final condition p(t f ) = xφ t=tf, reduces δ J to δ J = (H(x,v,p) H(x,u,p)) dt. If the original control u(t) is optimal, then we should have δ J 0, that is, H(x,v,p) H(x,u,p). We summarize our development in the following theorem. Theorem 1 Necessary conditions for u U to minimize (1) subject to (2) are: ṗ = where H = H(x,u,p) = F(x,u) + p f(x,u), ( ) H, x H(x,u,p ) = min u U H(x,u,p). If the final state, x(t f ), is free, then in addition to the above conditions it is required that the following end-point condition is satisfied, p(t f ) = xφ t=tf. 3

The equation, ṗ = is called in the literature the adjoint or costate equation. ( ) H (8) x We illustrate the above theorem with the following well-known example. Example 1 We consider a controlled object modeled by ẋ1 = 0 1 x 1 + 0 0 0 1 ẋ 2 = Ax + bu x 2 u with the performance index The control is required to satisfy J = 0 dt. u(t) 1 for all t [0,t f ]. This constraint means that the control must have magnitude no greater than 1. Our objective is to find admissible control that minimizes J and transfers the system from a given initial state x 0 to the origin. We begin by finding the Hamiltonian function for the problem, H = 1 + p (Ax + bu) [ ] = 1 + p 1 p 2 0 1 0 0 = 1 + p 1 x 2 + p 2 u. x 1 x 2 + 0 1 u The costate equations are ṗ1 ṗ 2 = H x 1 H x 2 = 0 p 1. Solving the costate equations yields p 1 = d 1 and p 2 = d 1 t + d 2, 4

where d 1 and d 2 are integration constants. We now find an admissible control minimizing the Hamiltonian, arg u min H = arg u min(1 + p 1 x 2 + p 2 u) Hence, = arg u min(p 2 u) u(t) = 1 if p 2 < 0 = u(t) = 0 if p 2 = 0 u(t) = 1 if p 2 > 0. u (t) = sign(p 2) = sign( d 1 t + d 2 ), where the signum function is defined as sign(z) = 1 if z > 0 0 if z = 0 1 if z < 0. Thus, the optimal control law is piecewise constant taking the values 1 or 1. This control law has at most two intervals of constancy because the argument of the sign function is a linear function, d 1 t + d 2, that changes its sign at most once. This type of control is called a bang-bang control because it switches back and forth between its extreme values. System trajectories for u = 1 and u = 1 are families of parabolas. Segments of the two parabolas through the origin form the switching curve, x 1 = 1 2 x2 2 sign(x 2 ). This means that if an initial state is above the switching curve, then u = 1 is used until the switching curve is reached. Then, u = 1 is used to reach the origin. For an initial state below the switching curve, the control u = 1 is used first to reach the switching curve, and then the control is switched to u = 1. We implement the above control action as 1 if v < 0 u = sign(v) = 0 if v = 0 1 if v > 0, 5

Figure 1: A phase-plane portrait of the time optimal closed-loop system of Example 1. where v = v(x 1,x 2 ) = x 1 + 1 2 x2 2sign(x 2 ) = 0 is the equation describing the switching curve. We can use the above equation to synthesize a closed-loop system such that starting at an arbitrary initial state in the state plane, the trajectory will always be moving in an optimal fashion towards the origin. Once the origin is reached, the trajectory will stay there. A phase portrait of the closed-loop time optimal system is given in Figure 1. References [1] D. G. Luenberger. Introduction to Dynamic Systems: Theory, Models, and Applications. John Wiley & Sons, New York, 1979. 6