Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems.

Similar documents
Suppose that we have a specific single stage dynamic system governed by the following equation:

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Homework Solution # 3

Linear Quadratic Regulator (LQR) Design II

Mathematical Methods - Lecture 9

Deterministic Dynamic Programming

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

2.3 Calculus of variations

Modeling and Analysis of Dynamic Systems

b) The system of ODE s d x = v(x) in U. (2) dt

UNIVERSITY OF MANITOBA

Partial Differential Equations Separation of Variables. 1 Partial Differential Equations and Operators

z x = f x (x, y, a, b), z y = f y (x, y, a, b). F(x, y, z, z x, z y ) = 0. This is a PDE for the unknown function of two independent variables.

Learning From Data Lecture 25 The Kernel Trick

MA 201: Method of Separation of Variables Finite Vibrating String Problem Lecture - 11 MA201(2016): PDE

Optimization of Linear Systems of Constrained Configuration

Math 211. Lecture #6. Linear Equations. September 9, 2002

Partial Differential Equations, Winter 2015

MATH 2250 Final Exam Solutions

Problem 1 Cost of an Infinite Horizon LQR

EE C128 / ME C134 Final Exam Fall 2014

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

Applied Math Qualifying Exam 11 October Instructions: Work 2 out of 3 problems in each of the 3 parts for a total of 6 problems.

Lecture 4 Continuous time linear quadratic regulator

The integrating factor method (Sect. 1.1)

Linear-Quadratic Optimal Control: Full-State Feedback

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Robotics. Control Theory. Marc Toussaint U Stuttgart

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

SWEEP METHOD IN ANALYSIS OPTIMAL CONTROL FOR RENDEZ-VOUS PROBLEMS

Lecture 6. Foundations of LMIs in System and Control Theory

EE C128 / ME C134 Feedback Control Systems

An Introduction to Numerical Methods for Differential Equations. Janet Peterson

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

Linear Systems Theory

Pontryagin s maximum principle

Math 342 Partial Differential Equations «Viktor Grigoryan

Output Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x)

MATH 220: PROBLEM SET 1, SOLUTIONS DUE FRIDAY, OCTOBER 2, 2015

The second-order 1D wave equation

The Relativistic Heat Equation

AM 205: lecture 14. Last time: Boundary value problems Today: Numerical solution of PDEs

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

MA 201, Mathematics III, July-November 2016, Partial Differential Equations: 1D wave equation (contd.) and 1D heat conduction equation

Formula Sheet for Optimal Control

2 The Linear Quadratic Regulator (LQR)

CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

MATH 425, HOMEWORK 5, SOLUTIONS

Switching, sparse and averaged control

4F3 - Predictive Control

Linear Differential Equations. Problems

Class Meeting # 1: Introduction to PDEs

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

Stochastic optimal control theory

Lecture 2: Discrete-time Linear Quadratic Optimal Control

Linear Quadratic Regulator (LQR) II

Alexei F. Cheviakov. University of Saskatchewan, Saskatoon, Canada. INPL seminar June 09, 2011

16. Working with the Langevin and Fokker-Planck equations

ECE504: Lecture 8. D. Richard Brown III. Worcester Polytechnic Institute. 28-Oct-2008

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Steady State Kalman Filter

Resolvent Estimates and Quantification of Nonlinear Stability

Stochastic and Adaptive Optimal Control

A differential Lyapunov framework for contraction analysis

Prof. Krstic Nonlinear Systems MAE281A Homework set 1 Linearization & phase portrait

Static and Dynamic Optimization (42111)

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim

Principles of Optimal Control Spring 2008

UNIVERSITY of LIMERICK OLLSCOIL LUIMNIGH

Week 4 Lectures, Math 6451, Tanveer

MATH 425, HOMEWORK 3 SOLUTIONS

Analysis III (BAUG) Assignment 3 Prof. Dr. Alessandro Sisto Due 13th October 2017

Chapter 8 Stabilization: State Feedback 8. Introduction: Stabilization One reason feedback control systems are designed is to stabilize systems that m

CDS 110b: Lecture 2-1 Linear Quadratic Regulators

Pilot Waves and the wave function

Introduction to Seismology

Advanced Control Theory

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA

Continuous Time Finance

MATH 220: MIDTERM OCTOBER 29, 2015

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

10 Transfer Matrix Models

The Necessity of the Transversality Condition at Infinity: A (Very) Special Case

ACM/CMS 107 Linear Analysis & Applications Fall 2016 Assignment 4: Linear ODEs and Control Theory Due: 5th December 2016

Lecture 5: Linear Systems. Transfer functions. Frequency Domain Analysis. Basic Control Design.

Math 201 Assignment #11

Ordinary differential equations. Phys 420/580 Lecture 8

Solving First Order PDEs

Production and Relative Consumption

Lyapunov Stability Analysis: Open Loop

Dr. Allen Back. Sep. 10, 2014

Numerics and Control of PDEs Lecture 7. IFCAM IISc Bangalore. Feedback stabilization of a 1D nonlinear model

Computing High Frequency Waves By the Level Set Method

Linear algebra and differential equations (Math 54): Lecture 19

Review: control, feedback, etc. Today s topic: state-space models of systems; linearization

Transcription:

ES22 Lecture Notes #11 Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems. J = φ(x( ) + L(x,u,t)dt ; x= f(x,u,t) ; x( t )= x ; ψ(x( ))= (P) t Suppose all necessary condition for (P) have been satisfied. What about sufficiency condition. Recall for static optimization problem with equality constraints, the trick is to expand the augmented criterion to second order and the constraints to first order, i.e.? J = φ(x( ) + λ T (f- x) +L(x,u,t)dt =( φ x -λ T ) δx( )+ [(λ+h x )δx+ H u t? δj = [φ x -λ T ] δx( )+ [(λ T +H x )δx(t)+ H u δu(t)]dt + 1 2 (δxt φ xx δx) + [δx δu] T t to H xx H xu δx H ux H uu δu dt and δx= f x δx + f u δu=- H λ ; δx( t )= δx = ; λ T =- H x, λ T ( )= φ x( ) ; H u = for all t choosing λ T =- H x, λ T ( )= φ x( ) ; H u = for all t we get, δx= f x δx + f u δu=- H λ ; δx( t )= δx = ; and δj = 1 2 (δxt φ xx δx) + t [δx δu] T H xx H xu δx H ux H uu δu dt which is recognized as a LQ problem. Thus if we can show that the minimum of this accessory problem, which is δx(t o ) T S(t o )δx(t o ), is zero or S(t )> then the stationary solution must be a local minimum also. Consequently, the sufficient condition is simply the existence of a positive definite solution of the Riccati equation (also know as the Jacobi condition or the conjugate point condition. See optics example). Practical Rationale for LQ problems Aerospace guidance and control Chemical process control YCHo 11/2/98 1

All kinds of automative applications: cruise control, engine control, temperature control Economic growth and resource models for nation, industry, and firms communication networks, computer systems manufacturing plants stock market seismic data processing image analysis weather prediction Infinite time regulator problem the case of in the LQ problem with A, B, Q, R, constant. Intuition ds/dt= since the Q term (decreased S) and the SBR - 1 B T S term (increases S) fight to a stand still. If S( -)=constant than u = - R -1 B T Sx = Kx ==> x = (A +BK)x is a constant coefficient linear system. Question: is it stable? Or does optimality imply stability? Answer: Stability the optimal return function x T Sx V(x) is a Lyapunov function. We have dv/dt = x T Sdx/dt + x T ds/dtx + dx/dt T Sx = x T [(A+BK) T S -SA-A T S-Q+SBR -1 B T S + (A+BK)S]x = x T [-Q-SBR -1 B T S]x < ==> stability Convergence If the system is controllable than we know a finite time control sequence followed by zero u(t) will stabilize the system at zero. The optimal control sequence must have a smaller value for J ==> convergence since nonzero integral of J is constantly increasing. Inhomogeneous LQ problem: J = 1 2 (( x- x f) T S f (x- x f )) + [x- x u- u] T t Q N T N R x- x u- u dt x= Ax + Bu + f(t) ; x( t )= x where x, u, f are given functions of time and x f is given constant. The optimal V(x,t) is (1/2)x T Sx +α T x +β where S still obeys the same Riccati equation and α and β obey linear ODEs dependent on S(t). It is also possible to add linear terms in x and u to the terminal and in-flight portion of the cirterion without changing the general nature of the solution. Two special cases: (i) Terminal Constraints: J = 1 2 u T R u dt ; x = A(t)x +B(t)u ; x( )= Consider t YCHo 11/2/98 2

J = ν T x( ) + 1 2 u T R u dt t and the HJBPDE, we get after assuming V(x,t)=(1/2)x T Sx +α T x +β, - 1 2 xt Sx - α T x - β = 1 2 xt [SA + A T S - SB R -1 B T S]x + α T Ax - 1 2 αt BR -1 B T α - α T BR -1 B T Sx Collecting and equating terms in x T x, x, and scalars, we get -S = SA + A T S - SB R -1 B T S ; S( ) = - α = ( A T - SB R -1 B T )α ; a T ( ) = ν T - β = - 1 2 αt BR -1 B T α ; β( ) = which implies S(t) = for all t and dα/dt = -A T α and α(t) = Φ(, t) ν β(t) = ν T { Φ(, τ)b R -1 B T Φ T (, τ)dτ}ν t We can solve for ν assuming controllability via x( ) = = - Φ(, t o )x +{ Φ(, τ)b R -1 B T Φ T (, τ)dτ}ν t NOTE: This solution should be taken with a grain of salt. Since ν depends on x and α(t) depends on ν. Thus we don't really have a feedback solution in the strict sense. (ii) The Least square fit Problem (was a take home quiz problem in 199) Let the scalar time function z(t) sin(t) for t [,]. Let x(t) a+bt. Determine a and b such that J = 1 2 (z-x ) 2 dt is minimized. (i) (ii) (iii) (iv) (2%) convert this problem to an nonstandard inhomogeneous Linear-Quadratic problem in which the optimizing variables are the constants a and b. (I have lectured in class extensively on how this portion of the problem can be solved and have warned everybody that this part of my lecture is important and subject to quiz). (35%) solve this problem from first principles (i.e., derive the appropriate conditions for optimum ; don't just assert them) using the LaGrangian methods (35%) solve the same problem again from first principle using dynamic programming and show that you can get the same answer. (1%) argue from intuitive grounds that the optimal b must be equal to zero. YCHo 11/2/98 3

(v) explain in what sense this problem captures the essential elements of the course YCHo 11/2/98 4

Solution: Define x(t)=x 1 (t), dx 1 /dt=x 2, dx 2 /dt= (1) Then we can let x 1 ()=a, x 2 ()=b. (2) (i) (ii) The inhomogeneous LQ problem is to choose a and b to minimize J subject to (1) and (2). Adjoin (1) to J via multiplier functions λ 1 (t) and λ 2 (t), we get J= 1 2 (z- x 1 ) 2 dt + [λ 1 (x 2 -x 1 )+ λ 2 (-x 2 )]dt integrating by parts J= [- λ 1 x 1 -λ 2 x 2 ] + [ 1 2 (z- x 1 )2 + λ 1 x 2 +λ 1 x 1 +λ 2 x 2 ]dt Taking variations on J, we get δj = [- λ 1 δx 1 -λ 2 δx 2 ] + [-(z- x1 )δx 1 +( λ 1 +λ 2 )δx 2 +λ 1 δx 1 ]dt Let us choose for convenience λ 1 = z- x 1 ; λ 1 ()= (3) then, λ 2 = - λ 1 ; λ 2 ()= (4) δj = λ 1 () δx 1 ()+ λ 2 () δx 2 ()= λ 1 () δa+ λ 2 () δb (5) For optimum a and b, we must have in addition, λ 1 ()= λ 2 ()= (6) Thus, the necessary conditions are (4),(6), and x 1 (t)=a+bt, i.e., (1). Integrating (4) using (6), we get λ 1 ()= zdt - a - 1 2 b2 t λ 2 ()= - zdτdt- 1 2 a2-1 6 b3 o The optimum a and b are now obtained by solving YCHo 11/2/98 5

- - 1 2 2-1 2 2-1 6 3 a b = - o t zdt zd τdt For z(t) =sin(t), we obtain from (7 ), a=2/ and b=. (7) (iii) For the dynamic programming solution, define V(x 1,x 2,t) as the value of J when starting at t with x 1 (t) and x 2 (t), then the usual DP argument yields V(x 1,x 2,t) = V( x 1 +² x 1, x 2 +² x 2,t+²t) + 1 2 (z- x 1 )2 ²t ==> - V t = V x 1 x 2 +V x 2 * + 1 2 (z(t)- x 1 )2 (8) Try a solution of the PDE (8) with V(x 1,x 2,t)= (1/2)S 11 (t)x 2 1 + S 12 (t)x 1 x 2 + (1/2)S 22 (t)x 2 2 + α l (t)x 1 +α 2 (t)x 2 +β(t). Substituting into (8) we get, - [ 1 2 S 11 x 1 2 +S 12 x 1 x 2 + 1 2 S 22 x 2 2 +α 1 x 1 +α 2 x 2 +β]=[s 11 x 1 + S 12 x 2 +α 1 ]x 2 + 1 2 (z- x 1 )2 Collecting terms in x 1 2, x 2 2, x 1, x 2 and equating their coefficients, we get S 11 = -1 ; S 12 = - S 11 ; S 22 = - 2S 12 with S 11 ()=, S 12 ()=, S 22 ()= - α 1 = - z ; - α 2 = α 1 with α 2 ()=, α 1 ()= β= - 1 2 z2 ; β()= (9) Integrating (9), S 11 (t)=-t, S 12 (t)= 2 /2 + (1/2)t 2 - t, S 22 (t)=(1/3) 3-2 t-(1/3)t 3 +t 2, α 1 (t) = -1-cos(t), α 2 (t)=-+t+sint. Now for optimum a and b, we differentiate V(x 1 (),x 2 (),) w.r.t. a and b. we have V/b=,V/a= ==> S 11 () S 12 () a -α = 1 () S 12 () S 22 () b -α 2 () (1) Solving (1), we obtain once again b= a=2/. (iv) symmetry dictates that b=. YCHo 11/2/98 6