UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

Similar documents
UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE)

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems

Feedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field

Prashant Mhaskar, Nael H. El-Farra & Panagiotis D. Christofides. Department of Chemical Engineering University of California, Los Angeles

ESC794: Special Topics: Model Predictive Control

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

EE C128 / ME C134 Feedback Control Systems

Converse Lyapunov Functions for Inclusions 2 Basic denitions Given a set A, A stands for the closure of A, A stands for the interior set of A, coa sta

Optimal Control Using an Algebraic Method for Control-Affine Nonlinear Systems. Abstract

Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 2015

POSITIVE CONTROLS. P.O. Box 513. Fax: , Regulator problem."

7.1 Linear Systems Stability Consider the Continuous-Time (CT) Linear Time-Invariant (LTI) system

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Deterministic Dynamic Programming

Problem Description The problem we consider is stabilization of a single-input multiple-state system with simultaneous magnitude and rate saturations,

Techinical Proofs for Nonlinear Learning using Local Coordinate Coding

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Robotics. Control Theory. Marc Toussaint U Stuttgart

Locally optimal controllers and application to orbital transfer (long version)

Time-Invariant Linear Quadratic Regulators!

Markov processes Course note 2. Martingale problems, recurrence properties of discrete time chains.

Global Linearization, and nally Model Predictive (Receding Horizon) Control (MPC), including hybrid approaches such as Model Predictive Control combin

A GENERALIZATION OF SONTAG S FORMULA FOR HIGH-PERFORMANCE CLF-BASED CONTROL. J. Willard Curtis III. A dissertation submitted to the faculty of

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

Stochastic Nonlinear Stabilization Part II: Inverse Optimality Hua Deng and Miroslav Krstic Department of Mechanical Engineering h

Game Theory Extra Lecture 1 (BoB)

Abstract. Previous characterizations of iss-stability are shown to generalize without change to the

4F3 - Predictive Control

LMI Methods in Optimal and Robust Control

Enlarged terminal sets guaranteeing stability of receding horizon control

Homework Solution # 3

[A + 1 ] + (1 ) v: : (b) Show: the derivative of T at v = v 0 < 0 is: = (v 0 ) (1 ) ; [A + 1 ]

Steady State Kalman Filter

minimize x subject to (x 2)(x 4) u,

WHEN deciding where spacecraft should land a balance exists between engineering safety and scientific (or programmatic)

Lecture 5 Linear Quadratic Stochastic Control

LINEAR-CONVEX CONTROL AND DUALITY

and the nite horizon cost index with the nite terminal weighting matrix F > : N?1 X J(z r ; u; w) = [z(n)? z r (N)] T F [z(n)? z r (N)] + t= [kz? z r

EML5311 Lyapunov Stability & Robust Control Design

Advanced Mechatronics Engineering

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko

Georgia Institute of Technology Nonlinear Controls Theory Primer ME 6402

Robust Control 5 Nominal Controller Design Continued

ON THE ASYMPTOTIC STABILITY IN TERMS OF TWO MEASURES FOR FUNCTIONAL DIFFERENTIAL EQUATIONS. G. Makay

Neighboring feasible trajectories in infinite dimension

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

Chapter 13 Internal (Lyapunov) Stability 13.1 Introduction We have already seen some examples of both stable and unstable systems. The objective of th

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

MARKOV CHAINS: STATIONARY DISTRIBUTIONS AND FUNCTIONS ON STATE SPACES. Contents

Applications of Nonlinear Control Using. the State-Dependent Riccati Equation THESIS. David K. Parrish. Captain, USAF

The Important State Coordinates of a Nonlinear System

Robust SDC Parameterization for a Class of Extended Linearization Systems

Benchmark problems in stability and design of. switched systems. Daniel Liberzon and A. Stephen Morse. Department of Electrical Engineering

Lecture 18: Optimization Programming

Streamlining of the state-dependent Riccati equation controller algorithm for an embedded implementation

6. Linear Quadratic Regulator Control

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

Stochastic and Adaptive Optimal Control

Lecture 4 Continuous time linear quadratic regulator

Chap. 3. Controlled Systems, Controllability

Computational Issues in Nonlinear Dynamics and Control

MCE693/793: Analysis and Control of Nonlinear Systems

Optimal control of nonlinear systems with input constraints using linear time varying approximations

Pointwise convergence rate for nonlinear conservation. Eitan Tadmor and Tao Tang

Design of Non-Linear Controller for a flexible rotatory beam using State-Dependent Riccati Equation (SDRE) control

Optimal nonlinear control for time delay system. Delsys 2013

ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko

Numerical Optimization

A Proof of the EOQ Formula Using Quasi-Variational. Inequalities. March 19, Abstract

Decentralized control with input saturation

Nonlinear Control. Nonlinear Control Lecture # 8 Time Varying and Perturbed Systems

Mathematical Economics. Lecture Notes (in extracts)

Denis ARZELIER arzelier

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini

Math 273a: Optimization Basic concepts

Numerical Optimal Control Overview. Moritz Diehl

A Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

Optimization: Interior-Point Methods and. January,1995 USA. and Cooperative Research Centre for Robust and Adaptive Systems.

On the Stability of the Best Reply Map for Noncooperative Differential Games

Introduction to Optimization!

Lecture Note 7: Switching Stabilization via Control-Lyapunov Function

Eects of small delays on stability of singularly perturbed systems

Suppose that we have a specific single stage dynamic system governed by the following equation:

The servo problem for piecewise linear systems

Outline. 1 Linear Quadratic Problem. 2 Constraints. 3 Dynamic Programming Solution. 4 The Infinite Horizon LQ Problem.

Linear Systems. Manfred Morari Melanie Zeilinger. Institut für Automatik, ETH Zürich Institute for Dynamic Systems and Control, ETH Zürich

Optimality Conditions for Constrained Optimization

Controllability Issues innonlinear State-Dependent. Riccati Equation Control

Lyapunov Stability Theory

3 Stability and Lyapunov Functions

Transcription:

Constrained Innite-Time Nonlinear Quadratic Optimal Control V. Manousiouthakis D. Chmielewski Chemical Engineering Department UCLA 1998 AIChE Annual Meeting

Outline Unconstrained Innite-Time Nonlinear Quadratic-Optimal Control Hamilton-Jacobi-Bellman Equation State Dependent Riccati Equation and the Curl Condition Closed-Loop Stability Example Constrained Innite-Time Nonlinear Quadratic-Optimal Control Family of Finite Time Problems Equivalence to the Innite Time Problem Example

Nonlinear Quadratic Optimal Control Innite-time Problem @J @x Z 1 min u + T Q(x)x ; 1 @J f(x) x @x 4 x T Q(x)x + u T R(x)u dt ;1 (x)b T (x) @J B(x)R = @x optimal is given by u (x) = ; 1 2 R;1 (x)b T (x) @J @x policy The T s.t. _x = f(x) + B(x)u The solution can be found by solving the HJB equation T T The determination of J(x) is nearly impossible in all but the simplest cases.

@J = 2 @x (x) 1 @P (x)^=p xt + @x 2 Innite-time Nonlinear Quadratic Optimal Control J(x) can be written as: J(x) = x T P (x)x where P (x) is symmetric. 1 In this case: @P + 1 B @P xt (x) @x 2 C x A Dene Then the HJB becomes: x T (x)a(x) + A T (x)(x) + Q(x) ; (x)b(x)r ;1 (x)b T (x)(x) x = where A(x) s.t. f(x) = A(x)x.

(H1) R(x) >, 8 x 2 < n, and Q(x) >, 8 x 2 D < n The gradients @(A(x)x) @x (H4) @b i(x) @x and Assumptions (H2) A(x) and B(x) are analytic matrix valued functions 8 x 2 D (H3) The pair (A(x) B(x)) is controllable (in the linear system sense) 8 x 2 D exist and are continuous and bounded 8 x 2 D (where b i (x) is the ith column of B(x))

State Dependent Riccati Equation If (x) satises the SDRE: (x)a(x) + A T (x) ; (x)b(x)r ;1 (x)b T (x)(x) + Q(x) = then the HJB will appear to be satised. In this case, the optimal policy is u (x) = ;R ;1 (x)b T (x)(x)x and the optimal value function is J(x) = x T P (x)x

@J @x Gradient of a Scalar Function For the HJB to actually be satised 2x T (x) must equal The conditions for a vector function, v(x), to be a gradient are : i @v = @v j j @x @x i 8 i j Assumption: (H5) A(x) is s.t. the solution to the SDRE, (x), satises: curl(x T (x)) = 8 x 2 D

Existence of a (x) there exists a solution to the HJB equation, If then there exists A(x) s.t. the solution J(x), @x = 2xT (x): The following guarantees the existence of a SDRE such that the curl condition is satised (Lu and Huang, 199). to the SDRE equation, (x), satises @J

= x T Z 1 (x)d! x J(x) Stability and Optimality and Chmielewski, 1998) Theorem:(Manousiouthakis (H1)-(H5) hold. Then the scalar function Let the necessary conditions for optimality (HJB) for all x 2 D. satises the feedback policy Furthermore, u(x) = ;R ;1 (x)b T (x)(x)x is asymptotically stabilizing. Let (H1)-(H5) hold, and assume D = < n. Corollary: the feedback policy is globally asymptotically stabilizing. Then

An Inverse Method Consider (x) and A(x) to be design functions. First we must guarantee (x) > curl(x T (x)) = A(x)x = f(x) Then simply calculate Q(x) from the SDRE. Q(x) = (x)b(x)r ;1 (x)b T (x)(x) ; A T (x)(x) ; (x)a(x)

> 8 x 2 Z An Inverse Method (continued) Let Z be dened as (a positively invariant set) Z = ( " n jj(x) J(x) = x T Z ) 1 (x)d# x x < 2 Dene s.t. x T Q(x)x + u T R(x)u = x T Q(x) + (x)b(x)r ;1 (x)b T (x)(x) x Then a sucient condition closed-loop stability is that x() 2 Z.

_x 1 ;2k a x 2 1 ; 4k a C Ass + F V x1 + 2k b x 2 x 1 = C A ; C Ass x 2 = C B ; C Bss u = C Ain ; C Ainss k a Example a k *) Consider a CSTR Reactor with reaction 2A k b B. 2 3 3 2 3 2 F V 4 5 = 4 5 4 u + 5 _x 2 a 2 + 2k ac Ass x 1 ; F V b x2 + k 1 x k F k 2 CAss = 1 2 (C Ainss ; C Ass ) b + V C Bss =

Example (continued) The following SS design parameters are used (all in SI units) k a = :5 k b = :1 F = :2 V C Ainss = 1 C Ass = 1:59 C Bss = 4:21

; 4k a C Ass + F V 4 2k a C Ass 2k b and Q o = 5 4 o 11 q q o 22 5 Example (continued) Let (x) = P o where P o satises A T o P o + P o A T o + Q o ; P o BR ;1 B T P o = where R = (V=F) 2, 2 3 2 3 F V + k b ; A o = Clearly this choice of (x) satises the \Curl" condition.

o 11 + 2k ax 1 (2p 11 ; p 12 ) + p 2 11 q k a x 1 (p 22 ; 2p 12 ) q o 22 q o Example (continued) Q(x) is calculated, via the SDRE, to be: 2 3 q o 11 + 2k ax 1 (2p 11 ; p 12 ) k a x 1 (p 22 ; 2p 12 ) Q(x) = 4 5 Then Q(x) + P o BR ;1 B T P o is positive denite i q o 11 + 2k ax 1 (2p 11 ; p 12 ) + p 2 11 > 22 ; (k a x 1 (p 22 ; 2p 12 )p 11 p 12 ) 2 >

Z = 2 < 2 j T P o Example (continued) For q o 11 = 1, and qo 22 = :1 x 1 2 ( ;1 39 ) ) Q(x) + P o BR ;1 B T P o > The largest positively invariant set such that Q(x) + P o BR ;1 B T P o > 8 x 2 Z, corresponds to = 1:8

Example 1: Closed-Loop Stability Region Q(x) + P o B R -1 B T P o > 1 Concentration of B 5 Q(x) + P o B R -1 B T P o > -5-1 -5 5 1 Concentration of A

Example 1: Closed-Loop Simulations 5 4 Concentration 3 2 C B C A 1 Nominal Inputs Optimal Control 5 1 15 2 25 Time

Z 1 Constrained ITNQOC Problem x T Qx + T (u)r(u) dt = inf u () s.t. _x = A(x)x + B(x)(u) x() = where () : < M! U, is dened as = arg min (u) 2U jju ; jj Further Assumptions: (H) U is convex and contains the origin in its interior (H) 2 X o ^= f 2 < n j 9 u s.t. () < 1g

( Z T " x T Z 1 (x)d# x = J(x) Constrained FTNQOC Problem Consider the following family of nite-time optimal control problems. x T Qx + T (u)r(u) dt + J(x(T ))) T () = inf u s.t. _x = A(x)x + B(x)(u) x() = where

of Constrained and Unconstrained Equivalence Problems X = f 2 < n j K() 2 Ug O 1 ^= 2 < n j x(t) 2 X 8 t > Dene where K(x) is the unconstrained optimal feedback gain. Lemma: Let K(x) be bounded for all x in a neighborhood of the origin. Then 2 intfug =) 2 intfo 1 g

(i) 1 ^=lim T!1 T exists for all 2 X o (ii) T = 1, T = T + 8, u T = u T + 8, x T (T ) 2 O 1 (iii) If 9 T > such that x T (T ) 2 O 1 8 2 X o 9 T s.t. x T (T ) 2 O 1 Finite-Time Solution to Innite-Time Problem Lemma: Let (H1-H) hold and assume the solution to T is unique. Then then T = Theorem: Let (H1-H) hold and assume the solution to T is unique. Then

8 < 2 n j 9 u s.t. lim jjx(t)jj < = t!1 : X max = X o 9 = Equivalence of Initial Condition Sets Dene the set of constrained stabilizable initial conditions as X max = Theorem: Let (H1-H) hold and assume the solution to T is unique. Then

Z 1 @4k B C Ass + F a V x1 (t) 2 + ru(t) 2 dt Example 2 Consider the CSTR of Example 1 with k b = 1 _x 1 = ;2k a x 2 1 ; C x 1 + F A u V with objective function (r = (F=V ) 2 ) and constraints u(t) + C Ainss 2 [ 1] 8 t ) X max = (;3:89 1)

Example 2 (continued) The resulting SDRE is: 2(;2k a x 1 ; 2k a C Ass ) (x 1 ) + 1 ; 2 (x 1 ) = Which has solution! (x 1 ) = 2 k a x 1 + k a C Ass + r (k a x 1 + k a C Ass ) 2 + 1

u(x) = ; V F k ax 1 + k a C Ass + r (k a x 1 + k a C Ass ) 2 + 1! x 1 Example 2 (continued) The Unconstrained Optimal Control Policy is and O 1 = (;:4 1) Finally, the Closed-Loop System is _x = ;x 1 r (ka x 1 + k a C Ass ) 2 + 1

Example 2: Optimal State Trajectories 1..8 Concentration of A..4 No Constraints Constrained Input Nominal Input.2. 5 1 15 2 25 Time

Example 2: Optimal Input Policies No Constraints Constrained Input C A in 4 2 5 1 15 2 25 Time

Acknowledgments The Following Financial Support is Gratefully Acknowledged: NSF GER-95545 NSF CTS-9312489 DOEd P2A432-95