Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Similar documents
Robotics. Control Theory. Marc Toussaint U Stuttgart

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

H 2 Optimal State Feedback Control Synthesis. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

Suppose that we have a specific single stage dynamic system governed by the following equation:

Lecture 4 Continuous time linear quadratic regulator

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

An Introduction to Linear Matrix Inequalities. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

6 OUTPUT FEEDBACK DESIGN

Homework Solution # 3

2 The Linear Quadratic Regulator (LQR)

Course Outline. Higher Order Poles: Example. Higher Order Poles. Amme 3500 : System Dynamics & Control. State Space Design. 1 G(s) = s(s + 2)(s +10)

9 Controller Discretization

Raktim Bhattacharya. . AERO 632: Design of Advance Flight Control System. Norms for Signals and Systems

Raktim Bhattacharya. . AERO 632: Design of Advance Flight Control System. Preliminaries

Lecture 2 and 3: Controllability of DT-LTI systems

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

Raktim Bhattacharya. . AERO 422: Active Controls for Aerospace Vehicles. Dynamic Response

EE C128 / ME C134 Feedback Control Systems

Steady State Kalman Filter

Semidefinite Programming Duality and Linear Time-invariant Systems

The Inverted Pendulum

Extensions and applications of LQ

EE363 Review Session 1: LQR, Controllability and Observability

6. Linear Quadratic Regulator Control

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

Lecture 5 Linear Quadratic Stochastic Control

Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 2015

EE221A Linear System Theory Final Exam

Denis ARZELIER arzelier

Problem 1 Cost of an Infinite Horizon LQR

Linear Quadratic Optimal Control Topics

Formula Sheet for Optimal Control

Appendix A Solving Linear Matrix Inequality (LMI) Problems

Linear Quadratic Regulator (LQR) Design II

Linear Quadratic Regulator (LQR) Design I

EE363 homework 7 solutions

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

CDS Solutions to the Midterm Exam

Introduction to Nonlinear Control Lecture # 3 Time-Varying and Perturbed Systems

7.1 Linear Systems Stability Consider the Continuous-Time (CT) Linear Time-Invariant (LTI) system

EE C128 / ME C134 Final Exam Fall 2014

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Robust Control 5 Nominal Controller Design Continued

ECEEN 5448 Fall 2011 Homework #4 Solutions

Raktim Bhattacharya. . AERO 422: Active Controls for Aerospace Vehicles. Basic Feedback Analysis & Design

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL

Grammians. Matthew M. Peet. Lecture 20: Grammians. Illinois Institute of Technology

Time-Invariant Linear Quadratic Regulators!

MODERN CONTROL DESIGN

CONTROL DESIGN FOR SET POINT TRACKING

OPTIMAL SPACECRAF1 ROTATIONAL MANEUVERS

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

EEE582 Homework Problems

EE363 homework 2 solutions

Stochastic and Adaptive Optimal Control

Lyapunov Stability Theory

Feedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation

Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1

Nonlinear Control Systems

Theory in Model Predictive Control :" Constraint Satisfaction and Stability!

Riccati Equations and Inequalities in Robust Control

Solution of Linear State-space Systems

Outline. 1 Linear Quadratic Problem. 2 Constraints. 3 Dynamic Programming Solution. 4 The Infinite Horizon LQ Problem.

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

Topic # Feedback Control Systems

Linear Quadratic Regulator (LQR) I

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

4F3 - Predictive Control

Automatic Control 2. Nonlinear systems. Prof. Alberto Bemporad. University of Trento. Academic year

LMI Methods in Optimal and Robust Control

5. Observer-based Controller Design

POLE PLACEMENT. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 19

Lecture 2: Discrete-time Linear Quadratic Optimal Control

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems.

Australian Journal of Basic and Applied Sciences, 3(4): , 2009 ISSN Modern Control Design of Power System

CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems

EE C128 / ME C134 Fall 2014 HW 9 Solutions. HW 9 Solutions. 10(s + 3) s(s + 2)(s + 5) G(s) =

Outline. Linear regulation and state estimation (LQR and LQE) Linear differential equations. Discrete time linear difference equations

Nonlinear Control. Nonlinear Control Lecture # 8 Time Varying and Perturbed Systems

6.241 Dynamic Systems and Control

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

TRACKING AND DISTURBANCE REJECTION

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points

Topics in control Tracking and regulation A. Astolfi

EG4321/EG7040. Nonlinear Control. Dr. Matt Turner

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems

NPTEL Online Course: Control Engineering

LECTURE NOTES IN CALCULUS OF VARIATIONS AND OPTIMAL CONTROL MSc in Systems and Control. Dr George Halikias

LINEAR-CONVEX CONTROL AND DUALITY

Static and Dynamic Optimization (42111)

Lecture 9: Discrete-Time Linear Quadratic Regulator Finite-Horizon Case

Model Predictive Control Short Course Regulation

OPTIMAL CONTROL AND ESTIMATION

X 2 3. Derive state transition matrix and its properties [10M] 4. (a) Derive a state space representation of the following system [5M] 1

Transcription:

.. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Quadratic Lyapunov Functions

Quadratic Stability Dynamical system is quadratically stable if ẋ = Ax,. V (x), V. Let V (x) = x T P x, P S n ++ (P = P T > ) Therefore, Therefore V (x) = ẋ T P x + x T P ẋ = x T A T P x + x T P Ax = x T ( A T P + P A ) x V = x T ( A T P + P A ) x = A T P + P A. AERO 632, Instructor: Raktim Bhattacharya 3 / 3

Lyapunov Equation We can write as for Q = Q T. A T P + P A. A T P + P A + Q = Interpretation For linear system ẋ = Ax, if V (x) = x T P x, V (x) = ẋ T P x + x T P ẋ = (Ax) T P x + x T P (Ax) = x T Qx. If V (x) = x T P x is generalized energy, V = x T Qx is generalized dissipation. AERO 632, Instructor: Raktim Bhattacharya 4 / 3

Stability Condition If P >, Q >, then ẋ = Ax is globally asymptotically stable Rλ i (A) <. Note that for P = P T >, eigenvalues are real = λ min (P ) x T x x T P x λ max (P ) x T x = V = x T Qx λ min (Q)x T x λ min(q) λ max (P ) xt P x = αv (x) AERO 632, Instructor: Raktim Bhattacharya 5 / 3

Lyapunov Integral If A is stable, then P =. e tat Qe ta dt, for any Q = Q T >. Proof: Substitute it in LHS of Lyapunov equation to get, ( ) A T P + P A = A T e tat Qe ta + e tat Qe ta A dt, = ( ) d dt etat Qe ta dt, = e tat Qe ta, = Q. AERO 632, Instructor: Raktim Bhattacharya 6 / 3

. Computation of x 2,Q Recall x 2 2 := Define weighted norm as x 2 2,Q := If x(t) is solution of ẋ = Ax, Substituting we get, x 2 2,Q = x T x dt. x(t) := e ta x. x T Qx dt. x T e tat Qe ta dt = x T P x assuming A is stable Cost-to-go interpretation AERO 632, Instructor: Raktim Bhattacharya 7 / 3

LQR Problem

Linear Quadratic Regulator Problem Statement Given system ẋ = Ax + Bu, Determine u (t) that solves Or min J := u(t) = y T y dt. y = Cx + Du. min y 2 with x() = x. u(t) ( x T C T Cx + x T C T Du + u T D T Cx + u T D T Du ) dt = ( x T C T Cx + u T D T Du ) dt. Assume C T D = for simplicity. AERO 632, Instructor: Raktim Bhattacharya 9 / 3

Linear Quadratic Regulator Solution as Optimal Control Problem. min u subject to ( x T Qx + u T Ru ) dt, Q = Q T, R = R T > ẋ = Ax + Bu, x() = x. Euler Lagrange Equations Hamilton-Jacobi-Bellman Equation Dynamic Programming AERO 632, Instructor: Raktim Bhattacharya 1 / 3

Euler Langrange Formulation

Linear Quadratic Regulator Solution as Optimal Control Problem EL Formulation. T min L(x, u)dt + Φ(x(T )), subject to ẋ = f(x, u). u Define H = L + λ T f. Euler-Lagrange Equations Our Problem H u = λt = H x λ(t ) = ϕ x (x(t )) min u T ( x T Qx + u T Ru ) dt, subject to ẋ = Ax + Bu. Define H = x T Qx + u T Ru + λ T (Ax + Bu). AERO 632, Instructor: Raktim Bhattacharya 12 / 3

Linear Quadratic Regulator Solution as Optimal Control Problem EL Formulation Our Problem 1 min u 2 T Define H = 1 2 EL Equations. ( x T Qx + u T Ru ) dt, subject to ẋ = Ax + Bu. ( x T Qx + u T Ru ) + λ T (Ax + Bu). (1) H u = = u T R + λ T B = = u = R 1 B T λ. (2) λt = H x = x T Q λ T A (3) λ(t ) =. AERO 632, Instructor: Raktim Bhattacharya 13 / 3

Linear Quadratic Regulator Solution as Optimal Control Problem EL Formulation Let λ(t) = P (t)x(t) From EL(2) we get λ = Qx A T P x. = λ = P x + P ẋ = P x + P (Ax + Bu), = P x + P (Ax BR 1 B T P x), = ( P + P A P BR 1 B T P )x. = ( P + P A + A T P P BR 1 B T P + Q)x = = P + P A + A T P P BR 1 B T P + Q =. Riccati Differential Equation In the steady-state T, P =, P A + A T P P BR 1 B T P + Q =. Algebraic Riccati Equation AERO 632, Instructor: Raktim Bhattacharya 14 / 3

Linear Quadratic Regulator Solution as Optimal Control Problem EL Formulation. 1 min u 2 is equivalent to ( x T Qx + u T Ru ) dt, subject to ẋ = Ax + Bu. P A + A T P P BR 1 B T P + Q =, u = R 1 B T P. AERO 632, Instructor: Raktim Bhattacharya 15 / 3

Hamilton-Jacobi-Bellman Formulation

. Hamilton-Jacobi-Bellman Approach Let subject to V 1 (x(t)) = min u[t, ) 2 t ẋ = Ax + Bu. (x T Qx + u T Ru)dt AERO 632, Instructor: Raktim Bhattacharya 17 / 3

. Hamilton-Jacobi-Bellman Approach contd. V (x(t)) = min = min u[t,t+ t] 1 2 u[t, ) { t+ t t Let V (x) := x T P x, therefore, t (x T Qx + u T Ru)dt 1 2 (xt Qx + u T Ru)dt + V (x(t + t)) } { 1 V (x(t)) = min u[t,t+ t] 2 (xt Qx + u T Ru) t + V (x(t))+ } (Ax + Bu) T P x t + x T P (Ax + Bu) t + H.O.T AERO 632, Instructor: Raktim Bhattacharya 18 / 3

. Hamilton-Jacobi-Bellman Approach contd. { 1 = min u[t,t+ t] 2 (xt Qx + u T Ru) + (Ax + Bu) T P x+ } x T P (Ax + Bu) + H.O.T =. lim t { 1 = min u 2 (xt Qx + u T Ru)+ } (Ax + Bu) T P x + x T P (Ax + Bu) = Quadratic in u, = u = R 1 B T P x. Optimal controller is state-feedback. AERO 632, Instructor: Raktim Bhattacharya 19 / 3

Variational Approach

A Third Approach Given dynamics ẋ = Ax + Bu,. with controller u = Kx, find K that minimizes J := 1 2 (x T Qx + u T Ru)dt = 1 2 The closed-loop dynamics is The solution is therefore, ẋ = Ax + Bu = (A + BK)x = A c x. x(t) = e ta c x. x T (Q + K T RK)xdt. The cost function is therefore, J := 1 ( ) 2 xt e tat c (Q + K T RK)e ta c dt x = 1 2 xt P x. AERO 632, Instructor: Raktim Bhattacharya 21 / 3

A Third Approach contd. Apply the following trick Or Or. d dt etat c (Q + K T RK)e ta c dt = ( ) A T c e tat c (Q + K T RK)e ta c dt ( ) + e tat c (Q + K T RK)e tac dt [ e tat c (Q + K T RK)e tac ] = AT c P + P A c A T c P + P A c + Q + K T RK =, (A + BK) T P + P (A + BK) + Q + K T RK =. A c AERO 632, Instructor: Raktim Bhattacharya 22 / 3

A Third Approach contd.. The optimal cost is therefore, J = 1 2 xt P x = J P =. P Variation δp from P should result in δj = Let P = P + δp, = J = 1 2 xt P x + 1 2 xt δp x }{{} δj δj = = δp = AERO 632, Instructor: Raktim Bhattacharya 23 / 3

A Third Approach. contd. Substitute P = P + δp, and K = K + δk in the equality constraint to get, or (A + BK) T P + P (A + BK) + Q + K T RK =, (A + B(K + δk)) T (P + δp ) + (P + δp )(A + B(K + δk)) + Q + (K + δk) T R(K + δk) =, (A + BK ) T P + P (A + BK ) + Q + K T RK + δp (A + BK ) + ( ) T + δk T (B T P + RK ) + ( ) T = K = R 1 B T P. + H.O.T =. AERO 632, Instructor: Raktim Bhattacharya 24 / 3

LMI Formulation

Problem Formulation Find gain K such that u = Kx minimizes. subject to dynamics and (x T Qx + u T Ru)dt, ẋ = Ax + Bu, x() = x. AERO 632, Instructor: Raktim Bhattacharya 26 / 3

. An Upper Bound on the Cost-to-go If V (x) > such that Integrating from [, T ], gives us or T dv dt (xt Qx + u T Ru). dv dt dt T V (x(t )) V (x()) Since V (x(t )) for any T = V (x()) (x T Qx + u T Ru)dt, T T (x T Qx + u T Ru)dt. (x T Qx + u T Ru)dt, AERO 632, Instructor: Raktim Bhattacharya 27 / 3

. An Upper Bound on the Cost-to-go Since V (x(t )) for any T or = V (x()) V (x()) T (x T Qx + u T Ru)dt, (x T Qx + u T Ru)dt.. Sufficient condition for upper-bound on cost-to-go.. If V (x) > such that. dv dt (xt Qx + u T Ru). Idea: Minimize upper-bound to get optimal K. AERO 632, Instructor: Raktim Bhattacharya 28 / 3

Optimization Problem. Find P = P T > and K such that with V := x T P x, subject to min P,K V (x()) = x()t P x() Cost Function V x T (Q + K T RK)x Constraint Function. Or equivalently. subject to min trp P,K. (A + BK) T P + P (A + BK) + Q + K T RK. AERO 632, Instructor: Raktim Bhattacharya 29 / 3

Optimization Problem. subject to min trp P,K (A + BK) T P + P (A + BK) + Q + K T RK. Learn about Linear Matrix Inequalities AERO 632, Instructor: Raktim Bhattacharya 3 / 3