Linear Quadratic Regulator (LQR) I

Similar documents
Linear Quadratic Regulator (LQR) Design I

Linear Quadratic Regulator (LQR) II

Linear Quadratic Regulator (LQR) Design II

Constrained Optimal Control I

Constrained Optimal Control. Constrained Optimal Control II

Using Lyapunov Theory I

Course Outline. Higher Order Poles: Example. Higher Order Poles. Amme 3500 : System Dynamics & Control. State Space Design. 1 G(s) = s(s + 2)(s +10)

Feedback Optimal Control for Inverted Pendulum Problem by Using the Generating Function Technique

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

Robust Control 5 Nominal Controller Design Continued

Gain Scheduling and Dynamic Inversion

Suppose that we have a specific single stage dynamic system governed by the following equation:

Chapter 2 SDOF Vibration Control 2.1 Transfer Function

Homework Solution # 3

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Lecture 4 Continuous time linear quadratic regulator

Control Systems. Design of State Feedback Control.

Quadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Robotics. Control Theory. Marc Toussaint U Stuttgart

4F3 - Predictive Control

Lecture 6 Classical Control Overview IV. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules

4F3 - Predictive Control

Feedback Linearization

Contents. 1 State-Space Linear Systems 5. 2 Linearization Causality, Time Invariance, and Linearity 31

EE363 homework 2 solutions

Lecture 5 Classical Control Overview III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

Lecture 9: Discrete-Time Linear Quadratic Regulator Finite-Horizon Case

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

5. Observer-based Controller Design

The Inverted Pendulum

Linear-Quadratic Optimal Control: Full-State Feedback

Optimal Control of Nonlinear Systems using RBF Neural Network and Adaptive Extended Kalman Filter

Linear control of inverted pendulum

Extensions and applications of LQ

6. Linear Quadratic Regulator Control

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

Lecture 2: Discrete-time Linear Quadratic Optimal Control

6 OUTPUT FEEDBACK DESIGN

Lecture : Feedback Linearization

Classical Numerical Methods to Solve Optimal Control Problems

1 Relative degree and local normal forms

Linear State Feedback Controller Design

Australian Journal of Basic and Applied Sciences, 3(4): , 2009 ISSN Modern Control Design of Power System

Solutions for Homework #8. Landing gear

Topic # Feedback Control Systems

Here represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.

Dynamic Inversion Design II

Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 2015

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

Numerics and Control of PDEs Lecture 1. IFCAM IISc Bangalore

Time-Invariant Linear Quadratic Regulators!

Optimization Tutorial 1. Basic Gradient Descent

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 8: Basic Lyapunov Stability Theory

Static and Dynamic Optimization (42111)

Optimal control of nonlinear systems with input constraints using linear time varying approximations

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

Reverse Order Swing-up Control of Serial Double Inverted Pendulums

Design and Comparison of Different Controllers to Stabilize a Rotary Inverted Pendulum

Dissipativity. Outline. Motivation. Dissipative Systems. M. Sami Fadali EBME Dept., UNR

Lecture 6. Foundations of LMIs in System and Control Theory

Theory in Model Predictive Control :" Constraint Satisfaction and Stability!

Exam. 135 minutes, 15 minutes reading time

Robust feedback linearization

Products and Convolutions of Gaussian Probability Density Functions

EE C128 / ME C134 Final Exam Fall 2014

OPTIMAL CONTROL AND ESTIMATION

Lyapunov Stability Analysis: Open Loop

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation

522. Optimal robust control of aeroelastic system vibrations

18-660: Numerical Methods for Engineering Design and Optimization

6.241 Dynamic Systems and Control

Semidefinite Programming Basics and Applications

MODERN CONTROL DESIGN

Hamilton-Jacobi-Bellman Equation Feb 25, 2008

EG4321/EG7040. Nonlinear Control. Dr. Matt Turner

Example: When describing where a function is increasing, decreasing or constant we use the x- axis values.

EE C128 / ME C134 Feedback Control Systems

Linearization problem. The simplest example

Outline. 1 Linear Quadratic Problem. 2 Constraints. 3 Dynamic Programming Solution. 4 The Infinite Horizon LQ Problem.

Introduction to centralized control

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures

Comparison of LQR and PD controller for stabilizing Double Inverted Pendulum System

Stabilization of a 3D Rigid Pendulum

Semidefinite Programming Duality and Linear Time-invariant Systems

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)

Topic # Feedback Control

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

y,z the subscript y, z indicating that the variables y and z are kept constant. The second partial differential with respect to x is written x 2 y,z

Lecture 8 Optimization

Denis ARZELIER arzelier

SIO 211B, Rudnick. We start with a definition of the Fourier transform! ĝ f of a time series! ( )

Switched systems: stability

Optimization of Linear Systems of Constrained Configuration

Chapter 8 Stabilization: State Feedback 8. Introduction: Stabilization One reason feedback control systems are designed is to stabilize systems that m

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

EL2520 Control Theory and Practice

SWEEP METHOD IN ANALYSIS OPTIMAL CONTROL FOR RENDEZ-VOUS PROBLEMS

10. Joint Moments and Joint Characteristic Functions

Math Review and Lessons in Calculus

Transcription:

Optimal Control, Guidance and Estimation Lecture Linear Quadratic Regulator (LQR) I Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore Generic Optimal Control Problem Perormance Index (to minimize / maximize): Path Constraint: Xɺ = t X U Boundary Conditions: X = ϕ (, ) + (,, ) J t X L t X U dt (,, ) t t t = X : Speciied ( ) : Fixed, X t : Free Pro. Radhakant Padhi, AE Dept., IISc-Bangalore

LQR Design: Problem Objective o drive the state X o a linear (rather linearized) system Xɺ = A X + BU to the origin by minimizing the ollowing quadratic perormance index (cost unction) J = ( X S X ) + ( X Q X + U RU ) dt where t S, Q (psd), R > (pd) t Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 3 LQR Design: Guideline or Selection o Weighting Matrices S (psd), Q (psd), R > (pd) hese are usually chosen as diagonal matrices, with s q r i i i = maximum expected/acceptable value o / ( xi ) = maximum expected/acceptable value o ( / xi ) = maximum expected/acceptable value o ( / u i ) Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 4

LQR Design: Some Facts to Remember { } he pair A, B needs to be controllable and the pair A, Q needs to be detectable { } S (psd), Q (psd), R > (pd) (these are usually chosen as diagonal matrices) By deault, it is assumed that t Constrained problems (i.e. problems with state and control inequality constraints) are not considered here. hose will be considered later. Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 5 LQR Design: Problem Statement Perormance Index (to minimize): Path Constraint: t J = ( X S X ) + X Q X + U RU dt t ϕ( X ) L( X, U ) Xɺ = A X + BU Boundary Conditions: X t = X : Speciied ( ) : Fixed, X t : Free Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 6

LQR Design: Necessary Conditions o Optimality erminal penalty: ϕ ( X ) = ( X S X ) Hamiltonian: H = X Q X + U RU + AX + BU State Equation: Xɺ = AX + BU Costate Equation: Optimal Control Eq.: Boundary Condition: ( H / X ) ( QX A ) ɺ = = + ( / ) = = H U U R B ( ϕ / X ) = = S X Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 7 LQR Design: Derivation o Riccati Equation Guess: Justiication: ( t) P ( t) X ( t) = From unctional analysis theory o normed linear space, lies in the "dual space" o X t o all continuous linear unctionals o X t. ( t), which is the space consisting Reerence: Optimization by Vector Space Methods D. G. Luenberger, John Wiley & Sons, 969. Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 8

LQR Design: Derivation o Riccati Equation Guess ( t) = P ( t) X ( t) ɺ = PX ɺ + PXɺ = PX ɺ + P AX + BU = PX ɺ + P( AX BR B ) = PX ɺ + P( AX BR B PX ) ( QX + A PX ) = ( Pɺ + PA PBR B P) X ( Pɺ + PA + A P PBR B P + Q) X = Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 9 LQR Design: Derivation o Riccati Equation Riccati equation ɺ + + + = P PA A P PBR B P Q Boundary condition ( ) = ( is ree) P t X S X X ( ) P t = S Pro. Radhakant Padhi, AE Dept., IISc-Bangalore

LQR Design: Solution Procedure Use the boundary condition P t = S and integrate the Riccati Equation backwards rom to t t Store the solution history or the Riccati matrix Compute the optimal control online U = R B P X = K X Pro. Radhakant Padhi, AE Dept., IISc-Bangalore LQR Design: Ininite ime Regulator Problem heorem (By Kalman) As t, or constant Q and R matrices, Pɺ t Algebraic Riccati Equation (ARE) Note: PA A P PBR B P Q + + = ARE is still a nonlinear equation or the Riccati matrix. It is not straightorward to solve. However, eicient numerical methods are now available. A positive deinite solution or the Riccati matrix is needed to obtain a stabilizing controller. Pro. Radhakant Padhi, AE Dept., IISc-Bangalore

Example : Stabilization o Inverted Pendulum Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore A Motivating Example: Stabilization o Inverted Pendulum System dynamics: ɺɺ θ = ω θ u, ω = g / L n n ( Linearized about vertical equilibrium point) u θ L mg System dynamics (state space orm): Deine: x = θ, x = ɺ θ xɺ x u x = ωn + x ɺ Xɺ A X B Perormance Index (to minimize): J = θ u dt + c Q =, R = c Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 4

A Motivating Example: Stabilization o Inverted Pendulum ARE: PA A P PBR B P Q Let + + = P p p = p p 3 (a symmetric matrix) pω n p pωn p3ω n c p c p p 3 + + p = 3ωn p p p c p p3 c p3 Equations: pωn c p + = p = 4 ω n ± ωn + c c p + p ω c p p = (repeated) 3 n 3 p c p = p = ± p c 3 3 u θ L mg Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 5 A Motivating Example: Stabilization o Inverted Pendulum However, Hence, p p 3 4 n n 3 3 n 3 3 3 n is a diagonal term, which needs to be real and positive. needs to be positive. hereore p = ω + ω + c, p = p c c Moreover, p + p ω c p p = p = c p p p ω (not needed in this problem) u θ L mg Gain Matrix: K = R B P = c p c p 3 Control: = = + ɺ ( θ 3θ ) u K X c p p Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 6

A Motivating Example: Stabilization o Inverted Pendulum Analysis Open-Loop System: I A = = ωn = ω = ± ω n n ( right hal pole: unstable system) u θ L mg Closed-Loop System: Deine: ACL = A BK = ωn c p c p 3 Closed-Loop Poles: I A = CL ω = ω + c 4 n p = + c ( ωn ω ) p3 = p = + c c ( ωn ω ) / Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 7 A Motivating Example: Stabilization o Inverted Pendulum Analysis Closed-Loop Poles: ( n ) / + ω + ω + ω =, = ω + ω ± ω ω 4 ( Note: ω = ωn + c > ωn ) ( n ) j ( n ) / / u θ L mg Both o the closed-loop poles are strictly in the let-hal plane. Hence, the closed-loop is guaranteed to be asymptotically stable. Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 8

Example : Finite-time time emperature Control Pro. Radhakant Padhi Dept. o Aerospace Engineering Indian Institute o Science - Bangalore Example: Finite ime emperature Control Problem System dynamics : ɺ θ = a θ θ + bu where a, b : Constants θ θ u a a : emperature : Ambient temperature (Constant = C) : Heat input Pro. Radhakant Padhi, AE Dept., IISc-Bangalore

Problem ormulations Case : Case : Cost Function: Cost Function: J θ = ( t ) t t u dt J = s ( θ 3) + u dt = θ = 3 C (Hard constraint) s > : Weightage ( t ) i.e. θ 3 C (Sot Constraint) Pro. Radhakant Padhi, AE Dept., IISc-Bangalore Solution: Solution: x ( θ θa ), θ ( ) = θa ( θ θ ) xɺ = ax + bu, x = = Η = + + ɺ Η = = a x Η = u = b u u ax bu a Pro. Radhakant Padhi, AE Dept., IISc-Bangalore a Necessary conditions xɺ = ax + bu ɺ = a u = b

Solution: Case (Hard constraint) a( t t ) a( t t) = e = e u = be a t ( t) a t xɺ = ax b e ( t) aking laplace transorm: sx ( s) x( ) at ax s b e = s a Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 3 Solution: Case (Hard constraint) s a at = b e a s a s + a Hence = a at X s = b e at at at x t b e e e Unknown ( θ θ ) However, x( t ) = = C a Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 4

Solution: Case (Hard constraint) at ( e ) at ( e ) at at at x( t ) = = b e e e b = a a = b x( t) = b a at at a e e at at at e ( e e ) = b ( e ) a e e at at at Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 5 Solution: Case (Hard constraint) Note : x( t ) = at ( e at e ) at ( e at e ) = (i.e. he boundary condition is "exactly met"!) Controller : u( t) = be at ( ) ae = b ( at ) ( at at e b e e ) a t t a Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 6

Solution: Case (Sot constraint) θ 3 C x C. Hence the cost unction is t J = s ( x ) + u dt = s ( x ) x = + s However, we have b x t e e e a at at at = ( ) Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 7 Solution: Case (Sot constraint) At t = t, x t = e = + a s b at b at i. e. + ( e ) = s a s a i. e. = at a + s b ( e ) Hence a( t ) t a t t s a = e = e a + s b e at Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 8

Solution: Case (Sot constraint) u t = b = be = be s a a + s b e a t t at at s abe s b ae + e e a t t at at at Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 9 Correlation between hard and sot constraint results As s, at abe lim u( t) = lim b ae + e e s s ( e ) S. C. s at at at at ae = = u t b e at at H. C. i. e. he "sot constraint" problem behaves like the "hard constraint" problem when s. Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 3

hanks or the Attention.!! Pro. Radhakant Padhi, AE Dept., IISc-Bangalore 3