Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE)

Similar documents
Feedback Optimal Control of Low-thrust Orbit Transfer in Central Gravity Field

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

WHEN deciding where spacecraft should land a balance exists between engineering safety and scientific (or programmatic)

Optimal Control Using an Algebraic Method for Control-Affine Nonlinear Systems. Abstract

EE C128 / ME C134 Feedback Control Systems

Robust SDC Parameterization for a Class of Extended Linearization Systems

SDRE BASED LEADER-FOLLOWER FORMATION CONTROL OF MULTIPLE MOBILE ROBOTS

Research Article Approximate Solutions to Nonlinear Optimal Control Problems in Astrodynamics

Homework Solution # 3

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

A Case Study of Feedback Controller and Estimator Designs for Nonlinear Systems via the State-dependent Riccati Equation Technique

Streamlining of the state-dependent Riccati equation controller algorithm for an embedded implementation

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules

Lecture 4 Continuous time linear quadratic regulator

NONLINEAR CONTROL FOR HIGH-ANGLE-OF-ATTACK AIRCRAFT FLIGHT USING THE STATE-DEPENDENT RICCATI EQUATION

Optimal control of nonlinear systems with input constraints using linear time varying approximations

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

Control Systems I. Lecture 2: Modeling. Suggested Readings: Åström & Murray Ch. 2-3, Guzzella Ch Emilio Frazzoli

CDS 110b: Lecture 2-1 Linear Quadratic Regulators

MODERN CONTROL DESIGN

Modelling and State Dependent Riccati Equation Control of an Active Hydro-Pneumatic Suspension System

Pierre Bigot 2 and Luiz C. G. de Souza 3

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

A New Algorithm for Solving Cross Coupled Algebraic Riccati Equations of Singularly Perturbed Nash Games

Optimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

Linear Quadratic Regulator (LQR) Design I

Approximate Solutions to Nonlinear Optimal Control Problems in Astrodynamics

Comparison of LQR and PD controller for stabilizing Double Inverted Pendulum System

ECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming

Suppose that we have a specific single stage dynamic system governed by the following equation:

Integrated Guidance and Control of Missiles with Θ-D Method

Linear Quadratic Regulator (LQR) I

EN Nonlinear Control and Planning in Robotics Lecture 3: Stability February 4, 2015

Lecture 5 Linear Quadratic Stochastic Control

6. Linear Quadratic Regulator Control

Performance Comparison of PSO Based State Feedback Gain (K) Controller with LQR-PI and Integral Controller for Automatic Frequency Regulation

NEAR-OPTIMAL FEEDBACK GUIDANCE FOR AN ACCURATE LUNAR LANDING JOSEPH PARSLEY RAJNISH SHARMA, COMMITTEE CHAIR MICHAEL FREEMAN KEITH WILLIAMS A THESIS

(q 1)t. Control theory lends itself well such unification, as the structure and behavior of discrete control

EML5311 Lyapunov Stability & Robust Control Design

Linear State Feedback Controller Design

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)

LMI Methods in Optimal and Robust Control

Time-Invariant Linear Quadratic Regulators!

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures

A Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley

Autonomous Mobile Robot Design

Stability of Nonlinear Systems An Introduction

Nonlinear H(Infinity) Missile Longitudinal Autopilot Design with Theta-D Method

Tracking control of elastic joint parallel robots via state-dependent Riccati equation

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

Linear-Quadratic Optimal Control: Full-State Feedback

Problem 1 Cost of an Infinite Horizon LQR

Optimal Control. McGill COMP 765 Oct 3 rd, 2017

A Robust Controller for Scalar Autonomous Optimal Control Problems

LINEAR-CONVEX CONTROL AND DUALITY

Design of Non-Linear Controller for a flexible rotatory beam using State-Dependent Riccati Equation (SDRE) control

Stabilization of constrained linear systems via smoothed truncated ellipsoids

The Rationale for Second Level Adaptation

WEIGHTING MATRICES DETERMINATION USING POLE PLACEMENT FOR TRACKING MANEUVERS

Coordinated Tracking Control of Multiple Laboratory Helicopters: Centralized and De-Centralized Design Approaches

On the stability of receding horizon control with a general terminal cost

Control Systems I. Lecture 2: Modeling and Linearization. Suggested Readings: Åström & Murray Ch Jacopo Tani

The Inverted Pendulum

A Novel Integral-Based Event Triggering Control for Linear Time-Invariant Systems

LYAPUNOV theory plays a major role in stability analysis.

ESC794: Special Topics: Model Predictive Control

Hybrid Systems Course Lyapunov stability

Locally optimal controllers and application to orbital transfer (long version)

Optimizing Economic Performance using Model Predictive Control

Comparison of Feedback Controller for Link Stabilizing Units of the Laser Based Synchronization System used at the European XFEL

An LQ R weight selection approach to the discrete generalized H 2 control problem

Model Predictive Regulation

Nonlinear Observer Design for Dynamic Positioning

MCE693/793: Analysis and Control of Nonlinear Systems

Optimal Selection of the Coefficient Matrix in State-Dependent Control Methods

Theory in Model Predictive Control :" Constraint Satisfaction and Stability!

EE C128 / ME C134 Final Exam Fall 2014

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

Automatic Generation Control Using LQR based PI Controller for Multi Area Interconnected Power System

Adaptive Predictive Observer Design for Class of Uncertain Nonlinear Systems with Bounded Disturbance

6.241 Dynamic Systems and Control

Piecewise Linear Quadratic Optimal Control

Intermediate Process Control CHE576 Lecture Notes # 2

Lecture 15: H Control Synthesis

Enlarged terminal sets guaranteeing stability of receding horizon control

6.241 Dynamic Systems and Control

Feedback linearization control of systems with singularities: a ball-beam revisit

Lecture 6. Foundations of LMIs in System and Control Theory

Linearization problem. The simplest example

Controllability Issues innonlinear State-Dependent. Riccati Equation Control

H State-Feedback Controller Design for Discrete-Time Fuzzy Systems Using Fuzzy Weighting-Dependent Lyapunov Functions

Australian Journal of Basic and Applied Sciences, 3(4): , 2009 ISSN Modern Control Design of Power System

Linear Feedback Control Using Quasi Velocities

MS-E2133 Systems Analysis Laboratory II Assignment 2 Control of thermal power plant

DERIVATIVE FREE OUTPUT FEEDBACK ADAPTIVE CONTROL

Computational Issues in Nonlinear Control and Estimation. Arthur J Krener Naval Postgraduate School Monterey, CA 93943

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

5. Observer-based Controller Design

Transcription:

Nonlinear Optimal Tracking Using Finite-Horizon State Dependent Riccati Equation (SDRE) AHMED KHAMIS Idaho State University Department of Electrical Engineering Pocatello, Idaho USA khamahme@isu.edu D. SUBBARAM NAIDU Idaho State University Department of Electrical Engineering Pocatello, Idaho USA naiduds@isu.edu Abstract: A number of computational techniques have been proposed for synthesizing nonlinear control based on state dependent Riccati equation (SDRE). Most of these techniques focusing on infinite-horizon regulator problems. This paper presents a new and computationally efficient approximate online technique used for finite-horizon nonlinear tracking problems. This technique based on change of variables that converts the differential Riccati equation to a linear Lyapunov differential equation. Illustrative examples are given to show the effectiveness of the proposed technique. Key Words: Optimal control, finite-horizon state dependent Riccati equation, nonlinear tracking 1 Introduction The need to improve performance in controlled systems leads to more and more accurate modeling. However, if a model is a good representation of the real system over a wide range operating points, it is most often nonlinear. Therefore, the ordinarily used linear control techniques become inadequate and it becomes necessary to use some other nonlinear control techniques. And the competitive era of rapid technological change has motivated the rapid development of nonlinear control theory for application to challenging complex dynamical real-world problems 1]. There exist many nonlinear control design techniques, each has benefits and weaknesses. Most of them are limited in their range of applicability, and use of certain nonlinear control technique for a specific system usually demands choosing between different factors, e.g. performance, robustness, optimality, and cost. Some of the well-known nonlinear control techniques are feedback linearization, adaptive control, nonlinear predictive control, sliding mode control, and approximating sequence of Riccati Equations. One of the highly promising and rapidly emerging techniques for nonlinear optimal controllers designing is the SDRE technique. Inspired by the great potential of SDRE for regulation of infinite horizon nonlinear systems 2], SDRE technique for infinite horizon tracking of nonlinear systems 3], and the development of finite- horizon SDRE for finite horizon regulation of nonlinear systems 4], in this paper, we develop the SDRE technique for finite horizon optimal tracking of nonlinear systems based on a change of variable 5], that converts the differential Riccati equation (DRE) to a linear differential Lyapunov equation 6], which are solved in real time at each time step. The reminder of this paper is organized as follows:an overview of infinite-horizon SDRE technique is discussed in Section II. Section III presents the SDRE in finite horizon regulator problem. Section IV presents the nonlinear finite horizon tracking technique via SDRE. Illustrative examples are discussed in Section V. Finally conclusions of this paper are given in Section VI. 2 Overview of Infinite-Horizon SDRE Technique SDRE, which is also referred to as the Frozen Riccati Equation (FRE) 7], first proposed by Pearson (1962) and later expanded by Wernli & Cook (1975), and studied by Mracek & Cloutier (1998) 8]. SDRE has become as a very attractive tool for the systematic design of nonlinear controllers, very common within the control community over the last decade, providing an extremely effective algorithm for nonlinear feedback control design by allowing nonlinearities in the system states while additionally offering great design flexibility through design matrices 1]. The method involves factorization of the nonlinear dynamics into product of a matrix-valued function. Thus, the SDRE algorithm captures the nonlinearities of the system, ISBN: 978-96-474-318-6 37

the amount of time that each approximation is applied. Because of its approximating nature, the SDRE technique is considered a suboptimal solution. However, with the proper choices for the A(x) and B(x) matrices, and with the proper amount of sample times, the SDRE technique can provide a very adequate optimal solution. 3 Nonlinear Finite-Horizon Regulator Via SDRE 3.1 Problem Formulation The nonlinear system considered in this paper is assumed to be in the form: Figure 1: Process of Infinite-Horizon SDRE Technique transforming the original nonlinear system to a linearlike structure with state dependent coefficient (SDC) matrices, and minimizing a nonquadratic performance index with a quadratic-like structure 9]. The Riccati equation using the SDC matrices is then solved online to give the sub optimum control law. Moreover, with enough sample points, the suboptimal solution can be made to be very close to optimal solution. The coefficients of this Riccati equation vary with the each point in state space. The algorithm thus involves solving, at a given point in state space, a SDRE whose point wise stabilizing solution during state evolution yields the SDRE nonlinear feedback control law. As the SDRE depends only on the current state, the computation can be carried out online, in which case the SDRE is defined along the state trajectory. In addition, a primary advantage offered by SDRE to the control designer is the opportunity to make tradeoffs between control effort and state errors by tuning the SDC. Fig.1 shows a process of the infinite-horizon SDRE technique. At each sample time, the following procedure is accomplished. First, the current state vector x is used to calculate numerical values for A(x) and B(x). Then, using the LQR equations, P and K are calculated. Control input u is then calculated and applied to the system. This procedure is then repeated at the next sample time. For the SDRE technique, the ARE is solved at every sample time for each new value of A(x) and B(x). This causes the nonlinear system to be approximated as a series of linear systems. Therefore, shorter time increments increase the accuracy of the control law, because this decreases ẋ(t) = f(x) + g(x)u(t), (1) y(t) = h(x). (2) That nonlinear system can be expressed in a statedependent like linear form, as: ẋ(t) = A(x)x(t) + B(x)u(t), (3) y(t) = C(x)x(t), (4) where f(x) = A(x)x(t), B(x) = g(x), h(x) = C(x)x(t). The goal is to find a state feedback control law of the form u(x) = kx(t), that minimizes a cost function given by 1]: + 1 2 J(x,u) = 1 2 x (t f )Fx(t f ) (5) tf t x (t)q(x)x(t) + u (x)r(x)u(x) ] dt, where Q(x) and F is a symmetric positive semidefinite matrix, and R(x) is a symmetric positive definite matrix. Moreover, x Q(x)x is a measure of control accuracy and u (x)r(x)u(x) is a measure of control effort. 3.2 Approximate Solution for Finite- Horizon SDRE Regulator To minimize the above cost function (5), a state feedback control law can be given as u(x) = kx(t) = R 1 (x)b (x)p(x)x(t), (6) wherep(x, t) is a symmetric, positive-definite solution of the state-dependent Diffrential Riccati Equation (SDDRE) of the form ISBN: 978-96-474-318-6 38

Ṗ(x) = P(x)A(x)+ A (x)p(x) (7) P(x)B(x)R 1 (x)b (x)p(x) + Q(x), with the final condition F = P(x, t f ). (8) The resulting SDRE-controlled trajectory becomes the solution of the state-dependent closed-loop dynamics ẋ(t) = A(x) B(x)R 1 (x)b (x)p(x)]x(t) (9) As the SDRE function of (x, t), we do not know the value of the states ahead of present time step. Consequently, the state dependent coefficients cannot be calculated to solve (7) with the final condition (8) by backward integration from t f to t. To overcome the problem, an approximate analytical approach is used 4, 5, 6]. which converts the original nonlinear Ricatti equation to a differential Lyapunov equation. At each time step, the Lyapunov equation can be solved in closed form. In order to solve the DRE (7), one can follow the following steps at each time step: Solve Algebraic Riccati Equation (ARE) to calculate the steady state value P ss (x) P ss (x)a(x) + A (x)p ss (x) P ss (x)b(x)r 1 (x)b (x)p ss (x) + Q(x) =. (1) Use changing of variables technique and assume that K(x,t) = P(x,t) P ss (x)] 1. Calculate the value of A cl (x) as A cl (x) = A(x) B(x)R 1 B (x)p ss (x). Solve the algebraic Lyapunov equation 11] A cl D + DA cl BR 1 B =. (11) Solve the differential Lyapunov equation K(x,t) = K(x,t)A cl(x) + A cl (x)k(x,t) (12) B(x)R 1 B (x). The solution of (12), as shown by 12], is given by K(t) = e Acl(t tf) (K(x,t f ) D)e Acl (t t f) + D. (13) Calculate the value of P(x,t) from the equation P(x,t) = K 1 (x,t) + P ss (t). (14) Finally, calculating the value of the optimal control u(x, t) as u(x, t) = R 1 B (x)p(x,t)x. (15) 4 Nonlinear Finite-Horizon Tracking Using SDRE 4.1 Problem Formulation Consider nonlinear system given in (1) and (2),which can be redescibed in the form (3) and (4), Let z(t) be the desired output. The goal is to find a state feedback control law that minimizes a cost function given by : + 1 2 J(x,u) = 1 2 e (t f )Fe(t f ) (16) tf t e (t)q(x)e(t) + u (x)r(x)u(x) ] dt, where e(t) = z(t) y(t). 4.2 Approximate Solution for Finite- Horizon SDDRE Tracking To minimize the above cost function (16), a feedback control law can be given as u(x) = R 1 B (x)p(x)x g(x)], (17) where P(x) is a symmetric, positive-definite solution of the state-dependent Diffrential Riccati Equation (SDDRE) of the form Ṗ(x) = P(x)A(x) + A (x)p(x) (18) P(x)B(x)R 1 B (x)p(x) + C (x)q(x)c(x), with the final condition P(x, t f ) = C (t f )FC(t f ). (19) The resulting SDRE-controlled trajectory becomes the solution of the state-dependent closed-loop dynamics ẋ(t) = A(x) B(x)R 1 (x)b (x)p(x)]x(t) (2) +B(x)R 1 (x)b (x)g(x), where g(x) is a solution of the state-dependent nonhomogeneous vector differential equation ġ(x) = A(x) B(x)R 1 (x)b (x)p(x)] g(x) (21) with the final condition C (x)q(x)z(x), g(x, t f ) = C (t f )Fz(t f ). (22) Simillar to Section 3, an approximate analytical approach is used and the DRE can be solved in the following steps at each time step: ISBN: 978-96-474-318-6 39

Solve for P(x,t) similar to the SDDRE regulator problem in Section 3 Calculate the steady state value g ss (x) from the equation g ss (x) = A(x) B(x)R 1 (x)b (x)p ss (x)] 1 (23) C (x)q(x)z(x). Use changing of variables technique and assume that K g (x,t) = g(x,t) g ss (x)]. Solve the differential equation 5 Illustrative Examples In this section, two illustrative examples, which have analytical solution for finite-horizon SDDRE tracking control design, are given. The first example is linear tracking and the second example is nonlinear tracking. 5.1 Example 1 Consider the second order system 1 (t) = x 2, (t) (27) K g (x, t) = e (A BR 1 B P) (t t f ) g(x, t f ) g ss (x)]. (24) 2 (t) = 2x 1 (t) 3x 2 (t) + u(t), Calculate the value of g(x, t) from the equation g(x,t) = K g (x,t) + g ss (x). (25) Calculate the value of the optimal control u(x, t) as u(x, t) = R 1 (x)b (x)p(x, t)x g(x, t)].(26) Fig.2 summarized the overview of the process of finite-horizon SDDRE tracking technique y(t) = x 1 (t), (28) with the initial conditions x =, ]. (29) The reference output is z(t) = 3, (3) and the selected weighted matrices are Q = diag(1, 1),R =.4, F = diag(2, 2). (31) The simulations were performed for final time of 1 seconds and the resulting output trajectory is shown in Fig. 3, and the optimal control is shown in Fig. 4. In Fig. 3, the solid line denotes the actual output trajectory of the finite-horizon tracking controller, the doted line denotes the reference output signal. It s clear that the algorithm gives very good result as the actual output is making almost ideal tracking to the reference output and it shows that the algorithm is able to solve the state dependent differential Riccati equation (SDDRE) finite-horizon tracking problem with minimum error. 5.2 Example 2 The dynamic equation for forced damped pendulum is: Figure 2: Overview of The Process of Finite-Horizon SDDER Tracking Note : It is easily seen that this technique with finitehorizon SDDRE can be used for linear systems and the resulting SDDRE becomes the standard DRE 1]. ml 2 θ = mglsin(θ) k θ + T, (32) where, θ.. angle of pendulum, l.. length of rod, m.. mass of pendulum, g.. gravitational constant, k.. damping (friction) constant, T.. driving torque. The system nonlinear state equations can be written in the form: ISBN: 978-96-474-318-6 4

Optimal Output 4 3.5 3 2.5 2 1.5 actual output reference output The simulations were performed for final time of 12 seconds and the resulting output trajectory is shown in Fig. 5, and the optimal control is shown in Fig. 6. In Fig. 5, the solid line denotes the actual output trajectory of the finite-horizon tracking controller, the doted line denotes the reference output signal. 1.5 1 2 3 4 5 6 7 8 9 1 Figure 3: Output Trajectories for Example 1 States 1.5 1.5.5 actual output reference output 8 1 6 4 1.5 2 4 6 8 1 12 Optimal Control 2 2 Figure 5: Output Trajectories for Example 2 4 6 12 8 1 2 3 4 5 6 7 8 9 1 Figure 4: Optimal Control for Example 1 Optimal Control 1 8 6 4 2 1 = x 2, (33) 2 = g l sin(x 1) k ml 2x 2 + 1 ml 2u, 2 2 4 6 8 1 12 Figure 6: Optimal Control for Example 2 y = x 1, (34) where: θ = x 1, θ = x 2, T = u. Or alternatively in state dependent form: x1 x 2 ] = 1 4.9sin(x 1 ) x 1.25 Let the reference output is ] x1 x 2 ] +.25 ] u. (35) z(t) = sin(t), (36) and the selected weighted matrices are Q = diag(1,),r =.7, F = diag(1, 1). (37) Comparing these trajectories in Fig. 5, it s clear that the finite-horizon SDDRE nonlinear tracking algorithm gives very good results as the actual optimal output is making very good tracking to the reference output, and the developed algorithm is able to solve the SDDRE finite-horizon nonlinear tracking problem. 6 Conclusion The paper offered a new approximate online technique used for finite-horizon nonlinear tracking problems. This technique based on change of variables that converts the differential Riccati equation to a linear Lyapunov equation. During online implementation, the Lyapunov equation is solved in a closed form at the given time step. Two Illustrative examples are in- ISBN: 978-96-474-318-6 41

cluded to demonstrate the effectiveness of the developed technique. References: 1] Çimen, T. State-dependent Riccati equation (SDRE) control: a survey. Proceedings of the 17th IFAC World Congress, 28, pp. 3761-3775. 2] Cloutier, J. R., State-Dependent Riccati Equation Techniques: An Overview, Proc. American Control Conference, Vol. 2, 1997, pp.932-936. 3] Çimen, T. Development and validation of a mathematical model for control of constrained nonlinear oil tanker motion. Mathematical and Computer Modelling of Dynamical Systems, 15,29, pp. 1749 4] Heydari, A., and Balakrishnan, S. N.,Path Planning Using a Novel Finite-Horizon Suboptimal Controller,Journal of Guidance, Control, and Dynamics, 213, pp.1 5 5] Nguyen, T., and Gajic, Z., Solving the Matrix Differential Riccati Equation: a Lyapunov Equation Approach, IEEE Trans. Automatic Control, Vol. 55, No. 1, 21, pp. 191-194. 6] Nazarzadeh, J., Razzaghi, M., and Nikravesh, K., Solution of the Matrix Riccati Equation for the Linear Quadratic Control Problems, Mathematical and Computer Modelling, Vol. 27, No. 7, 1998, pp. 51-55. 7] J. Doyle, Y. Huang, J. Primbs, R. Freeman, R. Murray, A. Packard, and M. Krstic.Nonlinear control: Comparisons and case studies. In Notes from the Nonlinear Control Workshop conducted at the American Control Conference, Albuquerque, New Mexico, 1998. 8] Cloutier, J. R., & Stansbery, D. T. Nonlinear, hybrid bank-to-turn/skid-to-turn autopilot design. In Proceedings of the AIAA Guidance, Navigation, and Control Conference, 21. 9] Çimen, T. Survey of State-Dependent Riccati Equation in Nonlinear Optimal Feedback Control Synthesis. AIAA Journal of Guidance, Control, and Dynamics, Vol. 35, No. 4, 212, pp. 125 147. 1] D. S. Naidu. Optimal Control Systems (Boca Raton: CRC Press, 23). 11] Z. Gajic and M. Qureshi, The Lyapunov Matrix Equation in System Stability and Control. New York: Dover Publications, 28. 12] Barraud, A., A New Numerical Solution of Xdot=A1*X+X*A2+D, X()=C,IEEE Transaction on Automatic Control, Vol. 22, No.6, Dec. 1977, pp. 976-977. ISBN: 978-96-474-318-6 42