Z i Q ij Z j. J(x, φ; U) = X T φ(t ) 2 h + where h k k, H(t) k k and R(t) r r are nonnegative definite matrices (R(t) is uniformly in t nonsingular).
|
|
- Ariel Marshall
- 5 years ago
- Views:
Transcription
1 2. LINEAR QUADRATIC DETERMINISTIC PROBLEM Notations: For a vector Z, Z = Z, Z is the Euclidean norm here Z, Z = i Z2 i is the inner product; For a vector Z and nonnegative definite matrix Q, Z Q = Z, QZ is L 2 -norm with the Kernel Q: Z, QZ = i,j Z i Q ij Z j. is transposition symbol; grad is symbol of gradient; A controlled process X t is defined by a linear vector differential equation with control action Ut: Ẋ t = atx t + ctut 2.1 subject to the fixed initial condition X = x. Here X t and Ut are vectors of sizes k and r, and at k k, ct k r are matrix valued functions of the argument t. The control Ut has to be chosen such that to track a smooth vector valued function φt, of the size k, in a sense of the minimization of cost functional Jx, φ; U = X T φt 2 h + T X t φt 2 Ht + Ut 2 Rt dt, 2.2 where h k k, Ht k k and Rt r r are nonnegative definite matrices Rt is uniformly in t nonsingular Preliminaries. Set Y t = X t φt. Since φt is assumed to be differentiable, by 2.1 we find the differential equation for Y t Ẏ t = aty t + [ atφt φt ] + ctut 2.3 subject to Y = X φ := y. Notice also that the cost functional Jx, φ : U is transformed to T Jy, φ; U = Y T 2 h + Y t 2 Ht + Ut 2 Rt dt. 2.4 As in Lecture 1, we introduce the Bellman function T V t, y = Y s 2 Hs + Us 2 Rs ds min Y T 2 h + Us:s t T t 1
2 and apply Bellman s principle of optimality: t+δ V t, y = min Us:t s t+δ where Y t = y and for s > t t Y s 2 Hs + Us 2 Rs Ẏ s = asy s + [ asφs φs ] + csus. ds + V t + δ, Y t+δ, 2.5 With a help of 2.5 we derive heuristically the Bellman equation V t, y = min y 2 Ht + U 2 Rt t U [ + grad y V t, y aty + [ atφt φt ] ] ctu subject to boundary condition V T, y = y 2 h The optimal control. We find now [ U t, y = argmin y 2 Ht+ U 2 Rt+grad y V t, y aty+ [ atφy φt ] ] +ctu U Obviously, this procedure is reduced to a minimization in U of the quadratic form QU := U 2 Rt + grad y V t, yctu. Since QU is the quadratic form U solves the equation grad U QU =. Notice that grad U U 2 Rt = 2U Rt, so that Hence, grad U QU = 2U Rt + grad y V t, yct. U t, y = 1. 2 R 1 tc t grad y V t, y 2.7 Then, the Bellman equation is transformed to V t, y = y 2 Ht + 1 grad t 4 y V t, yctr 1 2 Rt + grad y V t, y {aty + [ ]} atφy φt 1 2 grad yv t, yctr 1 yc t grad y V t, y. 2
3 Owing to grad y V t, yctr 1 tc 1 t grad y V t, y = grad 4 y V t, yctr 1 t we find that V t, y = y 2 Ht 1 grad t 4 y V t, yctr 1 2 Rt { + grad y V t, y aty + [ atφt φt ]}. 2 Rt 2.8 As in Lecture 1, we shall find a solution of 2.8 as quadratic form in y with nonnegative definite matrix Γt: V t, y = y Γty + y Bt + Qt; in particular, then grad y V t, y = 2y Γt + B t. Substituting that V t, y and grad y V t, y in 2.8 we arrive at the identity y Γty + y Ḃt + Qt y 2 Ht 1 2y Γt + B t ctr 1 tc t 2y Γt + B t 4 + 2y Γt + B t aty + [ ] atφt φt which provide, with 2y Γtaty y Γtaty + y a tγty, the differential equations Γt = Ht ΓtctR 1 c tγt + Γtat + a tγt a Ḃt = t ΓtctR 1 c t Bt + 2Γt [ atφt φt ] Qt = 1 4 B tctr 1 tc tbt 2.9 subject to the boundary conditions ΓT = h, QT = and QT =. Hence, by 2.7, U t, y = R 1 tc t Γty + Bt. 2.1 Thus, the optimal control for the original problem is defined as follows U t, Xt = R 1 tc t Γt { Xt φt } + Bt where Ẋ t = atx t + ctu t, X t. Remark. If φt atφt, then Bt and Qt. 3
4 2.3. Infinite horizon. From application point of view, it makes sense to analyze a case of very large time T, that is T =. To simplify an analysis, we assume that all matrices are time-independent, i.e. φt, and Jx, U = Ẋ t = ax t + cut, 2.11 X t 2 H + Ut 2 R dt The main problem for this setting is a requirement that there exists a control action Ut for which Jx, U <. We give conditions, guaranteeing the latter, expressed in terms of matrices a, c and H. Recall that matrix R is nonsingular. With T k, we consider a family of cost functionals J Tk x, U k 1. For every k, we have Tk J Tk x, U = X t 2 H + Ut 2 R dt. For fixed k, denote the optimal control by U k t. From the obtained above result we know that by Remark Bt, Qt U k t = R 1 c Γ k t, where Γ k t solves the Riccati equation Γ k t = a Γ k t + Γ k a + H Γ k tcr 1 c Γ k subject to the boundary condition Γ k T =. Moreover, min J Tk x, U = x 2 Γ Us: s T k. k We show that x 2 is an increasing sequence. Γ k k 1 Denote by Xt k the controlled process associated with optimal control U k t on [, T k ]. Then Tk x 2 Γ k = Xt k 2 H + U k t R 2 dt and so x 2 Γ k+1 = Tk+1 Tk Tk X k+1 t 2 H + U k+1 t 2 R dt X k+1 t 2 H + U k+1 t 2 R dt since T k+1 > T k X k t 2 H + U k t 2 R dt since U k is optimal = x 2 Γ k. 4
5 Consequently, lim k x 2 exists but we can not be sure that this limit is finite. Γ k To prove lim k x 2 Γ k <, notice that it suffices to choose some control action Ũt such that Jx, Ũ <. Then, whereas x 2 J Γ k T k x, Ũ Jx, Ũ, for any k, we get the desired property. Assuming the matrix 1 A = e as cc e a s ds is nonsingular. Then, taking { c Ũt = e a t 1x e as cc e ds a s t 1, t > 1, where x = X and Xt = a X t + cũt, we find X 1 = e a x + = e a [I =. e at cũtdt e at cc e a t dt 1 ] e at cc e a t dt x The latter allows us to conclude that X t, t 1, and at the same time Jx, Ũ = J 1x, Ũ <. So, it remains to prove that A is nonsingular matrix. Theorem. The matrix A is nonsingular, if the block-matrix of size k kr is of the full rank k Ga, c = c ac... a k 1 c Proof. Assume A is singular. Since A is nonnegative definite, there is vector Z so that = Z e at cc e a t dsz = Z e at cc e a t Zdt. On the other hand, Z e at cc e a t Z, being nonnegative and continuous in t, is equal zero for all t [, 1]. Hence, Z e at c. Differentiating this identity in t and letting 5
6 t =, we get z e at c t= = z c = d z e at c = z ac = dt t= d j dt j z e at c t= = z a j c = Consequently, Z Ga, cg a, cz = k 1 j= Z ca j a j c Z =, so that the matrix Ga, cg a, c is singular. The contradiction obtained, finishing the proof. Definition. The pair of matrices a, c is said to be controllable, if the matrix Ga, c is of rank k. Thus, if a, c is controllable, lim k x 2 exists for any x and is finite. Then, Γ k obviously, lim k Γ k exists as well and denote this limit by Γ. Since Γ k t = a Γ k t + Γ k ta + H Γ k tcr 1 c Γ k t subject to Γ k T =, the limit Γ solves the algebraic Riccati equation a Γ + Γa + H ΓcR 1 c Γ = The next fact we use without proof is that 2.15 possesses the unique positive definite solution provided that the matrix H Hc gh, c =. Hc k 1 is of the full rank k. In particular, it holds if H is nonsingular. Definition. The pair of matrices H, c is said to be observable, if the matrix gh, c is of rank k The optimal control for infinite horizon. Since xγ k 2 J Tk x, U Jx, U for any k, and lim k x 2 Γ k = xγ 2, we find the lower bound The optimal control regarding to [, T k ] is Jx, U x 2 Γ U t, X t = R 1 tc tγ k tx t. 6
7 Hence a control U t, X t = R 1 tc tγx t. is a candidate to be the optimal control for the infinite horizon. The check that supposition, let us consider a quadratic form X t 2 Γ. Write d dt X t 2 Γ = Ẋ t ΓXt + Xt ΓẊ t = X t a + U X t c ΓXt + Xt Γ axt + cu Xt ] [ = Xt [a ΓcR 1 c ΓXt + Xt Γ a cr 1 c Γ ]X t = Xt a Γ + Γa 2P cr 1 c Γ Xt = Xt H + ΓcR 1 c Γ Xt Hence, we find that for any T > = X t HX t X t ΓcR 1 RR 1 c ΓX t = X t HX t U X t RU X t. J T x, U = x 2 Γ X T 2 Γ x 2 Γ. Thereby, Jx, U = lim T J T x, U x 2 Γ. So, by 2.16 we have Jx, U = x 2 Γ, that is U X t is the optimal control. 1. Home work: Vector case. Feedback control. Bellman equation Let X t, t =, 1,..., N be the controlled vector sequence sequence defined by linear vector-matrix recursion X t+1 = ax t + cut, where U t is the control. X t and U t are vectors of sizes kand r respectively. a = a k k and c = c k r are known matrices. X = x and x is known. It is required to choose the optimal control U t minimizing the cost functional h k k, H k k are nonnegative definite matrices, R r r is positive definite matrix T 1 Jx, U = X T 2 h + X t 2 H + Ut 2 R Derive Bellman s equation for the Bellman function [ T 1 V x, t = X T 2 h + min U s:t s T 2. Find the optimal control. t=1 7 s=t X s 2 H + Us 2 R ]
OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationControl, Stabilization and Numerics for Partial Differential Equations
Paris-Sud, Orsay, December 06 Control, Stabilization and Numerics for Partial Differential Equations Enrique Zuazua Universidad Autónoma 28049 Madrid, Spain enrique.zuazua@uam.es http://www.uam.es/enrique.zuazua
More informationOn Controllability of Linear Systems 1
On Controllability of Linear Systems 1 M.T.Nair Department of Mathematics, IIT Madras Abstract In this article we discuss some issues related to the observability and controllability of linear systems.
More informationChapter III. Stability of Linear Systems
1 Chapter III Stability of Linear Systems 1. Stability and state transition matrix 2. Time-varying (non-autonomous) systems 3. Time-invariant systems 1 STABILITY AND STATE TRANSITION MATRIX 2 In this chapter,
More informationLecture 5 Linear Quadratic Stochastic Control
EE363 Winter 2008-09 Lecture 5 Linear Quadratic Stochastic Control linear-quadratic stochastic control problem solution via dynamic programming 5 1 Linear stochastic system linear dynamical system, over
More informationA note on linear differential equations with periodic coefficients.
A note on linear differential equations with periodic coefficients. Maite Grau (1) and Daniel Peralta-Salas (2) (1) Departament de Matemàtica. Universitat de Lleida. Avda. Jaume II, 69. 251 Lleida, Spain.
More information6.241 Dynamic Systems and Control
6.241 Dynamic Systems and Control Lecture 24: H2 Synthesis Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology May 4, 2011 E. Frazzoli (MIT) Lecture 24: H 2 Synthesis May
More information1. Find the solution of the following uncontrolled linear system. 2 α 1 1
Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +
More informationApplied Math Qualifying Exam 11 October Instructions: Work 2 out of 3 problems in each of the 3 parts for a total of 6 problems.
Printed Name: Signature: Applied Math Qualifying Exam 11 October 2014 Instructions: Work 2 out of 3 problems in each of the 3 parts for a total of 6 problems. 2 Part 1 (1) Let Ω be an open subset of R
More informationGeneralized Riccati Equations Arising in Stochastic Games
Generalized Riccati Equations Arising in Stochastic Games Michael McAsey a a Department of Mathematics Bradley University Peoria IL 61625 USA mcasey@bradley.edu Libin Mou b b Department of Mathematics
More informationEE363 homework 2 solutions
EE363 Prof. S. Boyd EE363 homework 2 solutions. Derivative of matrix inverse. Suppose that X : R R n n, and that X(t is invertible. Show that ( d d dt X(t = X(t dt X(t X(t. Hint: differentiate X(tX(t =
More informationMath 273, Final Exam Solutions
Math 273, Final Exam Solutions 1. Find the solution of the differential equation y = y +e x that satisfies the condition y(x) 0 as x +. SOLUTION: y = y H + y P where y H = ce x is a solution of the homogeneous
More informationLECTURE 10: THE PARALLEL TRANSPORT
LECTURE 10: THE PARALLEL TRANSPORT 1. The parallel transport We shall start with the geometric meaning of linear connections. Suppose M is a smooth manifold with a linear connection. Let γ : [a, b] M be
More informationNonlinear Observers. Jaime A. Moreno. Eléctrica y Computación Instituto de Ingeniería Universidad Nacional Autónoma de México
Nonlinear Observers Jaime A. Moreno JMorenoP@ii.unam.mx Eléctrica y Computación Instituto de Ingeniería Universidad Nacional Autónoma de México XVI Congreso Latinoamericano de Control Automático October
More informationChapter 3. LQ, LQG and Control System Design. Dutch Institute of Systems and Control
Chapter 3 LQ, LQG and Control System H 2 Design Overview LQ optimization state feedback LQG optimization output feedback H 2 optimization non-stochastic version of LQG Application to feedback system design
More informationState Variable Analysis of Linear Dynamical Systems
Chapter 6 State Variable Analysis of Linear Dynamical Systems 6 Preliminaries In state variable approach, a system is represented completely by a set of differential equations that govern the evolution
More informationM. S. N. Murty, B. V. Appa Rao, and G. Suresh Kumar
Bull. Korean Math. Soc. 43 (2006), No. 1, pp. 149 159 CONTROLLABILITY, OBSERVABILITY, AND REALIZABILITY OF MATRIX LYAPUNOV SYSTEMS M. S. N. Murty, B. V. Appa Rao, and G. Suresh Kumar Abstract. This paper
More informationLS.5 Theory of Linear Systems
LS.5 Theory of Linear Systems 1. General linear ODE systems and independent solutions. We have studied the homogeneous system of ODE s with constant coefficients, (1) x = Ax, where A is an n n matrix of
More informationThe goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T
1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationLinear System Theory
Linear System Theory Wonhee Kim Lecture 3 Mar. 21, 2017 1 / 38 Overview Recap Nonlinear systems: existence and uniqueness of a solution of differential equations Preliminaries Fields and Vector Spaces
More informationLinear-quadratic control problem with a linear term on semiinfinite interval: theory and applications
Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications L. Faybusovich T. Mouktonglang Department of Mathematics, University of Notre Dame, Notre Dame, IN
More informationECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming
ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear
More information2 Statement of the problem and assumptions
Mathematical Notes, 25, vol. 78, no. 4, pp. 466 48. Existence Theorem for Optimal Control Problems on an Infinite Time Interval A.V. Dmitruk and N.V. Kuz kina We consider an optimal control problem on
More informationJordan Canonical Form
October 26, 2005 15-1 Jdan Canonical Fm Suppose that A is an n n matrix with characteristic polynomial. p(λ) = (λ λ 1 ) m 1... (λ λ s ) ms and generalized eigenspaces V j = ker(a λ j I) m j. Let L be the
More informationDifferential Equations Preliminary Examination
Differential Equations Preliminary Examination Department of Mathematics University of Utah Salt Lake City, Utah 84112 August 2007 Instructions This examination consists of two parts, called Part A and
More informationOn the Stability of the Best Reply Map for Noncooperative Differential Games
On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University
More informationCourse Summary Math 211
Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.
More informationRobotics. Control Theory. Marc Toussaint U Stuttgart
Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,
More informationLecture 4 Continuous time linear quadratic regulator
EE363 Winter 2008-09 Lecture 4 Continuous time linear quadratic regulator continuous-time LQR problem dynamic programming solution Hamiltonian system and two point boundary value problem infinite horizon
More information1 Lyapunov theory of stability
M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability
More informationExamples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling
1 Introduction Many natural processes can be viewed as dynamical systems, where the system is represented by a set of state variables and its evolution governed by a set of differential equations. Examples
More informationConjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
More informationFirst and Second Order Differential Equations Lecture 4
First and Second Order Differential Equations Lecture 4 Dibyajyoti Deb 4.1. Outline of Lecture The Existence and the Uniqueness Theorem Homogeneous Equations with Constant Coefficients 4.2. The Existence
More informationNonlinear Control Systems
Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f
More informationOn perturbations in the leading coefficient matrix of time-varying index-1 DAEs
On perturbations in the leading coefficient matrix of time-varying index-1 DAEs Institute of Mathematics, Ilmenau University of Technology Darmstadt, March 27, 2012 Institute of Mathematics, Ilmenau University
More informationStability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games
Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,
More informationME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ
ME 234, Lyapunov and Riccati Problems. This problem is to recall some facts and formulae you already know. (a) Let A and B be matrices of appropriate dimension. Show that (A, B) is controllable if and
More informationModeling and Analysis of Dynamic Systems
Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability
More informationKalman Filter and Parameter Identification. Florian Herzog
Kalman Filter and Parameter Identification Florian Herzog 2013 Continuous-time Kalman Filter In this chapter, we shall use stochastic processes with independent increments w 1 (.) and w 2 (.) at the input
More informationEE363 Review Session 1: LQR, Controllability and Observability
EE363 Review Session : LQR, Controllability and Observability In this review session we ll work through a variation on LQR in which we add an input smoothness cost, in addition to the usual penalties on
More informationTopic # Feedback Control Systems
Topic #17 16.31 Feedback Control Systems Deterministic LQR Optimal control and the Riccati equation Weight Selection Fall 2007 16.31 17 1 Linear Quadratic Regulator (LQR) Have seen the solutions to the
More informationChap. 3. Controlled Systems, Controllability
Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :
More informationChapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.
Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space
More informationConstrained continuous-time Markov decision processes on the finite horizon
Constrained continuous-time Markov decision processes on the finite horizon Xianping Guo Sun Yat-Sen University, Guangzhou Email: mcsgxp@mail.sysu.edu.cn 16-19 July, 2018, Chengdou Outline The optimal
More informationLecture 9. Systems of Two First Order Linear ODEs
Math 245 - Mathematics of Physics and Engineering I Lecture 9. Systems of Two First Order Linear ODEs January 30, 2012 Konstantin Zuev (USC) Math 245, Lecture 9 January 30, 2012 1 / 15 Agenda General Form
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationState Regulator. Advanced Control. design of controllers using pole placement and LQ design rules
Advanced Control State Regulator Scope design of controllers using pole placement and LQ design rules Keywords pole placement, optimal control, LQ regulator, weighting matrixes Prerequisites Contact state
More informationAdvanced Mechatronics Engineering
Advanced Mechatronics Engineering German University in Cairo 21 December, 2013 Outline Necessary conditions for optimal input Example Linear regulator problem Example Necessary conditions for optimal input
More informationMODULE 8 Topics: Null space, range, column space, row space and rank of a matrix
MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix Definition: Let L : V 1 V 2 be a linear operator. The null space N (L) of L is the subspace of V 1 defined by N (L) = {x
More informationLinear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,
More informationLecture 8: Linear Algebra Background
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected
More informationLecture 2 and 3: Controllability of DT-LTI systems
1 Lecture 2 and 3: Controllability of DT-LTI systems Spring 2013 - EE 194, Advanced Control (Prof Khan) January 23 (Wed) and 28 (Mon), 2013 I LTI SYSTEMS Recall that continuous-time LTI systems can be
More informationQuadratic Stability of Dynamical Systems. Raktim Bhattacharya Aerospace Engineering, Texas A&M University
.. Quadratic Stability of Dynamical Systems Raktim Bhattacharya Aerospace Engineering, Texas A&M University Quadratic Lyapunov Functions Quadratic Stability Dynamical system is quadratically stable if
More informationLecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?
KTH ROYAL INSTITUTE OF TECHNOLOGY Norms for vectors and matrices Why? Lecture 5 Ch. 5, Norms for vectors and matrices Emil Björnson/Magnus Jansson/Mats Bengtsson April 27, 2016 Problem: Measure size of
More informationMATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem
MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear
More informationNonlinear Systems Theory
Nonlinear Systems Theory Matthew M. Peet Arizona State University Lecture 2: Nonlinear Systems Theory Overview Our next goal is to extend LMI s and optimization to nonlinear systems analysis. Today we
More informationCONTROL DESIGN FOR SET POINT TRACKING
Chapter 5 CONTROL DESIGN FOR SET POINT TRACKING In this chapter, we extend the pole placement, observer-based output feedback design to solve tracking problems. By tracking we mean that the output is commanded
More information1. The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ = Ax + Bu is
ECE 55, Fall 2007 Problem Set #4 Solution The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ Ax + Bu is x(t) e A(t ) x( ) + e A(t τ) Bu(τ)dτ () This formula is extremely important
More informationNonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1
Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems p. 1/1 p. 2/1 Converse Lyapunov Theorem Exponential Stability Let x = 0 be an exponentially stable equilibrium
More informationProblem List MATH 5173 Spring, 2014
Problem List MATH 5173 Spring, 2014 The notation p/n means the problem with number n on page p of Perko. 1. 5/3 [Due Wednesday, January 15] 2. 6/5 and describe the relationship of the phase portraits [Due
More informationECE504: Lecture 8. D. Richard Brown III. Worcester Polytechnic Institute. 28-Oct-2008
ECE504: Lecture 8 D. Richard Brown III Worcester Polytechnic Institute 28-Oct-2008 Worcester Polytechnic Institute D. Richard Brown III 28-Oct-2008 1 / 30 Lecture 8 Major Topics ECE504: Lecture 8 We are
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationFundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved
Fundamentals of Linear Algebra Marcel B. Finan Arkansas Tech University c All Rights Reserved 2 PREFACE Linear algebra has evolved as a branch of mathematics with wide range of applications to the natural
More informationThe Inverse Function Theorem 1
John Nachbar Washington University April 11, 2014 1 Overview. The Inverse Function Theorem 1 If a function f : R R is C 1 and if its derivative is strictly positive at some x R, then, by continuity of
More information1 The Observability Canonical Form
NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)
More informationTheorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0).
Linear Systems Notes Lecture Proposition. A M n (R) is positive definite iff all nested minors are greater than or equal to zero. n Proof. ( ): Positive definite iff λ i >. Let det(a) = λj and H = {x D
More informationLECTURE 16: CONJUGATE AND CUT POINTS
LECTURE 16: CONJUGATE AND CUT POINTS 1. Conjugate Points Let (M, g) be Riemannian and γ : [a, b] M a geodesic. Then by definition, exp p ((t a) γ(a)) = γ(t). We know that exp p is a diffeomorphism near
More informationNonlinear Control Lecture 5: Stability Analysis II
Nonlinear Control Lecture 5: Stability Analysis II Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Fall 2010 Farzaneh Abdollahi Nonlinear Control Lecture 5 1/41
More informationLinear Algebra. Preliminary Lecture Notes
Linear Algebra Preliminary Lecture Notes Adolfo J. Rumbos c Draft date May 9, 29 2 Contents 1 Motivation for the course 5 2 Euclidean n dimensional Space 7 2.1 Definition of n Dimensional Euclidean Space...........
More informationA Second Course in Elementary Differential Equations
A Second Course in Elementary Differential Equations Marcel B Finan Arkansas Tech University c All Rights Reserved August 3, 23 Contents 28 Calculus of Matrix-Valued Functions of a Real Variable 4 29 nth
More informationLOCALLY POSITIVE NONLINEAR SYSTEMS
Int. J. Appl. Math. Comput. Sci. 3 Vol. 3 No. 4 55 59 LOCALLY POSITIVE NONLINEAR SYSTEMS TADEUSZ KACZOREK Institute of Control Industrial Electronics Warsaw University of Technology ul. Koszykowa 75 66
More informationLinear ODEs. Existence of solutions to linear IVPs. Resolvent matrix. Autonomous linear systems
Linear ODEs p. 1 Linear ODEs Existence of solutions to linear IVPs Resolvent matrix Autonomous linear systems Linear ODEs Definition (Linear ODE) A linear ODE is a differential equation taking the form
More informationProblem 1 Cost of an Infinite Horizon LQR
THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 5 Ahmad F. Taha October 12, 215 Homework Instructions: 1. Type your solutions in the LATEX homework
More informationHankel Optimal Model Reduction 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Hankel Optimal Model Reduction 1 This lecture covers both the theory and
More informationChap 4. State-Space Solutions and
Chap 4. State-Space Solutions and Realizations Outlines 1. Introduction 2. Solution of LTI State Equation 3. Equivalent State Equations 4. Realizations 5. Solution of Linear Time-Varying (LTV) Equations
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationLMI Methods in Optimal and Robust Control
LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 15: Nonlinear Systems and Lyapunov Functions Overview Our next goal is to extend LMI s and optimization to nonlinear
More informationTime-Invariant Linear Quadratic Regulators!
Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 17 Asymptotic approach from time-varying to constant gains Elimination of cross weighting
More informationPontryagin s maximum principle
Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin
More informationLinear Algebra. Preliminary Lecture Notes
Linear Algebra Preliminary Lecture Notes Adolfo J. Rumbos c Draft date April 29, 23 2 Contents Motivation for the course 5 2 Euclidean n dimensional Space 7 2. Definition of n Dimensional Euclidean Space...........
More informationECSE.6440 MIDTERM EXAM Solution Optimal Control. Assigned: February 26, 2004 Due: 12:00 pm, March 4, 2004
ECSE.6440 MIDTERM EXAM Solution Optimal Control Assigned: February 26, 2004 Due: 12:00 pm, March 4, 2004 This is a take home exam. It is essential to SHOW ALL STEPS IN YOUR WORK. In NO circumstance is
More informationLecture 2: Discrete-time Linear Quadratic Optimal Control
ME 33, U Berkeley, Spring 04 Xu hen Lecture : Discrete-time Linear Quadratic Optimal ontrol Big picture Example onvergence of finite-time LQ solutions Big picture previously: dynamic programming and finite-horizon
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More information6. Linear Quadratic Regulator Control
EE635 - Control System Theory 6. Linear Quadratic Regulator Control Jitkomut Songsiri algebraic Riccati Equation (ARE) infinite-time LQR (continuous) Hamiltonian matrix gain margin of LQR 6-1 Algebraic
More information16. Local theory of regular singular points and applications
16. Local theory of regular singular points and applications 265 16. Local theory of regular singular points and applications In this section we consider linear systems defined by the germs of meromorphic
More informationA note on the monotonicity of matrix Riccati equations
DIMACS Technical Report 2004-36 July 2004 A note on the monotonicity of matrix Riccati equations by Patrick De Leenheer 1,2 Eduardo D. Sontag 3,4 1 DIMACS Postdoctoral Fellow, email: leenheer@math.rutgers.edu
More informationLecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.
Lecture 4 Chapter 4: Lyapunov Stability Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 4 p. 1/86 Autonomous Systems Consider the autonomous system ẋ
More informationSecond order forward-backward dynamical systems for monotone inclusion problems
Second order forward-backward dynamical systems for monotone inclusion problems Radu Ioan Boţ Ernö Robert Csetnek March 6, 25 Abstract. We begin by considering second order dynamical systems of the from
More informationSwitching, sparse and averaged control
Switching, sparse and averaged control Enrique Zuazua Ikerbasque & BCAM Basque Center for Applied Mathematics Bilbao - Basque Country- Spain zuazua@bcamath.org http://www.bcamath.org/zuazua/ WG-BCAM, February
More informationStochastic and Adaptive Optimal Control
Stochastic and Adaptive Optimal Control Robert Stengel Optimal Control and Estimation, MAE 546 Princeton University, 2018! Nonlinear systems with random inputs and perfect measurements! Stochastic neighboring-optimal
More informationPrinciples of Optimal Control Spring 2008
MIT OpenCourseWare http://ocw.it.edu 16.323 Principles of Optial Control Spring 2008 For inforation about citing these aterials or our Ters of Use, visit: http://ocw.it.edu/ters. 16.323 Lecture 10 Singular
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77
1/77 ECEN 605 LINEAR SYSTEMS Lecture 7 Solution of State Equations Solution of State Space Equations Recall from the previous Lecture note, for a system: ẋ(t) = A x(t) + B u(t) y(t) = C x(t) + D u(t),
More informationGame Theory Extra Lecture 1 (BoB)
Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,
More informationAN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES
AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim
More informationObserver design for a general class of triangular systems
1st International Symposium on Mathematical Theory of Networks and Systems July 7-11, 014. Observer design for a general class of triangular systems Dimitris Boskos 1 John Tsinias Abstract The paper deals
More informationSome lecture notes for Math 6050E: PDEs, Fall 2016
Some lecture notes for Math 65E: PDEs, Fall 216 Tianling Jin December 1, 216 1 Variational methods We discuss an example of the use of variational methods in obtaining existence of solutions. Theorem 1.1.
More information