Notes for ENEE 664: Optimal Control. André L. Tits

Size: px
Start display at page:

Download "Notes for ENEE 664: Optimal Control. André L. Tits"

Transcription

1 Notes for ENEE 664: Optimal Control André L. Tits DRAFT August 2008

2

3 Contents 1 Motivation and Scope Some Examples Scope of the Course Linear Quadratic Optimal Control Fixed end point problem Free end point problem Infinite horizon LTI problem More general optimal control problems Unconstrained Optimization First order condition of optimality Steepest descent method Introduction to convergence analysis Minimization of convex functions Second order optimality conditions Conjugate direction methods Rates of convergence Newton s method Variable metric methods Constrained Optimization Abstract Constraint Set Equality Constraints - First Order Conditions Equality Constraints Second Order Conditions Inequality Constraints First Order Conditions Mixed Constraints First Order Conditions Mixed Constraints Second order Conditions Glance at Numerical Methods for Constrained Problems Sensitivity Duality Linear programming Copyright c , André L. Tits. All Rights Reserved 1

4 CONTENTS 5 Dynamic Optimization Introduction to the calculus of variations Discrete-Time Optimal Control Continuous-Time Optimal Control - Linear Systems Continuous-Time Optimal Control; Nonlinear Systems Dynamic programming A Generalities on Vector Spaces 141 B On Differentiability and Convexity 155 B.1 Differentiability B.2 Some elements of convex analysis B.3 Acknowledgement Copyright c , André L. Tits. All Rights Reserved

5 Chapter 1 Motivation and Scope 1.1 Some Examples We give some examples of design problems in engineering that can be formulated as mathematical optimization problems. Although we emphasize here engineering design, optimization is widely used in other fields such as economics or operations research. Such examples can be found, e.g., in [15]. Example 1.1 Design of an operational amplifier (opamp) Suppose the following features (specifications) are desired 1. a large gain-bandwidth product 2. sufficient stability 3. low power dissipation In this course, we deal with parametric optimization. This means, for this example, that we assume the topology of the circuit has already been chosen, the only freedom left being the choice of the value of a number of design parameters (resistors, capacitors, various transistor parameters). In real world, once the parametric optimization has been performed, the designer will possibly decide to modify the topology of his circuit, hoping to be able to achieve better performances. Another parametric optimization is then performed. This loop may be repeated many times. To formulate the opamp design problem as an optimization problem, one has to specify one (possibly several) objective function(s) and various constraints. We decide for the following goal: minimize subject to the power dissipated gain-bandwidth product ق M 1 (given) frequency response ق M 2 at all frequencies. The last constraint will prevent two high a peaking in the frequency response, thereby ensuring sufficient closed-loop stability margin. We now denote by x the vector of design parameters Copyright c , André L. Tits. All Rights Reserved 3

6 Motivation and Scope x = (R 1, R 2,...,C 1, C 2,..., α i,...) R n For any given x, the circuit is now entirely specified and the various quantities mentioned above can be computed. More precisely, we can define P(x) = power dissipated GB(x) = gain-bandwidth product FR(x, ω) = frequency response, as a function of the frequency ω. We then write the optimization problem as min»p(x) GB(x) ق M 1, FR(x, ω) ق M 2 ω Ω (1.1) where Ω = [ω 1, ω 2 ] is a range of critical frequencies. To obtain a canonical form, we now define and we obtain f(x) := P(x) (1.2) g(x) := M 1 GB(x) (1.3) φ(x, ω) := FR(x, ω) M 2 (1.4) min»f(x) g(x) ق 0, φ(x, ω) ق 0 ω Ω (1.5) Note. We will systematically use notations such as min»f(x) g(x) ق 0 not to just indicate the minimum value, but rather as a short-hand for minimize f(x) subject to g(x) ق 0 More generally, one would have min»f(x) g i (x) ق 0, i = 1, 2,..., m, φ i (x, ω) ق 0, ω Ω i, i = 1,..., k. (1.6) If we define g : R n R m by g(x) = g 1 (x). g m (x) (1.7) 4 Copyright c , André L. Tits. All Rights Reserved

7 1.1 Some Examples and, assuming that all the Ω i s are identical, if we define φ : R n Ω R k by we obtain again φ(x, ω) = φ 1 (x, ω). φ k (x, ω) (1.8) min»f(x) g(x) ق 0, φ(x, ω) ق 0 ω Ω (1.9) [This is called a semi-infinite optimization problem: finitely many variables, infinitely many constraints.] Note. If we define (1.9) is equivalent to ψ i (x) = sup φ i (x, ω) (1.10) ω Ω min»f(x) g(x) ق 0, ψ(x) ق 0 (1.11) (more precisely,»x φ(x, ω) ق 0 ω Ω =»x ψ(x) ق 0 ) and ψ(x) can be absorbed into g(x). This transformation may not be advisable, for the following reasons: (i) some potentially useful information (e.g., what are the critical values of ω) is lost when replacing (1.9) by (1.11) (ii) for given x, ψ(x) may not be computable exactly in finite time (this computation involves another optimization problem) (iii) ψ may not be smooth even when φ is, as shown in the exercise below. Thus (1.11) may not be solvable by classical methods. Exercise 1.1 Prove the equivalence between (1.9) and (1.11). (To prove A = B, prove A B and B A.) Exercise 1.2 Suppose that φ: R n Ω R is continuous and that Ω is compact, so that the sup in (1.10) can be written as a max. Show that ψ is continuous. Exercise 1.3 Show, by exhibiting counterexamples, that (i) (compactness of Ω and) continuity of φ in each variable separately does not imply continuity of ψ (even though the sup in (1.10) is achieved under such assumptions) and that (ii) continuity of ψ does not follow either if the compactness assumption is dropped (even if the sup is achieved). Copyright c , André L. Tits. All Rights Reserved 5

8 Motivation and Scope Exercise 1.4 Exhibit an example where φ C (all derivatives exist and are continuous), where Ω is compact, but where ψ is not everywhere differentiable. However, in this course, we will limit ourselves mostly to classical (non semi-infinite) problems (and will generally assume continuous differentiability), i.e., to problems of the form min»f(x) g(x) ق 0, h(x) = 0 where f : R n R, g : R n R m, h : R n R l, for some positive integers n, m and l. Remark 1.1 To fit the opamp design problem into formulation (1.11) we had to pick one of the design specifications as objective (to be minimized). Intuitively more appealing would be some kind of multiobjective optimization problem. Example 1.2 Design of a p.i.d controller (proportional - integral - derivative) The scalar plant G(s) is to be controlled by a p.i.d. controller (see Figure 1.1). Again, the structure of the controller has already been chosen; only the values of three parameters have to be determined (x = [x 1, x 2, x 3 ] T ). R(s) + E(x,s) x Y(x,s) x + x s s G(s) T(x,s) Figure 1.1: Suppose the specifications are as follows low value of the ISE for a step input (ISE = integral of the square of the difference (error) between input and output, in time domain) enough stability short rise time, settling time, low overshoot We decide to minimize the ISE, while keeping the Nyquist plot of T(x, s) outside some forbidden region (see Figure 1.2) and keeping rise time, settling time, and overshoot under given values. The following constraints are also specified. 10 ق x 1 ق 10, 10 ق x 2 ق 10,.1 ق x 3 ق 10 Exercise 1.5 Put the p.i.d. problem in the form (1.6), i.e., specify f, g i, φ i, Ω i. 6 Copyright c , André L. Tits. All Rights Reserved

9 1.1 Some Examples parabola y=(a+1)x 2- b u(t) (-1,0) w=0 1 y(t) T(x,jw) w=w ι(t) Tr Ts T Figure 1.2: T(x, jω) has to stay outside the forbidden region ω [0, ω] For a step input, y(x, t) is desired to remain between l(t) and u(t) for t [0, T] Example 1.3 Consider again a plant, possibly nonlinear and time varying, and suppose we want to determine the best control u(t) to approach a desired response. ẋ = F(x, u, t) y = G(x, t) We may want to determine (مم) u to minimize the integral J(u) = T 0 (y u (t) v(t)) 2 dt where y u (t) is the output corresponding to control (مم) u and (مم) v is some reference signal. Various features may have to be taken into account: Constraints on (مم) u (for realizability) piecewise continuous u(t) ق U t T may be finite or infinite x(0), x(t) may be free, fixed, constrained The entire state trajectory may be constrained,((مم) x ) e.g., to keep the temperature reasonable One may require a closed loop control, e.g., u(t) = u(x(t)). It is well known that such feedback control systems are much less sensitive to perturbations and modeling errors. Copyright c , André L. Tits. All Rights Reserved 7

10 Motivation and Scope Unlike Example 1.1 and Example 1.2, Example 1.3 is an optimal control problem. Whereas discrete-time optimal control problems can be solved by classical optimization techniques, continuous-time problems involve optimization in infinite dimension spaces (a complete waveform has to be determined). To conclude this section we now introduce the class of problems that will be studied in this course. Consider the abstract optimization problem (P) min»f(x) x S where S is a subset of a vector space X and where f : X R is the cost or objective function. S is the feasible set. Any x in S is a feasible point. Definition 1.1 A point ˆx is called a (strict) global minimizer for (P) if ˆx S and f(ˆx) ق f(x) x S (<) ( x S, x ˆx) Assume now X is equipped with a norm. Definition 1.2 A point ˆx is called a (strict) local minimizer for (P) if ˆx S and ǫ > 0 such that f(ˆx) ق f(x) x S B(ˆx, ǫ) (<) ( x S B(ˆx, ǫ), x ˆx) 8 Copyright c , André L. Tits. All Rights Reserved

11 1.2 Scope of the Course 1.2 Scope of the Course 1. Type of optimization problems considered (i) Finite dimensional unconstrained equality constrained inequality [and equality] constrained linear, quadratic programs, convex problems multiobjective problems discrete optimal control (ii) Infinite dimensional calculus of variations (no control signal) (old: 1800) optimal control (new: 1950 s) Note: most types in (i) can be present in (ii) as well. 2. Results sought Essentially, solve the problem. The steps are conditions of optimality ( simpler characterization of solutions) numerical methods: solve the problem, generally by solving some optimality condition or, at least, using the insight such conditions provide. sensitivity: how good is the solutions in the sense of what if we didn t solve exactly the right problem? ( duality: some transformation of the original problem into a hopefully simpler optimization problem) Copyright c , André L. Tits. All Rights Reserved 9

12 Motivation and Scope 10 Copyright c , André L. Tits. All Rights Reserved

13 Chapter 2 Linear Quadratic Optimal Control References: [21, 1]. Background material for this chapter is given in Appendix A. Consider the linear control system ẋ(t) = A(t)x(t) + B(t)u(t) x(t 0 ) = x 0 (2.1) where x(t) R n, u(t) R m and (مم) A and (مم) B are matrix valued functions. Suppose,(مم) A (مم) B and (مم) u are continuous. Then (2.1) has the unique solution x(t) = Φ(t, t 0 )x 0 + t t 0 Φ(t, σ)b(σ)u(σ)dσ where the state transition matrix Φ satisfies the homogeneous differential equation with initial condition t Φ(t, t 0) = A(t)Φ(t, t 0 ) Φ(t 0, t 0 ) = I. Finally, for any t 1, t 2, the transition matrix Φ(t 1, t 2 ) is invertible and Φ(t 1, t 2 ) 1 = Φ(t 2, t 1 ). 2.1 Fixed end point problem Let t 1 > t 0. Suppose we wish to use the control function (مم) u : [t 0, t 1 ] R m in a certain class U of sufficiently regular functions. For simplicity, let U = C[t 0, t 1 ] m, the class of continuous such functions. Question: Given x 1 R n, does there exist u U such that, for system (2.1), x(t 1 ) = x 1? If the answer to the above is yes, we say that x 1 is reachable from (x 0, t 0 ) at time t 1. If moreover this holds for all x 0, x 1 R n then we say that the system (2.1) is reachable on [t 0, t 1 ]. Copyright c , André L. Tits. All Rights Reserved 11

14 Linear Quadratic Optimal Control There is no loss of generality in assuming that x 1 = θ, as shown by the following exercise. Exercise. Define ˆx(t) := x(t) Φ(t, t 1 )x 1. Then ˆx satisfies d = A(t)ˆx(t) + B(t)u(t). dtˆx(t) Conclude that, under dynamics (2.1), u steers (x 0, t 0 ) to (x 1, t 1 ) if and only if it steers (x 0 Φ(t 0, t 1 )x 1, t 0 ) to (θ, t 1 ). Since Φ(t 0, t 1 ) is invertible, it follows that system (2.1) is reachable on [t 0, t 1 ] if and only if it is controllable on [t 0, t 1 ], i.e., if and only if, given x 0, there exists u U that steers (x 0, t 0 ) to (θ, t 1 ). Note. Equivalence between reachability and controllability (to the origin) does not hold in the discrete-time case, where controllability is a weaker property than controllability. Now reachability of (θ, t 1 ) from (ξ, t 0 ), for some ξ is equivalent to solvability of the linear :((مم) u equation (in Φ(t 1, t 0 )ξ + t 1 t 0 Φ(t 1, σ)b(σ)u(σ)dσ = 0. Equivalently (multiplying on the left by the nonsingular matrix Φ(t 0, t 1 )), ξ t 1 = Φ(t 0, σ)b(σ)u(σ)dσ (2.2) t 0 =: L(u). with L : U R n, a linear map. Clearly, θ is reachable at time t 1 from (ξ, t 0 ) if and only if ξ R(L). It follows that (2.1) is controllable on [t 0, t 1 ] if and only if R(L) = R n. Suppose now θ is reachable at time t 1 from (ξ, t 0 ) and suppose we want to reach θ while spending the least energy, i.e., while minimizing u, u = t 1 t 0 u(t) T u(t)dt where,مم مم is the L 2 inner product over U (see Example A.4 in Appendix A). That is, we want to solve the minimum energy fixed endpoint problem (FEP) minimize u, u subject to (2.1), x(t 1 ) = θ, and u U. We know (see Appendix A) that the unique such control is given by u = L η 0, where L is the adjoint of L for inner product,مم, مم and where η 0 is any point in R n satisfying LL η 0 = ξ 12 Copyright c , André L. Tits. All Rights Reserved

15 2.1 Fixed end point problem (and such points do exist). It is shown in Appendix A that L is given by which yields (L x)(t) = B T (t)φ T (t 0, t)x t 1 LL x = Φ(t 0, σ)b(σ)(l x)(σ)dσ t 0 = t 1 t 0 Φ(t 0, σ)b(σ)b T (σ)φ T (t 0, σ)dσ x, x, i.e., LL : R n R n is given by LL = t 1 t 0 Φ(t 0, t)b(t)b T (t)φ T (t 0, t)dt =: W(t 0, t 1 ). Since R(L) = R(LL ), θ is reachable at t 1 from (ξ, t 0 ) if and only if The matrix W(t 0, t 1 ) has entries ξ R(W(t 0, t 1 )). (W(t 0, t 1 )) ij = (Φ i (t 0, ((مم) B (مم T, (Φ j (t 0, ((مم) B (مم T, where,مم مم is again the L 2 inner product, i.e., W(t 0, t 1 ) is the Gramian matrix (or Gram matrix, or Gramian) associated with the vectors (Φ 1 (t 0, ((مم) B (مم T,...,(Φ n (t 0, ((مم) B (مم T, (مم) B =) T (Φ 1 (t 0, ((مم T,.. (مم) B,. T (Φ n (t 0, ((مم T ), which are the columns of the matrix (مم) B T (Φ(t 0, ((مم T ). It is known as the controllability Gramian. It is invertible if and only if R(L) = R n, i.e., if and only if the system is controllable on [t 0, t 1 ]. Suppose this is the case and let ξ := ξ(x 0, t 0 ) := x 0 Φ(t 0, t 1 )x 1. The minimum energy control that steers (x 0, t 0 ) to (x 1, t 1 ) is then given by i.e. and the corresponding energy is given by u 0 = L (LL ) 1 ξ(x 0, t 0 ) u 0 (t) = B T (t)φ T (t 0, t)w(t 0, t 1 ) 1 ξ(x 0, t 0 ) (2.3) u 0, u 0 = ξ(x 0, t 0 ), W(t 0, t 1 ) 1 ξ(x 0, t 0 ). Note that, as expressed in (2.3), u 0 (t) depends explicitly, through ξ(x 0, t 0 ), on the initial state x 0 and initial time t 0. Consequently, if between t 0 and the current time t, the state Copyright c , André L. Tits. All Rights Reserved 13

16 Linear Quadratic Optimal Control has been affected by an external perturbation, u 0 as expressed by (2.3) is no longer optimal (minimum energy) over the remaining time interval [t, t 1 ]. Let us address this issue. At time t 0, we have u 0 (t 0 ) = B T (t 0 )Φ T (t 0, t 0 )W(t 0, t 1 ) 1 ξ(x 0, t 0 ) = B T (t 0 )W(t 0, t 1 ) 1 ξ(x 0, t 0 ). Intuitively, this must hold independently of the value of t 0, i.e., for the problem under consideration, Bellman s Principle of Optimality holds: independently of the initial state (at t 0 ), for u 0 to be optimal for (FEP), it is necessary that u 0 applied from the current time t ق t 0 up to the final time t 1, starting at the current state x(t), be optimal for the remaining problem, i.e., for the objective function t 1 t u(τ) T u(τ)dτ. Specifically, given x R n and t [t 0, t 1 ] such that x 1 is reachable at time t 1 from (x, t), denote by P(x, t; x 1, t 1 ) the problem of determining the control of least energy that steers (x, t) to (x 1, t 1 ), i.e., problem (FEP) with (x, t) replacing (x 0, t 0 ). Let (مم) x be the state trajectory that results when optimal control u 0 is applied, starting from x 0 at time t 0. Then Bellman s Principle of Optimality asserts that, for any t [t 0, t 1 ], the restriction of u 0 to [t, t 1 ] solves P(x(t), t; x 1, t 1 ). Exercise 2.1 Prove that Bellman s Principle of Optimality holds for the minimum energy fixed endpoint problem. [Hint: Observe that the entire derivation so far still works when U is redefined to be piecewise continuous on [t 0, t 1 ] (i.e., continuous except at finitely many points, with finite left and right limits at those points). Then, Assuming by contradiction existence of a lower energy control for P(x(t), t; x 1, t 1 ), show by construction that u 0 cannot be optimal for P(x 0, t 0 ; x 1, t 1 ).] It follows that, for any t such that W(t, t 1 ) is invertible, u 0 (t) = B T (t)w(t, t 1 ) 1 ξ(x(t), t) t (2.4) which yields the closed loop implementation ( independent of initial time and state ) depicted in Figure 2.1, where v(t) = B T (t)w 1 (t, t 1 )Φ(t, t 1 )x 1. In particular, if x 1 = θ, then v(t) = 0 t. If, for whatever reason, the state is perturbed at some point in time, (2.4) is still optimal over the remaining time interval (assuming there is no subsequent perturbation) while (2.3) is not. Exercise 2.2 Prove (2.4) from (2.3) directly, without invoking Bellman s principle. Example 2.1 (charging capacitor) d cv(t) = i(t) dt minimize t 1 0 ri(t) 2 dt s.t. v(0) = v 0, v(t 1 ) = v 1 14 Copyright c , André L. Tits. All Rights Reserved

17 2.1 Fixed end point problem v(t) + - u(t) B(t) + x(t) x(t) A(t) B T(t)W -1(t,t 1 ) Figure 2.1: Closed loop implementation r i c + - v Figure 2.2: Charging capacitor We obtain B(t) 1 c ; A(t) 0 W(0, t 1 ) = t c 2dt = t 1 c 2 η 0 = c 2v 0 v 1 t 1 i 0 (t) = 1 c 2 (v 0 v 1 ) = c(v 1 v 0 ) = constant. c t 1 t 1 The closed loop optimal feedback law is given by i 0 (t) = c t 1 t (v(t) v 1). Exercise 2.3 Discuss the same optimal control problem (with fixed end points) with the objective function replaced by J(u) t 1 t 0 u(t), R(t)u(t) dt Copyright c , André L. Tits. All Rights Reserved 15

18 Linear Quadratic Optimal Control where R(t) = R(t) T > 0 for all t [t 0, t 1 ] and (مم) R is continuous. [Hint (to be proved if you use it, as usual): There is no loss of generality in assuming that R(t) is symmetric. Further, the Cholesky factor of a symmetric positive definite matrix is unique and continuous. Consider a change of control variable. Other approach: define a new inner product on U.] Remark 2.1 In view of (2.2), a control u steers (x 0, t 0 ) to (x 1, t 1 ) if and only if it steers (x 0 Φ(t 0, t 1 )x 1, t 0 ) to (θ, t 1 ). Thus there is no loss of generality in assuming x 1 = θ. The controllability Gramian,مم) W (مم happens to satisfy certain simple equations. Recalling that W(t, t 1 ) = one easily verifies that W(t 1, t 1 ) = 0 and t 1 t Φ(t, σ)b(σ)b T (σ)φ(t, σ) T dσ d dt W(t, t 1) = A(t)W(t, t 1 ) + W(t, t 1 )A T (t) B(t)B T (t) (2.5) implying that, if W(t, t 1 ) is invertible, it satisfies d dt W(t, t 1) 1 = W(t, t 1 ) 1 A(t) A T (t)w(t, t 1 ) 1 + W(t, t 1 ) 1 B(t)B T (t)w(t, t 1 ) 1 (2.6) Exercise 2.4 Prove (2.6). (Hint: d dt M(t) 1 = M(t) 1 ( d dt M(t))M(t) 1.) Equation (2.5) is linear. It is a Lyapunov equation. Equation (2.6) is quadratic. It is a Riccati equation (for W(t, t 1 ) 1 ).,مم) W (مم also satisfies the functional equation W(t 0, t 1 ) = W(t 0, t) + Φ(t 0, t)w(t, t 1 )Φ T (t 0, t). As we will see later, the Riccati equation plays a fundamental role in optimal control systems involving linear dynamics and quadratic cost (linear-quadratic problems). At this point, simply note that, if x 1 = θ and W(t, t 1 ) is invertible, then u 0 (t) = B(t) T P(t)x(t), where P(t) = W(t, t 1 ) 1 solves Riccati equation (2.6). Example 2.2 (scalar Riccati equation) k = 1 k 2, k(t) = 0 Solution tanh 1 (k(t)) tanh 1 (k(t)) = t T k(t) = tanh (t T). 16 Copyright c , André L. Tits. All Rights Reserved

19 2.1 Fixed end point problem We have seen that, if W(t 0, t 1 ) is invertible, the optimal cost for problem (FEP) is given by J(u 0 ) = u 0, u 0 = x(t 0 ) Φ(t 0, t 1 )x 1, W(t 0, t 1 ) 1 (x(t 0 ) Φ(t 0, t 1 )x 1 ). (2.7) This is clearly true for any t 0, so that, from a given time t < t 1 (such that W(t, t 1 ) is invertible) the cost-to-go is given by x(t) Φ(t, t 1 )x 1, W(t, t 1 ) 1 (x(t) Φ(t, t 1 )x 1 ). In particular, if x 1 = θ, the cost-to-go takes value x(t), W(t, t 1 ) 1 x(t). We will see that such an expression arises in other types of linear-quadratic optimal control problems, with W(t, t 1 ) 1 replaced by some matrix K(t), solution of a certain Riccati equation. The following fundamental lemma, in which K(t) is arbitrary, expresses the variation of such expressions as a certain integral. Lemma 2.1 (Fundamental Lemma) Let,(مم) A (مم) B be continuous matrix-value functions and (مم) K = K T (مم) be an absolutely continuous matrix-valued function. Suppose K(t) exists on [t 0, t 1 ]. Then, if x(t) and u(t) are related by ẋ(t) = A(t)x(t) + B(t)u(t), (2.8) it holds x(t 1 ), K(t 1 )x(t 1 ) x(t 0 ), K(t 0 )x(t 0 ) = t 1 t 0 [ x(t) u(t) ], [ K(t) + A T (t)k(t) + K(t)A(t) K(t)B(t) B(t) T K(t) 0 ] [ x(t) u(t) ] dt (2.9) Proof. x(t 1 ), K(t 1 )x(t 1 ) x(t 0 ), K(t 0 )x(t 0 ) = t 1 t 0 d x(t), K(t)x(t) dt dt = t 1 t 0 ( ẋ(t), K(t)x(t) + x(t), K(t)x(t) + x(t), K(t)ẋ(t) )dt and the claim follows if one substitutes for ẋ(t) the right hand side of (2.8). Note that, while the integral in this lemma involves the paths x(t), u(t), t [t 0, t 1 ], its value depends only on the end points of the trajectory,(مم) x i.e., this integral is path independent. Now consider the more general quadratic cost J(u) := t 1 t 1 t 0 x(t), L(t)x(t) + u(t), u(t) dt = t 0 [ x(t) u(t) ], [ L(t) 0 0 I ] [ x(t) u(t) ] dt,(2.10) Copyright c , André L. Tits. All Rights Reserved 17

20 Linear Quadratic Optimal Control where (مم) L = (مم) L T is continuous. Let K(t) = K(t) T be some absolutely continuous timedependent matrix. Using the Fundamental Lemma we see that, since x 0 and x i are fixed, it is equivalent to minimize J(u) := J(u) + x 1, K(t 1 )x 1 x 0, K(t 0 )x 0 := t 1 [ ] x(t) =, u(t) t 0 t 1 t 0 ( x(t), L(t)x(t) + u(t), u(t) ) dt [ L(t) + K(t) + A T (t)k(t) + K(t)A(t) K(t)B(t) B(t) T K(t) 0 To complete the square, suppose there exists such K(t) that satisfies i.e., L(t) + K(t) + A T (t)k(t) + K(t)A(t) = K(t)B(t)B(t) T K(t) ] [ x(t) u(t) ] dt. K(t) = A T (t)k(t) K(t)A(t) + K(t)B(t)B(t) T K(t) L(t), (2.11) a Riccati Differential Equation (DRE). (Existence of a solution to this equation depends, in general, on.(مم) L As we will see late, when discussing the free endpoint problem, if L(t) is positive semidefinite for all t, then a solution exists, e.g., for every prescribed positive semidefinite final value K(t 1 ).) Then we get t 1 [ ] x(t) J(u) =, u(t) t 0 = t 1 [ K(t)B(t)B(t) T K(t) K(t)B(t) B(t) T K(t) 0 ] [ x(t) u(t) t 0 B(t) T K(t)x(t) + u(t), B(t) T K(t)x(t) + u(t) dt. Now, again supposing that some solution to (DRE) exists, let (مم) K be any such solution and let v(t) = B T (t)k(t)x(t) + u(t). It is readily verified that, in terms of the new control input v, the systems dynamics become and the cost function takes the form That is, we end up with the problem minimize ẋ(t) = [A(t) B(t)B T (t)k(t)]x(t) + Bv(t), t 1 t 0 J(u) = t 1 v(t), v(t) dt t 0 v(t), v(t) dt. ] dt. subject to ẋ(t) = [A(t) B(t)B T (t)π(t, K 1, t 1 )]x(t) + Bv(t) (2.12) x(t 0 ) = x 0 x(t 1 ) = x 1 18 Copyright c , André L. Tits. All Rights Reserved

21 2.1 Fixed end point problem where, following standard using, we have parametrized the solutions to DRE by there value (K 1 ) at time t 1, and denoted them by Π(t, K 1, t 1 ). This transformed problem (parameterized by K 1 ) is of an identical form to that we solved earlier. Denote by Φ A BB T Π and W A BB T Π the state transition matrix and controllability Gramian for (2.12) (for economy of notation, we have kept implicit the dependence of Π on K 1. Then, for a given K 1, we can write the optimal control v K 1 0 for the transformed problem as (assuming x 1 = θ for simplicity) v K 1 0 (t) = B(t)T W 1 A BB T Π x(t), and the optimal control u 0 for the original problem as u 0 (t) = v K 1 0 (t) B(t)T Π(t, K 1, t 1 )x(t) = B(t) T (Π(t, K 1, t 1 ) + W 1 A BB T Π )x(t). Still with x 1 = θ, we also obtain J(u 0 ) = J(u 0 ) + x 0, Π(t 0, K 1, t 1 )x 0 = x 0, (W A BB T Π(t 0, t 1 ) 1 + Π(t 0, K 1, t 1 ))x 0. The cost-to-go at time t is x(t), (W A BB T Π(t, t 1 ) 1 + Π(t, K 1, t 1 ))x(t) Finally, if L(t) is identically zero, we can pick K(t) identically zero and we recover the previous result. Exercise 2.5 Show that controllability of (2.1) on [t 0, t 1 ] implies invertibility of W A BB T Π and vice-versa. We obtain the block diagram depicted in Figure 2.3. Thus u 0 (t) = B(t) T p(t), with w = v 0 + u 0 - B(t) + + x A(t) B T(t)Π(t,K 1,t 1 ) B T(t)W A-BB TΠ (t,t 1 ) -1 Figure 2.3: Optimal feedback law p(t) = (Π(t, K 1, t 1 ) + W A BB T Π(t, t 1 ) 1 )x(t). Copyright c , André L. Tits. All Rights Reserved 19

22 Linear Quadratic Optimal Control If x 1 θ, then w 0 (t) = B T (t)w A BB T Π(t, t 1 ) 1 (x(t) Φ A BB T Π(t, t 1 )x 1 ). Note that while v 0 clearly depends on K 1, u 0 obviously cannot, since K 1 is an arbitrary symmetric matrix (subject to DRE having a solution with K(t 1 ) = K 1 ). Thus we could have assigned K 1 = 0 throughout the analysis. Check the details of this. The above is a valid closed-loop implementation as it does not involve the initial point (x 0, t 0 ) (indeed perturbations may have affected the trajectory between t 0 and the current time t). Π(t, K 1, t 1 ) can be precomputed (again, we must assume that such a solution exists.) Finally, note that the optimal cost J(u 0 ) is given by J(u 0 ) = J(u 0 ) ( x 1, K 1, x 1 x 0, Π(t 0, K 1, t 1 )x 0 ) = x 0 Φ A BB T Π(t 0, t 1 )x 1, W A BB T Π(t 0, t 1 ) 1 (x 0 Φ A BB T Π(t 0, t 1 )x 1 ) + x 0, Π(t 0, K 1, t 1 )x 0 x 1, K 1, x 1, and is independent of K 1. If x 1 = θ, we get J(u 0 ) = x 0, (W A BB T Π(t 0, t 1 ) 1 + Π(t 0, K 1, t 1 ))x 0 and the cost-to-go from any time t < t 1 and state x(t) is x(t), (W A BB T Π(t, t 1 ) 1 + Π(t, K 1, t 1 ))x(t) 2.2 Free end point problem We now consider optimal control problem for which the final value of the state is left unspecified (free end point). For such problems, there may be a cost associated with this value. In the context of quadratic objective functions we are led to consider t 1 minimize J(u) ( x(t), L(t)x(t) + u(t), u(t) )dt + x(t 1 ), Qx(t 1 ) t 0 subject to ẋ(t) = A(t)x(t) + B(t)u(t) (2.13) x(t 0 ) = x 0. Without loss of generality, L and Q are assumed symmetric (both can be zero). Using Lemma 2.1 one can both complete the square and do away with the final cost term. For this, let K(t) be solution of the Riccati equation K(t) = A T (t)k(t) K(t)A(t) L(t) + K(t)B(t)B T (t)k(t) with the condition i.e. K(t 1 ) = Q K(t) = Π(t, Q, t 1 ) 20 Copyright c , André L. Tits. All Rights Reserved

23 2.2 Free end point problem We assume that such a solution exists (this need not be the case). Then, eliminating the last term in J(u) by means of the lemma and performing manipulations identical to those operated in the fixed end point case, we obtain J(u) = t 1 t 0 B T (t)π(t, Q, t 1 )x(t) + u(t) 2 dt + x(t 0 ), Π(t 0, Q, t 1 )x(t 0 ). (2.14) Since the initial point is fixed (x(t 0 ) = x 0 ), the optimal control in closed loop form is given by u 0 (t) = B T (t)π(t, Q, t 1 )x(t) (2.15) and the optimal value is given by x(t 0 ), Π(t 0, Q, t 1 )x(t 0 ). The cost-to-go from any time t < t 1 and state x(t) is x(t), Π(t, Q, t 1 )x(t). Exercise 2.6 Let p(t) = Π(t, Q, t 1 )x(t). Then u 0 (t) = B(t) T p(t), [ ẋ(t) ṗ(t) ] [ A(t) B(t)B = T (t) L(t) A T (t) ] [ x(t) p(t) ] (2.16) and J(u 0 ) = x 0, p(t 0 ). Remark 2.2 We have not made any positive (or even nonnegative) definiteness assumption on L(t) or Q. The key assumption is that the stated Riccati equation has a solution Π(t, Q, t 1 ). Below we investigate conditions (in particular, on L and Q) which insure that this is the case. Exercise 2.7 Investigate the case of the more general cost function J(u) = t 1 t 0 ( x(t), L(t)x(t) + 2 u(t), S(t)x(t) + u(t), R(t)u(t) )dt, where R(t) is positive definite for all t. Hint: let v(t) = u(t)+m(t)x(t) with M(t) judiciously chosen. Remark 2.3 In all cases considered so far, the optimal control (when it exists) is in the range space of B T (t) for all t. Give a simple explanation for this fact. When L(t) = 0 for all t and Q is nonsingular, the solution of the Riccati equation is directly interpretable in terms of the controllability Gramian W(t, t 1 ). We observed above that if M(t) satisfies the linear equation Ṁ(t) = M(t)A T (t) + A(t)M(t) B(t)B T (t) (2.17) Copyright c , André L. Tits. All Rights Reserved 21

24 Linear Quadratic Optimal Control and if K(t) := M(t) 1 exists, then it satisfies the Riccati equation K(t) = A T (t)k(t) K(t)A(t) + K(t)B(t)B T (t)k(t). (2.18) Note that W(t, t 1 ) satisfies (2.17), with the terminal condition W(t 1, t 1 ) = 0. Its inverse (if it exists) satisfies (2.18) but does not satisfy the required condition at t 1 (in fact it does not exist at t = t 1 ). The solution to linear ordinary equation (2.17) taking on the value Q 1 (assuming Q is invertible) at t = t 1 is given by H(t) = W(t, t 1 ) + Φ(t, t 1 )Q 1 Φ T (t, t 1 ). (Prove this; the first term is the zero state response, the second one the zero input response.) Thus Π(t, Q, t 1 ) = H(t) 1 = (W(t, t 1 ) + Φ(t, t 1 )Q 1 Φ T (t, t 1 )) 1 Exercise 2.8 Assume that Q > 0. Show that H(t) is invertible t ق t 1 (even when W(t, t 1 ) is singular) We now turn to the case when L(t) is not identically zero. First, we derive some properties of adjoint equations. Given the equation ẋ(t) = A(t)x(t) its adjoint equation is given by ṗ(t) = A T (t)p(t). Denote by Φ A (t, t 0 ) and Φ A T (t, t 0 ) the respective state transition matrices. First, letting f(t) = x(t), p(t), one has so that f(t) = ẋ, p + x, ṗ = Ax, p + x, A T p = 0 x(t), p(t) = x(t 0 ), p(t 0 ) t. (2.19) It follows that Φ A T (t, t 0 ) = Φ T A(t 0, t). (Note the change of order of the arguments.) Corollary 2.1 d dt Φ A(t 0, t) = Φ A (t 0, t)a(t) Exercise 2.9 Prove the corollary. 22 Copyright c , André L. Tits. All Rights Reserved

25 2.2 Free end point problem We now turn back to the Riccati equation K(t) = A(t) T K(t) K(t)A(t) + K(t)B(t)B T (t)k(t) L(t) K(t 1 ) = Q (2.20) To analyze this equation, we consider the linear system ( ) [ ẋ(t) A(t) B(t)B = T (t) ṗ(t) L(t) A T (t) ] ( x(t) p(t) ) (2.21) evolving in R 2n. We will see that trajectories of (2.21) are directly related to optimal trajectories of our linear/quadratic optimal control problems. (Also see Exercise 2.6 above.) We first show that the Riccati equation (2.20) can be solved via system (2.21). Theorem 2.1 Let,مم) Ψ (مم be the 2n 2n state transition matrix for system (2.21) and let Then ( X(t) P(t) ) = Ψ(t, t 1 ) ( I Q ) Π(t, Q, t 1 ) = P(t)X(t) 1 solves (2.20) for t [τ, t 1 ] for any τ < t 1 such that X(t) 1 exists on [τ, t 1 ]. Proof. Just plug in and use the fact that ( ) [ Ẋ(t) A(t) B(t)B P(t) = T (t) L(t) A T (t) ] ( X(t) P(t) ). Note that, since X(t 1 ) = I, by continuity of the state transition matrix, X(t) is invertible for t close enough to t 1, and thus the Riccati equation has a solution for t close to t 1. This solution is unique since, over bounded subsets of R n n, the right-hand side of the Riccati equation satisfies a Lipschitz condition. Since clearly the transpose of a solution to the Riccati equation is also a solution, this unique solution must be symmetric, as required in the path-independent lemma. Also, the only way the solution can disappear is if X(t) becomes singular. We now show that, when Π(t, Q, t 1 ) disappears, it must first grow without bounds. Lemma 2.2 Let ˆt = inf»τ : Π(t, Q, t 1 ) exists t [τ, t 1 ]. If ˆt >,ف then,مم) Π Q, t 1 ) is unbounded on (ˆt, t 1 ]. Proof. Let so that (2.20) can be written ϕ(k, t) = A(t) T K KA(t) + KB(t)B T (t)k L(t), K(t) = ϕ(k(t), t), K(t 1 ) = Q. Copyright c , André L. Tits. All Rights Reserved 23

26 Linear Quadratic Optimal Control ϕ is continuous and is Lipschitz continuous in its first argument on every bounded set. The claim is then an immediate consequence of a classical result on the solution of ordinary differential equations: See, e.g., [11, Theorem 2.1], with D = (ˆt, t 1 ) R n(n+1) 2. Theorem 2.2 Suppose L(t) is positive semidefinite for all t and Q is positive semidefinite. Then Π(t, Q, t 1 ) exists t ق t 1. Proof. Again, let ˆt = inf»τ : Π(t, Q, t 1 ) exists t [τ, t 1 ], so that,مم) Π Q, t 1 ) exists in (ˆt, t 1 ]. In view of Lemma 2.2, it is enough to prove that,مم) Π Q, t 1 ) is bounded on (ˆt, t 1 ]. Let τ (ˆt, t 1 ]. For any x R n we have t 1 0 ق ) 1 x, Π(τ, Q, t 1 )x = min ( u(t), u(t) + x(t), L(t)x(t) )dt + x(t 1 ), Qx(t u U τ where x(t) satisfies (2.13) with initial condition x(τ) = x. Letting ˆx(t) be the solution to (2.13) corresponding to u(t) = 0 t and ˆx(τ) = x, we can write with ق )x x, Π(τ, Q, t 1 ق 0 F(τ) = t1 τ t 1 τ ˆx(t), L(t)ˆx(t) dt + ˆx(t 1 ), Qˆx(t 1 ) = x, F(τ)x Φ(t, τ) T L(t)Φ(t, τ)dt + Φ(t 1, τ) T QΦ(t 1, τ), a continuous function. Since this holds for all x R n and since Π(t, Q, t 1 ) is symmetric, it follows that, using, e.g., the spectral norm, Π(τ, Q, t 1 ) ق F(τ). Since (مم) F is continuous on [ˆt, t 1 ], it is bounded on [ˆt, t 1 ], hence on (ˆt, t 1 ]. The proof is complete. Corollary 2.2 If L(t) is positive semidefinite for all t, then the DRE (2.11 has a solution for some symmetric final value K 1 (indeed, for any symmetric positive semidefinite K 1 ), and hence the fixed end point problem has a solution. ق A 2 x, Fx for all x, then ق Ax x, ق Exercise 2.10 Prove that, if A = A T and 0 F 2. Thus, when L(t) is positive semidefinite for all t and Q is positive semidefinite, our problem has a unique optimal control given by (2.15). Finally, for all optimal control problems considered so far, the optimal control can be expressed in terms of the adjoint variable (or costate).(مم) p More precisely, the following holds. 24 Copyright c , André L. Tits. All Rights Reserved

27 2.2 Free end point problem Theorem 2.3 Consider either the fixed end point problem, with x 1 = θ, or the free end point problem studied above. Suppose that the controllability Gramian W(t 0, t 1 ) is nonsingular (for the fixed end point case) and that the relevant Riccati equation has a (unique) solution K(t) on [t 0, t 1 ] with K(t 1 ) = K 1. Let x(t) be the optimal trajectory and define p(t) by for the fixed end-point problem and by p(t) = (Π(t, K 1, t 1 ) + W A BB T Π(t, t 1 ) 1 )x(t) p(t) = Π(t, Q, t 1 )x(t) for the free end-point problem so that the optimal control is given by u 0 (t) = B T (t)p(t). Then ( ẋ(t) ṗ(t) ) ( A(t) B(t)B = T (t) L(t) A T (t) ) ( x(t) p(t) ) and the optimal cost is x(t 0 ), p(t 0 ). Exercise 2.11 Prove the theorem. Remark 2.4 Let H : R n R n R m R R be given by H(ξ, η, u, t) = 1 2 u, u + 1 ξ, L(t)ξ + η, A(t)ξ + B(t)u, 2 and let H : R n R n R R be defined by Thus H(ξ, η, t) = 1 2 ( ξ η H(ξ, η, t) = min u R m H(ξ, η, u, t). ), [ L(t) A T (t) A(t) B(t)B T (t) ] ( ξ η ) = 1 2 ξ, L(t)ξ + η, A(t)ξ 1 2 η, B(t)BT (t)η. The gradient of H with respect to the first 2n arguments is given by [ L(t) A H(ξ, η, t) = T ] ( ) (t) ξ A(t) B(t)B T. (t) η System (2.21) can then be equivalently written in the canonical form ẋ(t) = H (x(t), p(t), t) p ṗ(t) = H (x(t), p(t), t) x Copyright c , André L. Tits. All Rights Reserved 25

28 Linear Quadratic Optimal Control i.e., where z(t) = ( x(t) p(t) ) and J = ż(t) = J H(x(t), p(t), t) (2.22) ( 0 I I 0 ). This is one instance of Pontryagin s maximum principle ; see Chapter 5 for more details.) The function H is called pseudo-hamiltonian and H is called Hamiltonian (or true Hamiltonian) for the problem under consideration. Remark 2.5 Along trajectories of (2.21), since d dt H(x(t), p(t), t) = H(x(t), p(t), t), ż(t) + H(x(t), p(t), t) t = H(x(t), p(t), t) t H(x(t), p(t), t), ż(t) = H(x(t), p(t), t), J H(x(t), p(t), t) = 0 (since J T = J). In particular, if A, B and L do not depend on t, H(x(t), p(t), t) is constant along trajectories of (2.21). 2.3 Infinite horizon LTI problem We now turn our attention to the case of infinite horizon (t 1 =.(ف We also assume that A, B and L are constant. Assuming (as above) that L = L T ق 0, we write L = C T C, so that the problem can be written as subject to minimize J(u) = 0 ( u(t), u(t) + y(t), y(t) )dt ẋ = Ax + Bu y = Cx x(0) = x 0 Note that y is merely some linear image of x, and need not be a physical output. For example, it could be an error output. 26 Copyright c , André L. Tits. All Rights Reserved

29 2.3 Infinite horizon LTI problem We know that Π( t, 0, 0) exists for all t ق 0 and is symmetric positive definite. Intuition suggests that, since the time-to-go is infinite at any time t, and the system is time-invariant, the optimal feedback should not depend on t, i.e., that Π( t, 0, 0) converges to some limit Π as t,ف that its derivative converges to zero, and that Π thus satisfies the algebraic Riccati equation A T Π + Π A Π BB T Π = C T C (ARE) We now show that, under certain natural assumptions, this intuition is essentially correct. For any u U, t ق 0, define J t (u) = t 0 ( u(τ), u(τ) + x(τ), C T Cx(τ) )dτ. Let K be any constant symmetric matrix that satisfies (ARE), let u U, let x be the corresponding state trajectory, and let t ق 0. Note that, since K = 0, K also satisfies K = A T K KA + KBB T K C T C (DRE) It follows that J t (u) = t 0 u(τ) + B T Kx(τ) 2 x(t), Kx(t) + x 0, Kx 0. (2.23) Theorem 2.4 Let K = K T ق 0 satisfy (ARE) and let û U be determined by the feedback control law û = B T Kx. Then J(û) ق x 0, Kx 0. Proof. Since K ق 0 it follows from (2.23) that 0. ق x 0, Kx 0 t ق (û) J t.ف t The claim follows by letting Now note that, from Theorem 2.2, Π( t, 0, 0) exists for all t ق 0 and is symmetric, and that, by time-invariance of (DRE), 0. ق Π(0, 0, t) = Π( t, 0, 0) for all t Clearly any stabilizing feedback law yields a finite value for J(u). It turns out that stabilizability is sufficient for the existence of an optimal control. 0, ق Π Theorem 2.5 Suppose (A, B) is stabilizable. Then, for some matrix and Π solves (ARE). Moreover ف t Π(0, 0, t) Π as J(u) ق x 0, Π x 0 u U. Copyright c , André L. Tits. All Rights Reserved 27

30 Linear Quadratic Optimal Control Proof. Direct consequence of the following two lemmas..ف t Lemma 2.3 For (DRE), if (A, B) is stabilizable, Π( t, 0, 0) remains bounded as.ف t Furthermore there exists a matrix Π satisfying (ARE) such that Π(0, 0, t) Π as 0, ق Proof. Given x 0 R n we have, for t x 0, Π(0, 0, t)x 0 = min u U J t(u). It follows that Π(0, 0, t) ق 0 for t ق 0. Next, let û(t) = F x(t) be any stabilizing static state feedback and let ˆx be the corresponding solution of ẋ = (A + BF)x x(0) = x 0, i.e., ˆx(t) = e (A+BF)t x 0. Then, since A + BF is stable, ق 0 x 0, Π(0, 0, t)x 0 û(σ), û(σ) + ˆx(σ), C T Cˆx(σ) dσ = x 0, M x 0 with M = 0 e (A+BF)T t (F T F + C T C)e (A+BF)t dt 0. ق for all t Since Π(0, 0, t) is symmetric positive semidefinite, the first claim follows (prove it). Now, for all t ق 0, we have t x 0, Π(0, 0, t)x 0 = min u(σ), u(σ) + x(σ), C T Cx(σ) dσ. (2.24) u U 0 Nonnegative definiteness of C T C implies that x 0, Π(0, 0, t)x 0 increases monotonically as t increases. Since it is bounded, it must converge and its time derivative must go to zero for any x 0. 1 Using the fact that Π(0, 0, t) is symmetric it is easily shown that it converges (prove it), i.e., for some symmetric matrix Π, lim t Π(0, 0, t) = Π or equivalently lim Π( t, 0, 0) = Π, t and that its derivative goes to zero. Taking limits on both sides of (DRE) shows that Π satisfies (ARE). 1 The latter does not follow from mere convergence of x 0, Π(0, 0, t)x 0 : consider, e.g., the function (1/t)sin(t 2 ). 28 Copyright c , André L. Tits. All Rights Reserved

31 2.3 Infinite horizon LTI problem For any t ق 0 we now define J t (u) = t 0 ( u(σ), u(σ) + Cx(σ), Cx(σ) )dσ. Lemma 2.4 Assume that (A, B) is stabilizable. Then, for any u U, J(u) ق x 0, Π x 0. Proof. Since C T C is positive semidefinite, for any t > 0, J(u) ق J t (u) ق inf u U J t(u) = x 0, Π(0, 0, t)x 0. Letting t ف and using Lemma 2.3 proves the claim. Thus we have proved the following. Theorem 2.6 Suppose (A, B) is stabilizable, and let Π ق 0 be as in Theorem 2.5. Then Π solves (ARE), the control law û = B T Π x is optimal, and J(û) = x 0, Π x 0. This solves the infinite horizon LTI problem. Additional insight is provided by the following results. Theorem 2.7 If (C, A) is detectable and K ق 0 satisfies (ARE), then A BB T K is stable. Proof. (ARE) can be rewritten (A BB T K) T K + K(A BB T K) = KBB T K C T C. (2.25) Proceed now by contradiction. Let λ, with Reλ ق 0, and v 0 be such that (A BB T K)v = λv. (2.26) Multiplying (2.25) on the left by v and on the right by v we get 2(Reλ)(v Kv) = B T Kv 2 Cv 2. Since the left-hand side is nonnegative and the right-hand side nonpositive, both sides must vanish. Thus (i) Cv = 0 and, (ii) B T Kv = 0 which together with (2.26), implies that Av = λv. Since Reλ ق 0, this contradicts detectability of (C, A). Corollary 2.3 If (A, B) is stabilizable and (C, A) is detectable, then the optimal control law û = B T Π x is stabilizing. Copyright c , André L. Tits. All Rights Reserved 29

32 Linear Quadratic Optimal Control Theorem 2.8 Suppose (A, B) is stabilizable. Then, (i) if (C, A) is observable, then Π > 0; more generally, (ii) if Π x 0 = 0 (equivalently, x 0 Π x 0 = 0) for some x 0, then x 0 belongs to the unobservable subspace. Proof. J(û) = x 0, Π x 0 = 0 = 0 û, û + Cx, Cx dt. Thus û(t) = 0 for all t and Cx(t) = 0 for all t, so that x 0 is unobservable. Computation of Π Let [ A BB T H = L A T ], and let K + be a stabilizing solution to (ARE). Let [ I 0 T = K + I ]. Then T 1 = [ I 0 K + I ]. Now (elementary block column and block row operations) [ ] [ ] [ ] T 1 I 0 A BB T I 0 HT = K + I L A T K + I [ ] [ I 0 A BB = T K + BB T ] K + I L A T K + A T [ A BB = T K + BB T ] 0 (A BB T K + ) T since K + is a solution to (ARE). It follows that σ(h) = σ(a BB T K + ) σ( (A BB T K + )), where (مم) σ denotes the spectrum (set of eigenvalues). Thus, if (A, B) is stabilizable and (C, A) detectable, H cannot have any imaginary eigenvalues: It must have n eigenvalues in C and n eigenvalues in C +. Furthermore the first n columns of T form a basis for the stable invariant subspace of H, i.e., for the span of all generalized eigenvectors of H associated with stable eigenvalues ](see, e.g., Chapter 13 of [23] for more on this). Now let [ S11 S 21 be any basis for the stable invariant subspace of H, i.e., let S = [ S11 S 12 S 21 S Copyright c , André L. Tits. All Rights Reserved ]

33 2.4 More general optimal control problems be any invertible matrix such that S 1 HS = [ X Z 0 Y ] for some X, Y, Z such that σ(x) C and σ(y ) C +. (Note that σ(h) = σ(x) σ(y ).) Then it must hold that, for some nonsingular R, [ ] [ ] S11 I = K + R. S 21 It follows that S 11 = R and S 21 = K + R, thus We have thus proved the following. K + = S 21 S Theorem 2.9 Suppose (A, B) is stabilizable and (C, A) is detectable, ] and let K + be a stabilizing solution of (ARE). Then K + = S 21 S 1 11, where [ S11 S 21 is any basis for the stable invariant subspace of H. In particular, S 21 S 1 11 is symmetric and there is exactly one stabilizing solution to (ARE). Corollary 2.4 Suppose again that (A, B) is stabilizable and that (C, A) is detectable. Then there is a unique positive semidefinite solution to (ARE). Moreover, if (C, A) is observable, that solution is positive definite. Proof. The first statement follows directly from Theorem 2.7 and the last assertion in Theorem 2.9. Then second statement follows from Corollary 2.3 and Theorem 2.8. [ ] 0 I Exercise 2.12 Given J =, any real matrix H that satisfies J I 0 1 HJ = H T is said to be Hamiltonian. Show that if H is Hamiltonian and λ is an eigenvalue of H, then λ also is. 2.4 More general optimal control problems We have shown how to solve optimal control problems where the dynamics are linear, the objective function is quadratic, and the constraints are of a very simple type (fixed initial point, fixed final point). In most problems of practical interest, though, one or more of the following features is present. (i) nonlinear dynamics and objective function (ii) constraints on the control or state trajectories, e.g., u(t) U t, where U R m (iii) more general constraints on the initial and final state, e.g., g(x(t 1 )) ق 0. To tackle such more general optimization problems, we will make use of additional mathematic machinery. We first proceed to develop such machinery. Copyright c , André L. Tits. All Rights Reserved 31

34 Linear Quadratic Optimal Control 32 Copyright c , André L. Tits. All Rights Reserved

35 Chapter 3 Unconstrained Optimization References: [15, 19]. 3.1 First order condition of optimality We consider the problem min»f(x) x X (3.1) where, as assumed throughout, X is a normed vector space and f : X R is continuously (Fréchet) differentiable (see Appendix B). (In fact, many of the results we obtain below hold under the milder assumption that f is merely Fréchet differentiable at the local minimizer of interest.) Remark 3.1 This problem is technically very similar to the problem min»f(x) x Ω (3.2) where Ω is an open set of X, as shown in the following exercise. Exercise 3.1 Prove carefully, using the definitions given earlier, that ˆx is a local minimizer for (3.2) if and only if ˆx is a local minimizer for (3.1) and ˆx Ω. We now give a first order necessary condition for optimality. Theorem 3.1 Suppose ˆx is a local minimizer for (3.1). Then f (ˆx) = 0. x Proof. Since ˆx is a local minimizer for (3.1), there exists ǫ > 0 such that Since f is Fréchet-differentiable, we have, for all h X, f(ˆx + h) ق f(ˆx) h B(0, ǫ) (3.3) f(ˆx + h) = f(ˆx) + f (ˆx)h + o(h) (3.4) x Copyright c , André L. Tits. All Rights Reserved 33

36 Unconstrained Optimization with o(h) h 0 as h 0. Hence, from (3.3), whenever h B(0, ǫ) or, equivalently, f (3.5) 0 ق o(h) (ˆx)h + x and, dividing by α, for α 0 f (ˆx)(αh) + o(αh) ق 0 h B(0, ǫ), α [0, 1] (3.6) x f o(αh) (ˆx)h + ق 0 h B(0, ǫ), α (0, 1] (3.7) x α It is easy to show (see exercise below) that o(αh) α (3.7), we get 0 as α 0. Hence, letting α 0 in f 0 ق x (ˆx)h h B(0, ǫ). Since h B(0, ǫ) implies h B(0, ǫ), we have f (ˆx)( h) ق 0 thus x Hence f (ˆx)h ق 0 h B(0, ǫ). x f (ˆx)h = 0 h B(0, ǫ) x which implies (since f (ˆx) is linear) x f (ˆx)h = 0 h X x i.e., f (ˆx) = θ. x Exercise 3.2 If (مم) o is such that o(h) h 0 as h 0, then o(αh) α 0 as α 0 α R. Remark 3.2 The optimality condition above, like several other conditions derived in this course, is only a necessary condition, i.e., a point x satisfying this condition need not be optimal, even locally. However, if there is an optimal x, it has to be among those which satisfy the optimality condition. Also it is clear that this optimality condition is also necessary for a global minimizer (since a global minimizer is also a local minimizer). Hence, if a global minimizer is known to exist, it must be, among the points satisfying the optimality condition, the one with minimum value of f. 34 Copyright c , André L. Tits. All Rights Reserved

37 3.2 Steepest descent method Exercise 3.3 Check that, if f is merely Gâteaux differentiable, the above theorem is still valid. (Further, note that continuous differentiability is not required.) Suppose now that f f (ˆx) θ (hence ˆx is not a local minimizer). If h is such that (ˆx)h < 0 x x (such h exists; why?) then, for α > 0 small enough, we have and, hence, for some α 0 > 0, α» f o(αh) (ˆx)h + x α < 0 f(ˆx + αh) < f(ˆx) α (0, α 0 ]. Such h is called a descent direction for f at ˆx. The concept of descent direction is essential to numerical methods. If X is a Hilbert space, then f (ˆx)h = gradf(ˆx), h and a particular x descent direction is h = gradf(ˆx). (This is so irrespective of which inner product (and associated gradient) is used. We will return to this point when studying Newton s method and variable metric methods.) 3.2 Steepest descent method Suppose X = H, a Hilbert space. In view of what we just said, a natural algorithm for attempting to solve (3.1) would be the following. Algorithm 1 (steepest descent) Data x 0 H i = 0 while gradf(x i ) θ do» pick λ i arg min λ»f(x i λgradf(x i )) : λ ق 0 (if there is no such minimizer the algorithm fails) x i+1 = x i λ i gradf(x i ) i = i + 1 stop Notation: Given a real-valued function φ, the (possibly empty) set of global minimizers for the problem minimize φ(x) s.t. x S is denoted by arg min x»φ(x) : x S. f Exercise 3.4 Let x be such that gradf(x) θ (i.e., (x) θ). Show that ĥ := x gradf(x) (unit vector along gradf(x)) is indeed the unique direction of (local) gradf(x),gradf(x) 1/2 Copyright c , André L. Tits. All Rights Reserved 35

Notes for ENEE 664: Optimal Control. André L. Tits

Notes for ENEE 664: Optimal Control. André L. Tits Notes for ENEE 664: Optimal Control André L. Tits DRAFT July 2011 Contents 1 Motivation and Scope 3 1.1 Some Examples.................................. 3 1.2 Scope of the Course................................

More information

Notes for ENEE 664: Optimal Control

Notes for ENEE 664: Optimal Control Notes for ENEE 664: Optimal Control André L. Tits DRAFT September 2018 Copyright c 1993 2018, André L. Tits. All Rights Reserved 1 2 Copyright c 1993-2018, André L. Tits. All Rights Reserved Contents 1

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Balanced Truncation 1

Balanced Truncation 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ ME 234, Lyapunov and Riccati Problems. This problem is to recall some facts and formulae you already know. (a) Let A and B be matrices of appropriate dimension. Show that (A, B) is controllable if and

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 24: H2 Synthesis Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology May 4, 2011 E. Frazzoli (MIT) Lecture 24: H 2 Synthesis May

More information

EE363 homework 7 solutions

EE363 homework 7 solutions EE363 Prof. S. Boyd EE363 homework 7 solutions 1. Gain margin for a linear quadratic regulator. Let K be the optimal state feedback gain for the LQR problem with system ẋ = Ax + Bu, state cost matrix Q,

More information

Convexity of the Reachable Set of Nonlinear Systems under L 2 Bounded Controls

Convexity of the Reachable Set of Nonlinear Systems under L 2 Bounded Controls 1 1 Convexity of the Reachable Set of Nonlinear Systems under L 2 Bounded Controls B.T.Polyak Institute for Control Science, Moscow, Russia e-mail boris@ipu.rssi.ru Abstract Recently [1, 2] the new convexity

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Linear System Theory

Linear System Theory Linear System Theory Wonhee Kim Chapter 6: Controllability & Observability Chapter 7: Minimal Realizations May 2, 217 1 / 31 Recap State space equation Linear Algebra Solutions of LTI and LTV system Stability

More information

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems Robert M. Freund February 2016 c 2016 Massachusetts Institute of Technology. All rights reserved. 1 1 Introduction

More information

Theorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0).

Theorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0). Linear Systems Notes Lecture Proposition. A M n (R) is positive definite iff all nested minors are greater than or equal to zero. n Proof. ( ): Positive definite iff λ i >. Let det(a) = λj and H = {x D

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

3 Gramians and Balanced Realizations

3 Gramians and Balanced Realizations 3 Gramians and Balanced Realizations In this lecture, we use an optimization approach to find suitable realizations for truncation and singular perturbation of G. It turns out that the recommended realizations

More information

Solution of Linear State-space Systems

Solution of Linear State-space Systems Solution of Linear State-space Systems Homogeneous (u=0) LTV systems first Theorem (Peano-Baker series) The unique solution to x(t) = (t, )x 0 where The matrix function is given by is called the state

More information

Dichotomy, the Closed Range Theorem and Optimal Control

Dichotomy, the Closed Range Theorem and Optimal Control Dichotomy, the Closed Range Theorem and Optimal Control Pavel Brunovský (joint work with Mária Holecyová) Comenius University Bratislava, Slovakia Praha 13. 5. 2016 Brunovsky Praha 13. 5. 2016 Closed Range

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28 OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from

More information

Semidefinite Programming Duality and Linear Time-invariant Systems

Semidefinite Programming Duality and Linear Time-invariant Systems Semidefinite Programming Duality and Linear Time-invariant Systems Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 2 July 2004 Workshop on Linear Matrix Inequalities in Control LAAS-CNRS,

More information

Differential Equations Preliminary Examination

Differential Equations Preliminary Examination Differential Equations Preliminary Examination Department of Mathematics University of Utah Salt Lake City, Utah 84112 August 2007 Instructions This examination consists of two parts, called Part A and

More information

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers. Chapter 3 Duality in Banach Space Modern optimization theory largely centers around the interplay of a normed vector space and its corresponding dual. The notion of duality is important for the following

More information

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Module 07 Controllability and Controller Design of Dynamical LTI Systems Module 07 Controllability and Controller Design of Dynamical LTI Systems Ahmad F. Taha EE 5143: Linear Systems and Control Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ataha October

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

1 The Observability Canonical Form

1 The Observability Canonical Form NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)

More information

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)

Subject: Optimal Control Assignment-1 (Related to Lecture notes 1-10) Subject: Optimal Control Assignment- (Related to Lecture notes -). Design a oil mug, shown in fig., to hold as much oil possible. The height and radius of the mug should not be more than 6cm. The mug must

More information

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory

More information

w T 1 w T 2. w T n 0 if i j 1 if i = j

w T 1 w T 2. w T n 0 if i j 1 if i = j Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible -

More information

10 Transfer Matrix Models

10 Transfer Matrix Models MIT EECS 6.241 (FALL 26) LECTURE NOTES BY A. MEGRETSKI 1 Transfer Matrix Models So far, transfer matrices were introduced for finite order state space LTI models, in which case they serve as an important

More information

Topic # Feedback Control Systems

Topic # Feedback Control Systems Topic #17 16.31 Feedback Control Systems Deterministic LQR Optimal control and the Riccati equation Weight Selection Fall 2007 16.31 17 1 Linear Quadratic Regulator (LQR) Have seen the solutions to the

More information

Chapter III. Stability of Linear Systems

Chapter III. Stability of Linear Systems 1 Chapter III Stability of Linear Systems 1. Stability and state transition matrix 2. Time-varying (non-autonomous) systems 3. Time-invariant systems 1 STABILITY AND STATE TRANSITION MATRIX 2 In this chapter,

More information

The first order quasi-linear PDEs

The first order quasi-linear PDEs Chapter 2 The first order quasi-linear PDEs The first order quasi-linear PDEs have the following general form: F (x, u, Du) = 0, (2.1) where x = (x 1, x 2,, x 3 ) R n, u = u(x), Du is the gradient of u.

More information

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University. Lecture 4 Chapter 4: Lyapunov Stability Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 4 p. 1/86 Autonomous Systems Consider the autonomous system ẋ

More information

Lecture Notes of EE 714

Lecture Notes of EE 714 Lecture Notes of EE 714 Lecture 1 Motivation Systems theory that we have studied so far deals with the notion of specified input and output spaces. But there are systems which do not have a clear demarcation

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

1. The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ = Ax + Bu is

1. The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ = Ax + Bu is ECE 55, Fall 2007 Problem Set #4 Solution The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ Ax + Bu is x(t) e A(t ) x( ) + e A(t τ) Bu(τ)dτ () This formula is extremely important

More information

6. Linear Quadratic Regulator Control

6. Linear Quadratic Regulator Control EE635 - Control System Theory 6. Linear Quadratic Regulator Control Jitkomut Songsiri algebraic Riccati Equation (ARE) infinite-time LQR (continuous) Hamiltonian matrix gain margin of LQR 6-1 Algebraic

More information

Module 03 Linear Systems Theory: Necessary Background

Module 03 Linear Systems Theory: Necessary Background Module 03 Linear Systems Theory: Necessary Background Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html September

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

CHAPTER V DUAL SPACES

CHAPTER V DUAL SPACES CHAPTER V DUAL SPACES DEFINITION Let (X, T ) be a (real) locally convex topological vector space. By the dual space X, or (X, T ), of X we mean the set of all continuous linear functionals on X. By the

More information

Chap. 3. Controlled Systems, Controllability

Chap. 3. Controlled Systems, Controllability Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :

More information

Controllability, Observability, Full State Feedback, Observer Based Control

Controllability, Observability, Full State Feedback, Observer Based Control Multivariable Control Lecture 4 Controllability, Observability, Full State Feedback, Observer Based Control John T. Wen September 13, 24 Ref: 3.2-3.4 of Text Controllability ẋ = Ax + Bu; x() = x. At time

More information

Math Ordinary Differential Equations

Math Ordinary Differential Equations Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

More information

LINEAR-QUADRATIC OPTIMAL CONTROL OF DIFFERENTIAL-ALGEBRAIC SYSTEMS: THE INFINITE TIME HORIZON PROBLEM WITH ZERO TERMINAL STATE

LINEAR-QUADRATIC OPTIMAL CONTROL OF DIFFERENTIAL-ALGEBRAIC SYSTEMS: THE INFINITE TIME HORIZON PROBLEM WITH ZERO TERMINAL STATE LINEAR-QUADRATIC OPTIMAL CONTROL OF DIFFERENTIAL-ALGEBRAIC SYSTEMS: THE INFINITE TIME HORIZON PROBLEM WITH ZERO TERMINAL STATE TIMO REIS AND MATTHIAS VOIGT, Abstract. In this work we revisit the linear-quadratic

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

Zeros and zero dynamics

Zeros and zero dynamics CHAPTER 4 Zeros and zero dynamics 41 Zero dynamics for SISO systems Consider a linear system defined by a strictly proper scalar transfer function that does not have any common zero and pole: g(s) =α p(s)

More information

Partial Differential Equations

Partial Differential Equations Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,

More information

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri C. Melchiorri (DEI) Automatic Control & System Theory 1 AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS Claudio Melchiorri Dipartimento di Ingegneria dell Energia Elettrica e dell Informazione (DEI)

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Model reduction for linear systems by balancing

Model reduction for linear systems by balancing Model reduction for linear systems by balancing Bart Besselink Jan C. Willems Center for Systems and Control Johann Bernoulli Institute for Mathematics and Computer Science University of Groningen, Groningen,

More information

A brief introduction to ordinary differential equations

A brief introduction to ordinary differential equations Chapter 1 A brief introduction to ordinary differential equations 1.1 Introduction An ordinary differential equation (ode) is an equation that relates a function of one variable, y(t), with its derivative(s)

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

ME Fall 2001, Fall 2002, Spring I/O Stability. Preliminaries: Vector and function norms

ME Fall 2001, Fall 2002, Spring I/O Stability. Preliminaries: Vector and function norms I/O Stability Preliminaries: Vector and function norms 1. Sup norms are used for vectors for simplicity: x = max i x i. Other norms are also okay 2. Induced matrix norms: let A R n n, (i stands for induced)

More information

2. Review of Linear Algebra

2. Review of Linear Algebra 2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear

More information

Iterative Solution of a Matrix Riccati Equation Arising in Stochastic Control

Iterative Solution of a Matrix Riccati Equation Arising in Stochastic Control Iterative Solution of a Matrix Riccati Equation Arising in Stochastic Control Chun-Hua Guo Dedicated to Peter Lancaster on the occasion of his 70th birthday We consider iterative methods for finding the

More information

ECEEN 5448 Fall 2011 Homework #5 Solutions

ECEEN 5448 Fall 2011 Homework #5 Solutions ECEEN 5448 Fall 211 Homework #5 Solutions Professor David G. Meyer December 8, 211 1. Consider the 1-dimensional time-varying linear system ẋ t (u x) (a) Find the state-transition matrix, Φ(t, τ). Here

More information

1 Continuous-time Systems

1 Continuous-time Systems Observability Completely controllable systems can be restructured by means of state feedback to have many desirable properties. But what if the state is not available for feedback? What if only the output

More information

Generalized Riccati Equations Arising in Stochastic Games

Generalized Riccati Equations Arising in Stochastic Games Generalized Riccati Equations Arising in Stochastic Games Michael McAsey a a Department of Mathematics Bradley University Peoria IL 61625 USA mcasey@bradley.edu Libin Mou b b Department of Mathematics

More information

Reflected Brownian Motion

Reflected Brownian Motion Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide

More information

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced

More information

EE363 homework 8 solutions

EE363 homework 8 solutions EE363 Prof. S. Boyd EE363 homework 8 solutions 1. Lyapunov condition for passivity. The system described by ẋ = f(x, u), y = g(x), x() =, with u(t), y(t) R m, is said to be passive if t u(τ) T y(τ) dτ

More information

LINEAR-CONVEX CONTROL AND DUALITY

LINEAR-CONVEX CONTROL AND DUALITY 1 LINEAR-CONVEX CONTROL AND DUALITY R.T. Rockafellar Department of Mathematics, University of Washington Seattle, WA 98195-4350, USA Email: rtr@math.washington.edu R. Goebel 3518 NE 42 St., Seattle, WA

More information

Algebra II. Paulius Drungilas and Jonas Jankauskas

Algebra II. Paulius Drungilas and Jonas Jankauskas Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive

More information

Introduction to Nonlinear Control Lecture # 4 Passivity

Introduction to Nonlinear Control Lecture # 4 Passivity p. 1/6 Introduction to Nonlinear Control Lecture # 4 Passivity È p. 2/6 Memoryless Functions ¹ y È Ý Ù È È È È u (b) µ power inflow = uy Resistor is passive if uy 0 p. 3/6 y y y u u u (a) (b) (c) Passive

More information

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Optimal Control Ömer Özak SMU Macroeconomics II Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Review of the Theory of Optimal Control Section 1 Review of the Theory of Optimal Control Ömer

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7)

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7) EEE582 Topical Outline A.A. Rodriguez Fall 2007 GWC 352, 965-3712 The following represents a detailed topical outline of the course. It attempts to highlight most of the key concepts to be covered and

More information

The norms can also be characterized in terms of Riccati inequalities.

The norms can also be characterized in terms of Riccati inequalities. 9 Analysis of stability and H norms Consider the causal, linear, time-invariant system ẋ(t = Ax(t + Bu(t y(t = Cx(t Denote the transfer function G(s := C (si A 1 B. Theorem 85 The following statements

More information

Linear Quadratic Zero-Sum Two-Person Differential Games

Linear Quadratic Zero-Sum Two-Person Differential Games Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,

More information

Problem 2 (Gaussian Elimination, Fundamental Spaces, Least Squares, Minimum Norm) Consider the following linear algebraic system of equations:

Problem 2 (Gaussian Elimination, Fundamental Spaces, Least Squares, Minimum Norm) Consider the following linear algebraic system of equations: EEE58 Exam, Fall 6 AA Rodriguez Rules: Closed notes/books, No calculators permitted, open minds GWC 35, 965-37 Problem (Dynamic Augmentation: State Space Representation) Consider a dynamical system consisting

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

Uniformly Uniformly-ergodic Markov chains and BSDEs

Uniformly Uniformly-ergodic Markov chains and BSDEs Uniformly Uniformly-ergodic Markov chains and BSDEs Samuel N. Cohen Mathematical Institute, University of Oxford (Based on joint work with Ying Hu, Robert Elliott, Lukas Szpruch) Centre Henri Lebesgue,

More information

Nonlinear Optimization

Nonlinear Optimization Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

More information

Problem Set 5 Solutions 1

Problem Set 5 Solutions 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski Problem Set 5 Solutions The problem set deals with Hankel

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.4: Dynamic Systems Spring Homework Solutions Exercise 3. a) We are given the single input LTI system: [

More information

Criterions on periodic feedback stabilization for some evolution equations

Criterions on periodic feedback stabilization for some evolution equations Criterions on periodic feedback stabilization for some evolution equations School of Mathematics and Statistics, Wuhan University, P. R. China (Joint work with Yashan Xu, Fudan University) Toulouse, June,

More information

On Controllability of Linear Systems 1

On Controllability of Linear Systems 1 On Controllability of Linear Systems 1 M.T.Nair Department of Mathematics, IIT Madras Abstract In this article we discuss some issues related to the observability and controllability of linear systems.

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 5. Input-Output Stability DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Input-Output Stability y = Hu H denotes

More information

Control, Stabilization and Numerics for Partial Differential Equations

Control, Stabilization and Numerics for Partial Differential Equations Paris-Sud, Orsay, December 06 Control, Stabilization and Numerics for Partial Differential Equations Enrique Zuazua Universidad Autónoma 28049 Madrid, Spain enrique.zuazua@uam.es http://www.uam.es/enrique.zuazua

More information

Modeling and Analysis of Dynamic Systems

Modeling and Analysis of Dynamic Systems Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability

More information

CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73

CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73 CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS p. 1/73 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS Mixed Inequality Constraints: Inequality constraints involving control and possibly

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

EE363 homework 2 solutions

EE363 homework 2 solutions EE363 Prof. S. Boyd EE363 homework 2 solutions. Derivative of matrix inverse. Suppose that X : R R n n, and that X(t is invertible. Show that ( d d dt X(t = X(t dt X(t X(t. Hint: differentiate X(tX(t =

More information

Steady State Kalman Filter

Steady State Kalman Filter Steady State Kalman Filter Infinite Horizon LQ Control: ẋ = Ax + Bu R positive definite, Q = Q T 2Q 1 2. (A, B) stabilizable, (A, Q 1 2) detectable. Solve for the positive (semi-) definite P in the ARE:

More information

Theory of Ordinary Differential Equations

Theory of Ordinary Differential Equations Theory of Ordinary Differential Equations Existence, Uniqueness and Stability Jishan Hu and Wei-Ping Li Department of Mathematics The Hong Kong University of Science and Technology ii Copyright c 24 by

More information

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 Ali Jadbabaie, Claudio De Persis, and Tae-Woong Yoon 2 Department of Electrical Engineering

More information

Lyapunov Stability Theory

Lyapunov Stability Theory Lyapunov Stability Theory Peter Al Hokayem and Eduardo Gallestey March 16, 2015 1 Introduction In this lecture we consider the stability of equilibrium points of autonomous nonlinear systems, both in continuous

More information

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books. Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define

More information

1 Controllability and Observability

1 Controllability and Observability 1 Controllability and Observability 1.1 Linear Time-Invariant (LTI) Systems State-space: Dimensions: Notation Transfer function: ẋ = Ax+Bu, x() = x, y = Cx+Du. x R n, u R m, y R p. Note that H(s) is always

More information

L2 gains and system approximation quality 1

L2 gains and system approximation quality 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 24: MODEL REDUCTION L2 gains and system approximation quality 1 This lecture discusses the utility

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Newtonian Mechanics. Chapter Classical space-time

Newtonian Mechanics. Chapter Classical space-time Chapter 1 Newtonian Mechanics In these notes classical mechanics will be viewed as a mathematical model for the description of physical systems consisting of a certain (generally finite) number of particles

More information

Nonlinear Control. Nonlinear Control Lecture # 6 Passivity and Input-Output Stability

Nonlinear Control. Nonlinear Control Lecture # 6 Passivity and Input-Output Stability Nonlinear Control Lecture # 6 Passivity and Input-Output Stability Passivity: Memoryless Functions y y y u u u (a) (b) (c) Passive Passive Not passive y = h(t,u), h [0, ] Vector case: y = h(t,u), h T =

More information

Mathematics 530. Practice Problems. n + 1 }

Mathematics 530. Practice Problems. n + 1 } Department of Mathematical Sciences University of Delaware Prof. T. Angell October 19, 2015 Mathematics 530 Practice Problems 1. Recall that an indifference relation on a partially ordered set is defined

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

Observer design for a general class of triangular systems

Observer design for a general class of triangular systems 1st International Symposium on Mathematical Theory of Networks and Systems July 7-11, 014. Observer design for a general class of triangular systems Dimitris Boskos 1 John Tsinias Abstract The paper deals

More information

Geometric Optimal Control with Applications

Geometric Optimal Control with Applications Geometric Optimal Control with Applications Accelerated Graduate Course Institute of Mathematics for Industry, Kyushu University, Bernard Bonnard Inria Sophia Antipolis et Institut de Mathématiques de

More information

NOTES ON LINEAR ODES

NOTES ON LINEAR ODES NOTES ON LINEAR ODES JONATHAN LUK We can now use all the discussions we had on linear algebra to study linear ODEs Most of this material appears in the textbook in 21, 22, 23, 26 As always, this is a preliminary

More information