Introduction to Linear Matrix Inequalities (LMIs)

Size: px
Start display at page:

Download "Introduction to Linear Matrix Inequalities (LMIs)"

Transcription

1 ECE 680 Fall 2017 Introduction to Linear Matrix Inequalities (LMIs) by Stanislaw H. Żak and Guisheng Zhai October 18, 2017 Henry Ford offered customers the Model T Ford in any color the customer wants, as long as it is black. Henry Ford s policy resulted from subordination to the assembly line constraint that could not keep up with market demand. This was a resource constraint. To maximize throughput it was important for Ford not to change colors during production. This policy assured maximum profit and value for Ford by subordinating the whole system (including customers) to production. Henry Ford s failure was that he stuck to his policy even when new competitors entered the market; then the resource constraint became a market constraint resulting in a steep drop in sales and market share. The right step would have been to realize that there was now a market constraint and subordinate the system to the constraint. Competition and customers needs should have driven Ford to produce a variety of cars: different styles, colors, or amenities. Ronen and Pass 3, p Motivation A number of control design problems for linear and nonlinear systems can be formulated as convex optimization problems. To define a convex optimization problem, we need definitions of a convex set and a convex function. Definition 1 A set Ω R n is convex if for any x and y in Ω, the line segment between x and y lies in Ω, that is, αx + (1 α)y Ω for any α (0, 1). 1

2 Examples of convex sets include the empty set, R n, a subspace, a line, a line segment. Definition 2 A real-valued function f : Ω R defined on a convex set Ω R n is convex if for all x, y Ω and all α (0, 1), f(αx + (1 α)y) αf(x) + (1 α)f(y). Thus a real-valued function in n variables is convex over a convex set if for all x, y Ω the points on the line segment connecting (x, f(x)) and (y, f(y)) lie on or above the graph of f. For more information on the subject of convex function, we recommend 2, Section Definition 3 A convex optimization problem is the one where the objective function to be minimized is convex and the constraint set, over which we minimize the objective function, is a convex set. If f is a convex function and Ω is a convex set, then, in general, max f(x) subject to x Ω is NOT a convex optimization problem! However, if f is convex then the above problem is a convex optimization problem. 2 Linear Matrix Inequality and Its Properties Consider n + 1 real symmetric matrices F i = F i R m m, i = 0, 1,..., n and a vector x = x 1 x 2 x n R n. 2

3 Then, F (x) = F 0 + x 1 F x n F n n = F 0 + x i F i i=1 is an affine function of x rather than a linear function of x because F (x) is composed of a linear term, n i=1 x if i, and a constant term, F 0. Consider now an expression of the form F (x) = F 0 + x 1 F x n F n 0. The above is to be interpreted as follows: find a set of vectors x such that z F (x)z 0 for all z R m, that is, F (x) is positive semi-definite. Recall that F i s are constant matrices, x is unknown, and F (x) = F (x) is an affine function of x. The expression, F (x) = F 0 + x 1 F x n F n 0, is referred to as in the literature as the linear matrix inequality (LMI) although the term the affine matrix inequality would be a correct term to use. It is easy to verify that the set, {x : F (x) 0}, is a convex set. A system of LMIs, F 1 (x) 0, F 2 (x) 0,..., F k (x) 0, can be represented as one single LMI, F 1 (x) F (x) = F 2 (x)... F k (x) 0. A linear matrix inequality, involving an m-by-n constant matrix A, of the form, Ax b can be represented as m LMIs b i a i x 0, i = 1, 2,..., m, 3

4 where a i is the i-th row of the matrix A. We can view each scalar inequality as an LMI. We then represent m LMIs as one LMI, b 1 a 1 x b F (x) = 2 a 2 x b m a mx A semidefinite programming, problem is a convex optimization problem of the form minimize c x subject to F (x) 0. Note that the linear objective function c x is a convex function and the constraint set, {x : F (x) 0}, is a convex set. The matrix property that we discuss next is useful when converting LMIs into equivalent LMIs or converting some nonlinear matrix inequalities into linear matrix inequalities. We start with a simple observation. Lemma 1 Let P = P be a nonsingular n-by-n matrix and let x = Mz, where M R n n such that det M 0. Then, x P x 0 if and only if z M P Mz 0, that is, Similarly P 0 if and only if M P M 0. P 0 if and only if M P M 0. Suppose that we have a square block matrix A B B D, where A = A and D = D. Then, by Lemma 1, A B O I A B O I B D 0 if and only if I O B D I O 0, 4

5 where I is an identity matrix of appropriate dimension. In other words, A B D B 0 if and only if 0. B D B A Consider next a square block matrix of the form A11 A 12, A 21 A 22 where A 11 and A 22 are square and symmetric submatrices, and A 12 = A 21. Suppose that the matrix A 11 is invertible. Then, I O A11 A 21 I A 1 11 A 21 A 21 A 1 11 I A 21 A 22 O I Let = 11 = A 22 A 21 A 1 11 A 21. A11 O O A 22 A 21 A 1 11 A 21 The matrix 11 is called the Schur complement of A 11. Hence, by Lemma 1, A11 A 21 A11 O 0 if and only if 0, A 21 A 22 O 11 that is, A11 A 21 A 21 A 22 0 if and only if A 11 0 and Many problems of optimization, control design, and signal processing can be formulated in terms of LMIs. To test whether or not there exists a solution x to F (x) 0 is called a feasibility problem. We say that the LMI is non-feasible if no solution exists. Remark 1 Any feasible non-strict LMI can be reduced to an equivalent LMI that is strictly feasible, by eliminating implicit equality constraints and then reducing the resulting LMI by removing any constant nullspace see 1, Section Example 1 We present an example illustrating the LMI feasibility problem. It is well known that a constant square matrix A R n n has its eigenvalues in the open left half-complex 5.

6 plane if and only if for any real, symmetric, positive definite Q R n n, the solution P = P to the Lyapunov matrix equation A P + P A = Q, is positive definite. That is, the real parts of the eigenvalues of A are all negative if and only if there exists a real symmetric positive definite matrix P such that A P + P A 0, or equivalently, A P P A 0. Thus, the location of all eigenvalues of A in the open left half-complex plane is equivalent to feasibility of the following LMI, P O O A P P A that is, the existence of a symmetric positive definite P such that the above inequality holds. Finding P = P 0 such that A P + P A 0 can be reduced to solving the LMI x n x 2n 1 x q 0, feasibility problem. Indeed, let x 1 x 2 x n x P = 2 x n+1 x 2n 1.., where n(n + 1) q =. 2 We select the following basis matrices, P 1 =.., P =..,..., P q = Note that P i s are symmetric and have only non-zero elements corresponding to x i in P. Any symmetric matrix P can be expressed as a linear combination of the basis matrices. Let F i = A P i P i A, i = 1, 2,..., q. 6

7 We then write A P + P A = ( x 1 A P 1 + P 1 A ) ( + x 2 A P 2 + P 2 A ) ( + + x q A P q + P q A ) = x 1 F 1 x 2 F 2 x q F q 0. Let then if and only if F (x) = x 1 F 1 + x 2 F x q F q, P = P 0 and A P + P A 0 F (x) 0. More examples of linear matrix inequalities in system and control can be found in the book by Boyd et al LMI Solvers The LMI, F (x) = F 0 + x 1 F x n F n 0 is called the canonical representation of an LMI. The LMIs in the canonical form are inefficient from a storage view-point as well as from the efficiency of the LMI solvers view-point. Modern LMI solvers use a structured representation of LMIs that we discuss next. One can use MATLAB s toolbox to efficiently solve LMIs. This toolbox has three types of LMI solvers. 3.1 The Feasibility Problem Solver This solver computes a feasible solution, that is, it solves the feasible problem defined by a given system of LMI constraints. Using this solver, we can solve any system of LMIs of the form N L(X 1,..., X k )N M R(X 1,..., X k )M, 7

8 where X 1,..., X k are matrix variables, N is the left outer factor, M is the right outer factor, L(X 1,..., X k ) is the left inner factor and R(X 1,..., X k ) is the right inner factor. The matrices L( ) and R( ) are, in general, symmetric block matrices. We note that the term left-hand side refers to what is on the smaller side of the inequality 0 X. Thus in X 0, the matrix X is still on the right-hand side because it is on the larger side of the inequality. We next provide a description of an approach that can be used to solve the given LMI system feasibility problem. To initialize the LMI system description, we type setlmis(). Then we declare matrix variables using the command lmivar. The command lmiterm allows us to specify LMIs that constitute the LMI system under consideration. Next, we need to obtain an internal representation using the command getlmis. We next compute a feasible solution to the given LMI system using the command feasp. After that we extract matrix variable values with the command dec2mat. In summary, a general structure of a MATLAB program for finding a feasible solution to the set of LMIs has the form, setlmis() lmivar lmiterm... lmiterm getlmis feasp dec2mat We now analyze the above commands in some detail so that the reader can write MATLAB programs for solving LMIs after finishing reading this section. First, to create a new matrixvalued variable, say, X, in the given LMI system, we use the command X=lmivar(type,structure) The input type specifies the structure of the variable X. There may be three structures of matrix variables. When type=1, we have a symmetric block diagonal matrix variable. The input type=2 refers to a full rectangular matrix variable. Finally, type=3 refers to other 8

9 cases. The second input structure gives additional information on the structure of the matrix variable X. For example, the matrix variable X could have the form D 1 O O O D X = 2 O....., O O D r where each D i is a square symmetric matrix. For the above example we would use type=1. The above matrix variable has r blocks. The input structure is then an r 2 matrix whose i-th row describes the i-th block, where the first component of each row gives the corresponding block size, while the second element of each row specifies the block type. For example, X=lmivar(1,3 1) specifies a full symmetric 3 3 matrix variable. On the other hand, X=lmivar(2,2 3) specifies a rectangular 2 3 matrix variable. Finally, the matrix variable S of the form s s S =, 0 0 s 2 s s 3 s 4 can be declared as follows: S=lmivar(1,2 0;2 1) Note that in the above the second component of the first row of the second input has the value of zero, that is, structure(1,2)=0. This describes a scalar block matrix of the form D 1 = s 1 I 2. Note that the second block is a 2 2 symmetric full block. The purpose of the next command is to specify the terms of the LMI system of interest. This command has the form 9

10 lmiterm(termid,a,b,flag) We briefly describe each of the four inputs of this command. The first input, termid, is a row with four elements that specify the terms of each LMI of the LMI system. We have termid(1)=n to specify the left-hand side of the n-th LMI. We use termid(1)=-n to specify the right-hand side of the n-th LMI. The middle two elements of the input termid specify the block location. Thus termid(2,3)=i j refers to the term that belongs to the (i, j) block of the LMI specified by the first component. Finally, termid(4)=0 for the constant term, termid(4)=x for the variable term in the form AXB, while termid(4)=-x for the variable term in the form AX B. The second and the third inputs of the command lmiterm give the values of the left and right outer factors, that is, A and B give the value of the constant outer factors in the variable terms, AXB or in AX B. Finally, the fourth input to lmiterm serves as a compact way to specify the expression AXB + (AXB). Thus flag= s can be used to denote a symmetrized expression. We illustrate the above command on the following LMI, P A + A P 0. (1) We have one LMI with two terms. We could use the following description of this single LMI, lmiterm(1 1 1 P,1,A) lmiterm( P,A,1) Because P A + A P = P A + (P A), we can compactly describe (1) with the use of the flag as follows, lmiterm(1 1 1 P,1,A, s ) Now, to solve the feasibility problem we could have typed tmin,xfeas=feasp(lmis) In general, for a given LMI feasibility problem of the form find such that x L(x) R(x) 10

11 the command feasp solves the auxiliary convex problem minimize subject to t L(x) R(x) + ti. The system of LMIs is feasible if the minimal t is negative. We add that the current value of t is displayed by feasp at each iteration. Finally, we convert the output of the LMI solver into matrix variables using the command P=dec2mat(lmis,xfeas,P). 3.2 Minimizing Linear Objective Under LMI Constraints This solver is invoked using the function mincx. It solves the convex problem minimize subject to c x A(x) B(x). The notation A(x) B(x) is a shorthand notation for general structured LMI systems. Thus to solve a mincx problem, in addition to specifying the LMI constraints as in the feasp problem, we also declare the linear objective function. Then we invoke the function mincx. We illustrate both the feasp and mincx solvers in the following example. Example 2 We consider the optimization problem minimize c x subject to Ax b, where c = A = 4 5, , b = We first solve the feasibility problem, that is, we find an x such that Ax b, using the feasp solver. After that we solve the above minimization problem using the mincx solver. A simple MATLAB code accomplishing the above tasks is shown below. 11

12 % Enter problem data A =1 1;1 3;2 1; b= '; c= 4 5'; setlmis(); X=lmivar(2,2 1); lmiterm(1 1 1 X,A(1,:),1); lmiterm( , b(1)); lmiterm(1 2 2 X,A(2,:),1); lmiterm( , b(2)); lmiterm(1 3 3 X,A(3,:),1); lmiterm( , b(3)); lmis=getlmis; % disp(' feasp result ') tmin,xfeas=feasp(lmis); x feasp=dec2mat(lmis,xfeas,x) disp(' mincx result ') objective,x mincx=mincx(lmis,c, ) The feasp produces x feasp = The mincx produces x mincx = In the next example we discuss the function defcx that we can use to construct the vector c used by the LMI solver mincx. Example 3 Suppose that we wish to solve the optimization problem minimize trace(p ) subject to A P + P A 0, where P = P 0. 12

13 We can use the function mincx to solve the above problem. However, to use the function mincx, we need a vector c such that c x = trace(p ). After specifying the LMIs and obtaining their internal representation using, for example, the command lmisys=getlmis, we can obtain the desired c with the following MATLAB code, q=decnbr(lmisys); c=zeros(q,1); for j=1:q Pj=defcx(lmisys,j,P); c(j)=trace(pj); end Having obtained the vector c, we can use the function mincx to solve the optimization problem. 3.3 Generalized Eigenvalue Minimization Problem This problem can be stated as minimize subject to λ C(x) D(x) 0 B(x) A(x) λb(x). In the above, we need to distinguish between standard LMI constraints of the form C(x) D(x) and the linear-fractional LMIs of the form A(x) λb(x) that are concerned with the generalized eigenvalue λ, that is, the LMIs involving λ. The number of linear-fractional constraints is specified with nflc. The generalized eigenvalue minimization problem under LMI constraints is solved by calling the solver gevp. The basic structure of the gevp solver has the form, 13

14 lopt,xopt=gevp{lmisys,nflc} It returns lopt, which is the global minimum of the generalized eigenvalue, and xopt, which is the optimal decision vector variable. The argument lmisys is the system of LMIs, C(x) D(x), 0 B(x), and A(x) λb(x) for λ = 1. As in the previous solvers, the corresponding optimal values of the matrix variables are obtained using dec2mat. There are other inputs to gevp but they are optional. For more information on this type of the LMI solver, we refer to the LMI Lab in the MATLAB s Robust Control Toolbox user s guide. Example 4 Consider a system model of the form ẋ = Ax, x(0) = x 0, where the matrix A R n n is asymptotically stable. We wish to estimate the decay rate of the system s trajectory x(t), that is, we wish to find constants η > 0 and M > 0 such that x(t) e ηt M(x 0 ). (2) Because A is asymptotically stable, by Lyapunov s theorem, for any Q = Q 0 there exists P = P 0 such that A P + P A = Q, that is, Let Then, we have x ( A P + P A ) x = x Qx. V = x P x. V = x ( A P + P A ) x, which is the Lyapunov derivative of V evaluated on trajectories of ẋ = Ax. Let ( α = min V ). (3) V Then, we have α V V and since V = x P x 0 for x 0, we obtain V αv. (4) 14

15 Therefore, V (t) e αt V (0). (5) We refer to α as the decay rate of V. We represent (5) as x(t) P x(t) e αt x 0 P x 0. (6) We have λ min (P ) x 2 2 x P x. (7) Because P = P 0, we have that λ min (P ) > 0. Combining (6) and (7), and dividing both sides by λ min (P ) > 0 gives Hence, x(t) 2 e αt x 0 P x 0 λ min (P ). x(t) e α 2 t x 0 P x 0 λ min (P ). (8) We represent P = P as P = P 1/2 P 1/2. Hence, Taking (9) into account, we represent (8) as x 0 P x 0 = x 0 P 1/2 P 1/2 x 0 = P 1/2 x 0 2. (9) x(t) e α 2 t P 1/2 x 0 λmin (P ). (10) Comparing the above with (2) yields η = α/2 and M(x 0 ) = P 1/2 x 0 λmin (P ). In terms of LMIs, finding α that satisfies (4) is equivalent to minimizing α subject to P 0 A P + P A αp. 15

16 For example if A = , then finding α that satisfies (3) can be accomplished using the following LMIs: A = ; setlmis(); P = lmivar(1,3 1) lmiterm( P,1,1) lmiterm( ,.01) lmiterm(2 1 1 P,1,A,'s') lmiterm( P,1,1) lmis = getlmis; gamma,p opt=gevp(lmis,1); P=dec2mat(lmis,P opt,p) alpha= gamma The result is Recall that η = α/2. % P % P > 0.01*I α = and P = % linear fractional constraint left hand side % linear fractional constraint right hand side Notes A quick introduction to the MATLAB s LMI toolbox is the tutorial that can be accessed by typing in lmidem. It is very instructive and we recommend to try it. We also recommend a tutorial on the mathematical theory and process control applications of linear matrix inequalities (LMIs) and bilinear matrix inequalities (BMIs) by VanAntwerp and Braatz 4. 16

17 We mention that in addition to the MATLAB s LMI toolbox there is another toolbox for solving LMIs called LMITOOL, a built in software package in Scilab toolbox, developed at INRIA in France. Scilab offers free software for numerical optimization. There is a version of LMITOOL for MATLAB that can be uploaded from the website of the Scilab Consortium. Yet Another LMI Parser (YALMIP) for solving LMIs was developed in Switzerland in the Automatic Control Laboratory at ETH. YALMIP is an intuitive and flexible modelling language for solving optimization problems in MATLAB. YALMIP supports linear programming (LP), quadratic programming (QP), second order cone programming (SOCP), semidefinite programming, determinant maximization, mixed integer programming, posynomial geometric programming, semidefinite programs with bilinear matrix inequalities (BMI), and multiparametric linear and quadratic programming. A very popular software package for solving LMIs, called CVX, was developed by the group of Stephen Boyd at Stanford. It runs in MATLAB. The CVX example library has hundreds of examples. 5 Practice Problems PP 1 Show that the set, Ω = { x : x P x 1 } is a convex set, where P is a positive definite matrix. PP 2 1. Show that the function, f 1 (x) = x P x, is a convex function, where P is a positive definite matrix; 2. Show that the function, f 2 (x) = x x 1 x 2 +3x 2 2 5x 1 +6x 2 +10, is a convex function. PP 3 Show that if f(x) is a convex function, then Ω = {x : f(x) c} is a convex set for any real number c. PP 4 Let D and E be given real matrices of appropriate dimensions and let F satisfy F F I. Show that the following matrix inequality holds for any ɛ < 0. DF E + E F D ɛdd + 1 ɛ E E (11) PP 5 Given a positive scalar α, derive an LMI condition under which the real parts of the eigenvalues of a real matrix A R n n are smaller than α. 17

18 PP 6 Establish an LMI condition for asymptotic stability of the uncertain linear timeinvariant system, ẋ = A(λ)x, where { p } p A(λ) λ i A i : λ i 0, λ i = 1, i=1 i=1 and each individual A i is asymptotically stable. Use a common Lyapunov function approach. PP 7 Investigate asymptotic stability of the uncertain linear time-varying system ẋ = A + DF (t)e x using a quadratic Lyapunov function candidate V (x) = x P x, where A R n n, D R n m and E R r n are constant matrices, while F (t) R m r models uncertainties such that F (t)f (t) γ 2 I r, γ > 0. PP 8 Given a real matrix A R n n, construct a set of LMIs that determine whether or not its eigenvalues satisfy α < Reλ(A) < β, where α < β are two real scalars. Organize your computations using the obtained LMIs in the form of a MATLAB program PP 9 The problem of designing a stabilizing state feedback uk = Kxk for a discrete-time LTI system xk + 1 = Axk + Buk, can be reduced to finding a matrix P 0 such that (A BK) P (A BK) P 0, (12) which is a nonlinear matrix inequality (NMI) with respect to P 0 and K. Formulate LMIs that are equivalent to the above conditions. PP 10 Use LMIs obtained in Practice Problem PP 9 to compute a stabilizing state feedback for the linear discrete-time system, xk + 1 = xk + uk

19 PP 11 Consider the linear continuous-time system ẋ = Ax + Bw, x(0) = 0 z = Cx + Dw (13) where x R n is the state, w R q is the disturbance input, z R p is the controlled output, and A, B, C, D are constant matrices of appropriate dimensions. Show that if there exist P = P 0 and γ > 0, such that A P + P A P B C B P γi D C D γi then the system is asymptotically stable, and furthermore t 0 0, (14) t z(τ) z(τ) dτ γ 2 w(τ) w(τ) dτ, for all t 0. PP 12 For the interconnected linear continuous-time system, design a stabilizing decentralized state feedback 0 ẋ 1 = A 11 x 1 + A 12 x 2 + B 1 u 1 (15) ẋ 2 = A 21 x 1 + A 22 x 2 + B 2 u 2, (16) u 1 = K 1 x 1, u 2 = K 2 x 2 using a combined Lyapunov function candidate V (x) = x 1 P 1 x 1 + x 2 P 2 x 2, where P 1 and P 2 are positive definite matrices to be determined. Formulate the design problem in terms of LMIs. 6 Solutions to Practice Problems SPP 1 We will show that for any x and y in Ω, that is, x P x 1, y P y 1, we have αx + (1 α)y Ω for all α (0, 1). 19

20 Indeed, (αx + (1 α)y) P (αx + (1 α)y) = α 2 x P x + (1 α) 2 y P y + α(1 α)(x P y + y P x) α 2 + (1 α) 2 + α(1 α)(x P y + y P x). Taking into account that we obtain (x y) P (x y) 0, x P y + y P x x P x + y P y 2, and thus (αx + (1 α)y) P (αx + (1 α)y) α 2 + (1 α) 2 + 2α(1 α) = (α + (1 α)) 2 = 1, which implies αx + (1 α)y Ω. The proof is complete. SPP 2 1. First, we note that f 1 (x) is a real-valued function f 1 : R n R defined on R n, a convex set. Then, for any x, y R n and any α (0, 1), we obtain f 1 (αx + (1 α)y) = (αx + (1 α)y) P (αx + (1 α)y) = α 2 x P x + (1 α) 2 y P y + α(1 α)(x P y + y P x). Taking into account that (x y) P (x y) 0, we obtain x P y + y P x x P x + y P y = f 1 (x) + f 1 (y), and hence f 1 (αx + (1 α)y) α 2 f 1 (x) + (1 α) 2 f 1 (y) + α(1 α)(f 1 (x) + f 1 (y)) = αf 1 (x) + (1 α)f 1 (y). Therefore, f 1 (x) is a convex function. 20

21 2. We rewrite f 2 (x) as f 2 (x) = x P x + c x + 10 where 1 1 P =, c = Since P is positive definite, the function x P x in f 2 (x) is convex. The function c x + 10 is affine and thus convex. The sum of two convex functions is also convex. Therefore, f 2 (x) is convex. SPP 3 Given any x, y Ω, that is, f(x) c, f(y) c, we have f(αx + (1 α)y) αf(x) + (1 α)f(y) αc + (1 α)c = c for any α (0, 1). Thus, αx + (1 α)y Ω, which implies that Ω is a convex set. SPP 4 Using properties of symmetric nonnegative definite matrices, we obtain (ɛ 1 2 D ɛ 1 2 F E ) (ɛ 1 2 D ɛ 1 2 F E ) 0, which is equivalent to Since F F I, we obtain that DF E + E F D ɛdd + 1 ɛ E F F E. DF E + E F D ɛdd + 1 ɛ E E. SPP 5 The eigenvalues of A + αi, denoted λ i (A + αi), are λ i (A) + α. Therefore, the real parts of the eigenvalues of a real matrix A R n n are smaller than α is equivalent to the matrix A + αi being a Hurwitz matrix. This condition can be expressed in terms of an LMI for the existence of a positive definite matrix P such that (A + αi) P + P (A + αi) 0, or equivalently, P O O A P P A 2αP 0. 21

22 SPP 6 The system is asymptotically stable if we can find a symmetric P 0 such that A(λ) P + P A(λ) 0. (17) Note that if we can find a common P 0 such that A i P + P A i 0, i = 1, 2,..., p, (18) and then if we multiply each matrix inequality in (18) by a nonnegative scalar λ i and add them up, we obtain the matrix inequality (17). Therefore, the LMIs (18) together with P 0 constitute a sufficient condition for asymptotic stability of the uncertain system. SPP 7 The uncertain linear time-varying system ẋ = A + DF (t)e x is asymptotically stable if we can find symmetric P 0 such that (A + DF (t)e) P + P (A + DF (t)e) 0. (19) Using 0 ( P DF (t) E ) ( P DF (t) E ) = P DF (t)f (t) D P + E E E F (t) D P P DF (t)e γ 2 P DD P + E E E F (t) D P P DF (t)e, we obtain that E F (t) D P + P DF (t)e γ 2 P DD P + E E (20) holds for any F (t) satisfying F (t) F (t) γ 2 I r. Combining (19) and (20), we obtain that the uncertain system is asymptotically stable if P 0 and A P + P A + γ 2 P DD P + E E 0, which is equivalent to the LMI condition, A P + P A + E E P D D P γ 2 I m 0. The above equivalence can be established using the Schur complement. 22

23 SPP 8 Recall that the symbol R denotes the real part of a given complex number. Noticing that α < R λ(a) is equivalent to the matrix αi A being Hurwitz, which is equivalent to P 1 0, (αi A) P 1 + P 1 (αi A) 0. Similarly, the condition R λ(a) < β is equivalent to the matrix A βi being Hurwitz, which is equivalent to P 2 0, (A βi) P 2 + P 2 (A βi) 0. A simple MATLAB program for solving the above LMIs is shown below. n=size(a,1); setlmis(); P1=lmivar(1,n 1); P2=lmivar(1,n 1); lmiterm(1 1 1 P1,1,alpha*eye(n) A,`s'); lmiterm( P1,1,1); lmiterm(3 1 1 P2,1,A beta*eye(n),`s'); lmiterm( P2,1,1); lmis=getlmis; tmin,xfeas=feasp(lmis); if tmin<0 disp(' alpha < Re lambda(a) < beta? : YES ') else disp(' alpha < Re lambda(a) < beta? : NO ') end SPP 9 Since P 0, the matrix inequality (12) is equivalent to P 1 A BK 0, A K B P which can be verified using the Schur complement. Let W = P 1 and pre-multiply and post-multiply the above inequality by diag(i, W ) to obtain W AW BKW W A W K B W 0. 23

24 Let M = KW, then we obtain the LMI condition, W AW BM W A M B W 0. (21) When the above LMI is feasible, the controller gain is given by K = MW 1. SPP 10 The following is a simple MATLAB program solving the LMI (21). A=0 1; 2 3; B=0; 1; n=size(a,1); m=size(b,2); setlmis(); W=lmivar(1,n 1); M=lmivar(2,m n); lmiterm(1 1 1 W,1, 1); lmiterm(1 1 2 W,A,1); lmiterm(1 1 2 M,B, 1); lmiterm(1 2 2 W,1, 1); % (1,1): W % (1,2):AW % (1,2): BM % (2,2): W lmis=getlmis; tmin,xfeas=feasp(lmis); if tmin<0 WW = dec2mat(lmis,xfeas,w); MM = dec2mat(lmis,xfeas,m); K = MM*inv(WW) eig(a B*K) end SPP 11 We first observe from (14) that the (1,1) block of the LMI is negative definite, that is, A P + P A 0, and thus the system is asymptotically stable. 24

25 Next, using the Schur complement, we can show that (14) is equivalent to A P + P A P B C + γ 1 C D 0. B P γi D We pre-multiply the above by x w and post-multiply it by x w to obtain x ( A P + P A ) x + x P Bw + w B P x γw w + γ 1 z z 0. (22) We next evaluate the Lyapunov derivative of V (x) = x P x along solutions of the system to obtain V (x) = ẋ P x + x P ẋ = (Ax + Bw) P x + x P (Ax + Bw) = x ( A P + P A ) x + x P Bw + w B P x γw w γ 1 z z. Integrating both sides of the above inequality from 0 to t gives V (x(t)) V (x(0)) γ t Since V (x(t)) 0 and V (x(0)) = 0, we have t 0 0 t w(τ) w(τ) dτ γ 1 z(τ) z(τ) dτ. t z(τ) z(τ) dτ γ 2 w(τ) w(τ) dτ. SPP 12 Substituting the decentralized state feedback into the interconnected system, we obtain d x1 dt x 2 = A11 B 1 K 1 A 12 A 21 A 22 B 2 K x1 x 2. (23) We consider the following Lyapunov function candidate for the closed-loop system, V (x) = x 1 P 1 x 1 + x P 2 P 2 x 2 = x 1 x 1 O x1 2. O P 2 x 2 The closed-loop system (23) is asymptotically stable if A11 B 1 K 1 A 12 P 1 O A 21 A 22 B 2 K 2 O P 2 P 1 O A11 B 1 K 1 A 12 + O P 2 A 21 A 22 B 2 K

26 Pre-multiplying and post-multiplying the above by Q1 O P 1 O = O Q 2 O P 2 1 yields Q1 O A11 B 1 K 1 A 12 O Q 2 A 21 A 22 B 2 K 2 A11 B 1 K 1 A 12 Q1 O + A 21 A 22 B 2 K 2 O Q 2 0, where Q 1 = P 1 1 and Q 2 = P 1 2. Let M 1 = K 1 Q 1 and M 2 = K 2 Q 2, then performing some matrix multiplications gives A11 Q 1 + Q 1 A 11 B 1 M 1 M 1 B 1 Q 1 A 21 + A 12 Q 2 0, A 21 Q 1 + Q 2 A 12 A 22 Q 2 + Q 2 A 22 B 2 M 2 M 2 B 2 which, together with Q 1 0 and Q 2 0, constitutes the LMIs for asymptotic stability of the closed-loop system. If the LMIs are feasible, the controller gains are given by K 1 = M 1 Q 1 1, K 2 = M 2 Q 1 2. A simple MATLAB program for solving the above LMIs is given below. n1=size(a11,1); m1=size(b1,2); n2=size(a22,1); m2=size(b2,2); setlmis(); Q1=lmivar(1,n1 1); Q2=lmivar(1,n2 1); M1=lmivar(2,m1 n1); M2=lmivar(2,m2 n2); lmiterm(1 1 1 Q1, A11, 1, 's'); lmiterm(1 1 1 M1, B1, 1, 's'); lmiterm(1 1 2 Q1, 1, A21'); lmiterm(1 1 2 Q2, A12, 1); lmiterm(1 2 2 Q2, A22, 1, 's'); lmiterm(1 2 2 M2, B2, 1, 's'); lmiterm( Q1, 1, 1); 26

27 lmiterm( Q2, 1, 1); lmis=getlmis; tmin,xfeas=feasp(lmis); if tmin<0 q1 = dec2mat(lmis,xfeas,q1); q2 = dec2mat(lmis,xfeas,q2); m1 = dec2mat(lmis,xfeas,m1); m2 = dec2mat(lmis,xfeas,m2); K1 = m1*inv(q1); K2 = m2*inv(q2); Acl = A11 B1*K1 A12; A21 A22 B2*K2; eig(acl) end References 1 S. Boyd, L. El Ghaoui, E. Feron, and V. Balakrishnan. Linear Matrix Inequalities in System and Control Theory. SIAM Studies in Applied Mathematics. Society for Industrial and Applied Mathematics, Philadelphia, E. K. P. Chong and S. H. Żak. An Introduction to Optimization. John Wiley & Sons, Inc., Hoboken, New Jersey, fourth edition, B. Ronen and S. Pass. Focused Operations Management: Achieving More With Existing Resources. John Wiley & Sons, Hoboken, New Jersey, J. G. VanAntwerp and R. D. Braatz. A tutorial on linear and bilinear matrix inequalities. Journal of Process Control, 10(4): ,

ECE 680 Linear Matrix Inequalities

ECE 680 Linear Matrix Inequalities ECE 680 Linear Matrix Inequalities Stan Żak School of Electrical and Computer Engineering Purdue University zak@purdue.edu October 11, 2017 Stan Żak (Purdue University) ECE 680 Linear Matrix Inequalities

More information

Appendix A Solving Linear Matrix Inequality (LMI) Problems

Appendix A Solving Linear Matrix Inequality (LMI) Problems Appendix A Solving Linear Matrix Inequality (LMI) Problems In this section, we present a brief introduction about linear matrix inequalities which have been used extensively to solve the FDI problems described

More information

Linear Matrix Inequality Approach

Linear Matrix Inequality Approach 1 MEM800-007 Chapter 4a Linear Matrix Inequality Approach Reference: Linear matrix inequalities in system and control theory, Stephen Boyd et al. SIAM 1994. Linear Matrix Inequality (LMI) approach have

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Introduction to linear matrix inequalities Wojciech Paszke

Introduction to linear matrix inequalities Wojciech Paszke Introduction to linear matrix inequalities Wojciech Paszke Institute of Control and Computation Engineering, University of Zielona Góra, Poland e-mail: W.Paszke@issi.uz.zgora.pl Outline Introduction to

More information

Modern Optimal Control

Modern Optimal Control Modern Optimal Control Matthew M. Peet Arizona State University Lecture 19: Stabilization via LMIs Optimization Optimization can be posed in functional form: min x F objective function : inequality constraints

More information

Module 04 Optimization Problems KKT Conditions & Solvers

Module 04 Optimization Problems KKT Conditions & Solvers Module 04 Optimization Problems KKT Conditions & Solvers Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html September

More information

Linear Matrix Inequality (LMI)

Linear Matrix Inequality (LMI) Linear Matrix Inequality (LMI) A linear matrix inequality is an expression of the form where F (x) F 0 + x 1 F 1 + + x m F m > 0 (1) x = (x 1,, x m ) R m, F 0,, F m are real symmetric matrices, and the

More information

From Convex Optimization to Linear Matrix Inequalities

From Convex Optimization to Linear Matrix Inequalities Dep. of Information Engineering University of Pisa (Italy) From Convex Optimization to Linear Matrix Inequalities eng. Sergio Grammatico grammatico.sergio@gmail.com Class of Identification of Uncertain

More information

8 A First Glimpse on Design with LMIs

8 A First Glimpse on Design with LMIs 8 A First Glimpse on Design with LMIs 8.1 Conceptual Design Problem Given a linear time invariant system design a linear time invariant controller or filter so as to guarantee some closed loop indices

More information

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about Rank-one LMIs and Lyapunov's Inequality Didier Henrion 1;; Gjerrit Meinsma Abstract We describe a new proof of the well-known Lyapunov's matrix inequality about the location of the eigenvalues of a matrix

More information

Graph and Controller Design for Disturbance Attenuation in Consensus Networks

Graph and Controller Design for Disturbance Attenuation in Consensus Networks 203 3th International Conference on Control, Automation and Systems (ICCAS 203) Oct. 20-23, 203 in Kimdaejung Convention Center, Gwangju, Korea Graph and Controller Design for Disturbance Attenuation in

More information

7.1 Linear Systems Stability Consider the Continuous-Time (CT) Linear Time-Invariant (LTI) system

7.1 Linear Systems Stability Consider the Continuous-Time (CT) Linear Time-Invariant (LTI) system 7 Stability 7.1 Linear Systems Stability Consider the Continuous-Time (CT) Linear Time-Invariant (LTI) system ẋ(t) = A x(t), x(0) = x 0, A R n n, x 0 R n. (14) The origin x = 0 is a globally asymptotically

More information

Linear Matrix Inequalities in Control

Linear Matrix Inequalities in Control Linear Matrix Inequalities in Control Delft Center for Systems and Control (DCSC) Delft University of Technology The Netherlands Department of Electrical Engineering Eindhoven University of Technology

More information

Research Article An Equivalent LMI Representation of Bounded Real Lemma for Continuous-Time Systems

Research Article An Equivalent LMI Representation of Bounded Real Lemma for Continuous-Time Systems Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 28, Article ID 67295, 8 pages doi:1.1155/28/67295 Research Article An Equivalent LMI Representation of Bounded Real Lemma

More information

Marcus Pantoja da Silva 1 and Celso Pascoli Bottura 2. Abstract: Nonlinear systems with time-varying uncertainties

Marcus Pantoja da Silva 1 and Celso Pascoli Bottura 2. Abstract: Nonlinear systems with time-varying uncertainties A NEW PROPOSAL FOR H NORM CHARACTERIZATION AND THE OPTIMAL H CONTROL OF NONLINEAR SSTEMS WITH TIME-VARING UNCERTAINTIES WITH KNOWN NORM BOUND AND EXOGENOUS DISTURBANCES Marcus Pantoja da Silva 1 and Celso

More information

Stability of linear time-varying systems through quadratically parameter-dependent Lyapunov functions

Stability of linear time-varying systems through quadratically parameter-dependent Lyapunov functions Stability of linear time-varying systems through quadratically parameter-dependent Lyapunov functions Vinícius F. Montagner Department of Telematics Pedro L. D. Peres School of Electrical and Computer

More information

STABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS. Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec

STABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS. Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec STABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec DAS-CTC-UFSC P.O. Box 476, 88040-900 Florianópolis, SC,

More information

LMI Methods in Optimal and Robust Control

LMI Methods in Optimal and Robust Control LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 02: Optimization (Convex and Otherwise) What is Optimization? An Optimization Problem has 3 parts. x F f(x) :

More information

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 8 September 2003 European Union RTN Summer School on Multi-Agent

More information

Robust Stability. Robust stability against time-invariant and time-varying uncertainties. Parameter dependent Lyapunov functions

Robust Stability. Robust stability against time-invariant and time-varying uncertainties. Parameter dependent Lyapunov functions Robust Stability Robust stability against time-invariant and time-varying uncertainties Parameter dependent Lyapunov functions Semi-infinite LMI problems From nominal to robust performance 1/24 Time-Invariant

More information

15. Conic optimization

15. Conic optimization L. Vandenberghe EE236C (Spring 216) 15. Conic optimization conic linear program examples modeling duality 15-1 Generalized (conic) inequalities Conic inequality: a constraint x K where K is a convex cone

More information

LOW ORDER H CONTROLLER DESIGN: AN LMI APPROACH

LOW ORDER H CONTROLLER DESIGN: AN LMI APPROACH LOW ORDER H CONROLLER DESIGN: AN LMI APPROACH Guisheng Zhai, Shinichi Murao, Naoki Koyama, Masaharu Yoshida Faculty of Systems Engineering, Wakayama University, Wakayama 640-8510, Japan Email: zhai@sys.wakayama-u.ac.jp

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization Instructor: Farid Alizadeh Author: Ai Kagawa 12/12/2012

More information

EE363 homework 8 solutions

EE363 homework 8 solutions EE363 Prof. S. Boyd EE363 homework 8 solutions 1. Lyapunov condition for passivity. The system described by ẋ = f(x, u), y = g(x), x() =, with u(t), y(t) R m, is said to be passive if t u(τ) T y(τ) dτ

More information

Lecture Notes DISC Course on. Linear Matrix Inequalities in Control. Carsten Scherer and Siep Weiland

Lecture Notes DISC Course on. Linear Matrix Inequalities in Control. Carsten Scherer and Siep Weiland Lecture Notes DISC Course on Linear Matrix Inequalities in Control Carsten Scherer and Siep Weiland Version 2.: April, 1999 2 Preface Objectives In recent years linear matrix inequalities LMI s have emerged

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

Nonlinear Control Design for Linear Differential Inclusions via Convex Hull Quadratic Lyapunov Functions

Nonlinear Control Design for Linear Differential Inclusions via Convex Hull Quadratic Lyapunov Functions Nonlinear Control Design for Linear Differential Inclusions via Convex Hull Quadratic Lyapunov Functions Tingshu Hu Abstract This paper presents a nonlinear control design method for robust stabilization

More information

Linear Matrix Inequalities in Robust Control. Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University MTNS 2002

Linear Matrix Inequalities in Robust Control. Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University MTNS 2002 Linear Matrix Inequalities in Robust Control Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University MTNS 2002 Objective A brief introduction to LMI techniques for Robust Control Emphasis on

More information

Convex Optimization 1

Convex Optimization 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski Convex Optimization 1 Many optimization objectives generated

More information

Static Output Feedback Stabilisation with H Performance for a Class of Plants

Static Output Feedback Stabilisation with H Performance for a Class of Plants Static Output Feedback Stabilisation with H Performance for a Class of Plants E. Prempain and I. Postlethwaite Control and Instrumentation Research, Department of Engineering, University of Leicester,

More information

FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES. Danlei Chu, Tongwen Chen, Horacio J. Marquez

FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES. Danlei Chu, Tongwen Chen, Horacio J. Marquez FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES Danlei Chu Tongwen Chen Horacio J Marquez Department of Electrical and Computer Engineering University of Alberta Edmonton

More information

SYNTHESIS OF ROBUST DISCRETE-TIME SYSTEMS BASED ON COMPARISON WITH STOCHASTIC MODEL 1. P. V. Pakshin, S. G. Soloviev

SYNTHESIS OF ROBUST DISCRETE-TIME SYSTEMS BASED ON COMPARISON WITH STOCHASTIC MODEL 1. P. V. Pakshin, S. G. Soloviev SYNTHESIS OF ROBUST DISCRETE-TIME SYSTEMS BASED ON COMPARISON WITH STOCHASTIC MODEL 1 P. V. Pakshin, S. G. Soloviev Nizhny Novgorod State Technical University at Arzamas, 19, Kalinina ul., Arzamas, 607227,

More information

Robust and Optimal Control, Spring 2015

Robust and Optimal Control, Spring 2015 Robust and Optimal Control, Spring 2015 Instructor: Prof. Masayuki Fujita (S5-303B) D. Linear Matrix Inequality D.1 Convex Optimization D.2 Linear Matrix Inequality(LMI) D.3 Control Design and LMI Formulation

More information

Vector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n.

Vector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n. Vector Spaces Definition: The usual addition and scalar multiplication of n-tuples x = (x 1,..., x n ) R n (also called vectors) are the addition and scalar multiplication operations defined component-wise:

More information

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013 Convex Optimization (EE227A: UC Berkeley) Lecture 6 (Conic optimization) 07 Feb, 2013 Suvrit Sra Organizational Info Quiz coming up on 19th Feb. Project teams by 19th Feb Good if you can mix your research

More information

Dissipativity. Outline. Motivation. Dissipative Systems. M. Sami Fadali EBME Dept., UNR

Dissipativity. Outline. Motivation. Dissipative Systems. M. Sami Fadali EBME Dept., UNR Dissipativity M. Sami Fadali EBME Dept., UNR 1 Outline Differential storage functions. QSR Dissipativity. Algebraic conditions for dissipativity. Stability of dissipative systems. Feedback Interconnections

More information

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Module 07 Controllability and Controller Design of Dynamical LTI Systems Module 07 Controllability and Controller Design of Dynamical LTI Systems Ahmad F. Taha EE 5143: Linear Systems and Control Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ataha October

More information

Denis ARZELIER arzelier

Denis ARZELIER   arzelier COURSE ON LMI OPTIMIZATION WITH APPLICATIONS IN CONTROL PART II.2 LMIs IN SYSTEMS CONTROL STATE-SPACE METHODS PERFORMANCE ANALYSIS and SYNTHESIS Denis ARZELIER www.laas.fr/ arzelier arzelier@laas.fr 15

More information

Riccati Equations and Inequalities in Robust Control

Riccati Equations and Inequalities in Robust Control Riccati Equations and Inequalities in Robust Control Lianhao Yin Gabriel Ingesson Martin Karlsson Optimal Control LP4 2014 June 10, 2014 Lianhao Yin Gabriel Ingesson Martin Karlsson (LTH) H control problem

More information

An alternative proof of the Barker, Berman, Plemmons (BBP) result on diagonal stability and extensions - Corrected Version

An alternative proof of the Barker, Berman, Plemmons (BBP) result on diagonal stability and extensions - Corrected Version 1 An alternative proof of the Barker, Berman, Plemmons (BBP) result on diagonal stability and extensions - Corrected Version Robert Shorten 1, Oliver Mason 1 and Christopher King 2 Abstract The original

More information

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term EE 150 - Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko JPL Third Term 2011-2012 Due on Thursday May 3 in class. Homework Set #4 1. (10 points) (Adapted

More information

Linear Systems with Saturating Controls: An LMI Approach. subject to control saturation. No assumption is made concerning open-loop stability and no

Linear Systems with Saturating Controls: An LMI Approach. subject to control saturation. No assumption is made concerning open-loop stability and no Output Feedback Robust Stabilization of Uncertain Linear Systems with Saturating Controls: An LMI Approach Didier Henrion 1 Sophie Tarbouriech 1; Germain Garcia 1; Abstract : The problem of robust controller

More information

On optimal quadratic Lyapunov functions for polynomial systems

On optimal quadratic Lyapunov functions for polynomial systems On optimal quadratic Lyapunov functions for polynomial systems G. Chesi 1,A.Tesi 2, A. Vicino 1 1 Dipartimento di Ingegneria dell Informazione, Università disiena Via Roma 56, 53100 Siena, Italy 2 Dipartimento

More information

Convex Optimization Approach to Dynamic Output Feedback Control for Delay Differential Systems of Neutral Type 1,2

Convex Optimization Approach to Dynamic Output Feedback Control for Delay Differential Systems of Neutral Type 1,2 journal of optimization theory and applications: Vol. 127 No. 2 pp. 411 423 November 2005 ( 2005) DOI: 10.1007/s10957-005-6552-7 Convex Optimization Approach to Dynamic Output Feedback Control for Delay

More information

Convex optimization problems. Optimization problem in standard form

Convex optimization problems. Optimization problem in standard form Convex optimization problems optimization problem in standard form convex optimization problems linear optimization quadratic optimization geometric programming quasiconvex optimization generalized inequality

More information

Introduction and Math Preliminaries

Introduction and Math Preliminaries Introduction and Math Preliminaries Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Appendices A, B, and C, Chapter

More information

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis Topic # 16.30/31 Feedback Control Systems Analysis of Nonlinear Systems Lyapunov Stability Analysis Fall 010 16.30/31 Lyapunov Stability Analysis Very general method to prove (or disprove) stability of

More information

9 The LQR Problem Revisited

9 The LQR Problem Revisited 9 he LQR Problem Revisited Problem: Compute a state feedback controller u(t) = Kx(t) that stabilizes the closed loop system and minimizes J := lim E [ z(t) z(t) ] t for the LI system Assumptions: a) D

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Parameterized Linear Matrix Inequality Techniques in Fuzzy Control System Design

Parameterized Linear Matrix Inequality Techniques in Fuzzy Control System Design 324 IEEE TRANSACTIONS ON FUZZY SYSTEMS, VOL. 9, NO. 2, APRIL 2001 Parameterized Linear Matrix Inequality Techniques in Fuzzy Control System Design H. D. Tuan, P. Apkarian, T. Narikiyo, and Y. Yamamoto

More information

Computational Finance

Computational Finance Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples

More information

Lecture 6: Conic Optimization September 8

Lecture 6: Conic Optimization September 8 IE 598: Big Data Optimization Fall 2016 Lecture 6: Conic Optimization September 8 Lecturer: Niao He Scriber: Juan Xu Overview In this lecture, we finish up our previous discussion on optimality conditions

More information

Lecture: Convex Optimization Problems

Lecture: Convex Optimization Problems 1/36 Lecture: Convex Optimization Problems http://bicmr.pku.edu.cn/~wenzw/opt-2015-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Introduction 2/36 optimization

More information

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979) Course Outline FRTN Multivariable Control, Lecture Automatic Control LTH, 6 L-L Specifications, models and loop-shaping by hand L6-L8 Limitations on achievable performance L9-L Controller optimization:

More information

Figure 3.1: Unity feedback interconnection

Figure 3.1: Unity feedback interconnection Chapter 3 Small Gain Theorem and Integral Quadratic Constraints This chapter presents a version of the small gain theorem for L2 gains, and then proceeds to introduce the general technique of working with

More information

Lecture 2: Convex Sets and Functions

Lecture 2: Convex Sets and Functions Lecture 2: Convex Sets and Functions Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Network Optimization, Fall 2015 1 / 22 Optimization Problems Optimization problems are

More information

CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems

CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems CDS 101 1. For each of the following linear systems, determine whether the origin is asymptotically stable and, if so, plot the step response and frequency response for the system. If there are multiple

More information

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min. MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.

More information

Semidefinite Programming Duality and Linear Time-invariant Systems

Semidefinite Programming Duality and Linear Time-invariant Systems Semidefinite Programming Duality and Linear Time-invariant Systems Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 2 July 2004 Workshop on Linear Matrix Inequalities in Control LAAS-CNRS,

More information

EE C128 / ME C134 Feedback Control Systems

EE C128 / ME C134 Feedback Control Systems EE C128 / ME C134 Feedback Control Systems Lecture Additional Material Introduction to Model Predictive Control Maximilian Balandat Department of Electrical Engineering & Computer Science University of

More information

w T 1 w T 2. w T n 0 if i j 1 if i = j

w T 1 w T 2. w T n 0 if i j 1 if i = j Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible -

More information

Advances in Convex Optimization: Theory, Algorithms, and Applications

Advances in Convex Optimization: Theory, Algorithms, and Applications Advances in Convex Optimization: Theory, Algorithms, and Applications Stephen Boyd Electrical Engineering Department Stanford University (joint work with Lieven Vandenberghe, UCLA) ISIT 02 ISIT 02 Lausanne

More information

4. Convex optimization problems

4. Convex optimization problems Convex Optimization Boyd & Vandenberghe 4. Convex optimization problems optimization problem in standard form convex optimization problems quasiconvex optimization linear optimization quadratic optimization

More information

On Computing the Worst-case Performance of Lur'e Systems with Uncertain Time-invariant Delays

On Computing the Worst-case Performance of Lur'e Systems with Uncertain Time-invariant Delays Article On Computing the Worst-case Performance of Lur'e Systems with Uncertain Time-invariant Delays Thapana Nampradit and David Banjerdpongchai* Department of Electrical Engineering, Faculty of Engineering,

More information

A Characterization of the Hurwitz Stability of Metzler Matrices

A Characterization of the Hurwitz Stability of Metzler Matrices 29 American Control Conference Hyatt Regency Riverfront, St Louis, MO, USA June -2, 29 WeC52 A Characterization of the Hurwitz Stability of Metzler Matrices Kumpati S Narendra and Robert Shorten 2 Abstract

More information

An LMI Approach to the Control of a Compact Disc Player. Marco Dettori SC Solutions Inc. Santa Clara, California

An LMI Approach to the Control of a Compact Disc Player. Marco Dettori SC Solutions Inc. Santa Clara, California An LMI Approach to the Control of a Compact Disc Player Marco Dettori SC Solutions Inc. Santa Clara, California IEEE SCV Control Systems Society Santa Clara University March 15, 2001 Overview of my Ph.D.

More information

Control with Random Communication Delays via a Discrete-Time Jump System Approach

Control with Random Communication Delays via a Discrete-Time Jump System Approach Control with Random Communication Delays via a Discrete-Time Jump System Approach Lin Xiao Arash Hassibi Jonathan P How Information Systems Laboratory Stanford University Stanford, CA 9435, USA Abstract

More information

Application of LMI for design of digital control systems

Application of LMI for design of digital control systems BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 56, No. 4, 28 Application of LMI for design of digital control systems A. KRÓLIKOWSKI and D. HORLA Institute of Control and Information

More information

1 Strict local optimality in unconstrained optimization

1 Strict local optimality in unconstrained optimization ORF 53 Lecture 14 Spring 016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, April 14, 016 When in doubt on the accuracy of these notes, please cross check with the instructor s

More information

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures Preprints of the 19th World Congress The International Federation of Automatic Control Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures Eric Peterson Harry G.

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Support vector machines (SVMs) are one of the central concepts in all of machine learning. They are simply a combination of two ideas: linear classification via maximum (or optimal

More information

Lecture 2 and 3: Controllability of DT-LTI systems

Lecture 2 and 3: Controllability of DT-LTI systems 1 Lecture 2 and 3: Controllability of DT-LTI systems Spring 2013 - EE 194, Advanced Control (Prof Khan) January 23 (Wed) and 28 (Mon), 2013 I LTI SYSTEMS Recall that continuous-time LTI systems can be

More information

IMPROVED MPC DESIGN BASED ON SATURATING CONTROL LAWS

IMPROVED MPC DESIGN BASED ON SATURATING CONTROL LAWS IMPROVED MPC DESIGN BASED ON SATURATING CONTROL LAWS D. Limon, J.M. Gomes da Silva Jr., T. Alamo and E.F. Camacho Dpto. de Ingenieria de Sistemas y Automática. Universidad de Sevilla Camino de los Descubrimientos

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

ECE580 Partial Solution to Problem Set 3

ECE580 Partial Solution to Problem Set 3 ECE580 Fall 2015 Solution to Problem Set 3 October 23, 2015 1 ECE580 Partial Solution to Problem Set 3 These problems are from the textbook by Chong and Zak, 4th edition, which is the textbook for the

More information

d A 0 + m t k A k 0 whenever λ min (B k (x)) t k λ max (B k (x)) for k = 1, 2,..., m x n B n (k).

d A 0 + m t k A k 0 whenever λ min (B k (x)) t k λ max (B k (x)) for k = 1, 2,..., m x n B n (k). MATRIX CUBES PARAMETERIZED BY EIGENVALUES JIAWANG NIE AND BERND STURMFELS Abstract. An elimination problem in semidefinite programming is solved by means of tensor algebra. It concerns families of matrix

More information

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP)

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP) Agenda 1 Cone programming 2 Convex cones 3 Generalized inequalities 4 Linear programming (LP) 5 Second-order cone programming (SOCP) 6 Semidefinite programming (SDP) 7 Examples Optimization problem in

More information

Stability lectures. Stability of Linear Systems. Stability of Linear Systems. Stability of Continuous Systems. EECE 571M/491M, Spring 2008 Lecture 5

Stability lectures. Stability of Linear Systems. Stability of Linear Systems. Stability of Continuous Systems. EECE 571M/491M, Spring 2008 Lecture 5 EECE 571M/491M, Spring 2008 Lecture 5 Stability of Continuous Systems http://courses.ece.ubc.ca/491m moishi@ece.ubc.ca Dr. Meeko Oishi Electrical and Computer Engineering University of British Columbia,

More information

Copositive Plus Matrices

Copositive Plus Matrices Copositive Plus Matrices Willemieke van Vliet Master Thesis in Applied Mathematics October 2011 Copositive Plus Matrices Summary In this report we discuss the set of copositive plus matrices and their

More information

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 8: Basic Lyapunov Stability Theory

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 8: Basic Lyapunov Stability Theory MCE/EEC 647/747: Robot Dynamics and Control Lecture 8: Basic Lyapunov Stability Theory Reading: SHV Appendix Mechanical Engineering Hanz Richter, PhD MCE503 p.1/17 Stability in the sense of Lyapunov A

More information

Constrained interpolation-based control for polytopic uncertain systems

Constrained interpolation-based control for polytopic uncertain systems 2011 50th IEEE Conference on Decision and Control and European Control Conference CDC-ECC Orlando FL USA December 12-15 2011 Constrained interpolation-based control for polytopic uncertain systems H.-N.

More information

Analysis and design of switched normal systems

Analysis and design of switched normal systems Nonlinear Analysis 65 (2006) 2248 2259 www.elsevier.com/locate/na Analysis and design of switched normal systems Guisheng Zhai a,, Xuping Xu b, Hai Lin c, Anthony N. Michel c a Department of Mechanical

More information

On Bounded Real Matrix Inequality Dilation

On Bounded Real Matrix Inequality Dilation On Bounded Real Matrix Inequality Dilation Solmaz Sajjadi-Kia and Faryar Jabbari Abstract We discuss a variation of dilated matrix inequalities for the conventional Bounded Real matrix inequality, and

More information

EE5138R: Problem Set 5 Assigned: 16/02/15 Due: 06/03/15

EE5138R: Problem Set 5 Assigned: 16/02/15 Due: 06/03/15 EE538R: Problem Set 5 Assigned: 6/0/5 Due: 06/03/5. BV Problem 4. The feasible set is the convex hull of (0, ), (0, ), (/5, /5), (, 0) and (, 0). (a) x = (/5, /5) (b) Unbounded below (c) X opt = {(0, x

More information

Solving of the Modified Filter Algebraic Riccati Equation for H-infinity fault detection filtering

Solving of the Modified Filter Algebraic Riccati Equation for H-infinity fault detection filtering Acta Universitatis Sapientiae Electrical and Mechanical Engineering, 9 (2017) 57-77 DOI: 10.1515/auseme-2017-0010 Solving of the Modified Filter Algebraic Riccati Equation for H-infinity fault detection

More information

H 2 Suboptimal Estimation and Control for Nonnegative

H 2 Suboptimal Estimation and Control for Nonnegative Proceedings of the 2007 American Control Conference Marriott Marquis Hotel at Times Square New York City, USA, July 11-13, 2007 FrC20.3 H 2 Suboptimal Estimation and Control for Nonnegative Dynamical Systems

More information

Chapter 2 Set Theoretic Methods in Control

Chapter 2 Set Theoretic Methods in Control Chapter 2 Set Theoretic Methods in Control 2.1 Set Terminology For completeness, some standard definitions of set terminology will be introduced. For a detailed reference, the reader is referred to the

More information

Gramians based model reduction for hybrid switched systems

Gramians based model reduction for hybrid switched systems Gramians based model reduction for hybrid switched systems Y. Chahlaoui Younes.Chahlaoui@manchester.ac.uk Centre for Interdisciplinary Computational and Dynamical Analysis (CICADA) School of Mathematics

More information

arzelier

arzelier COURSE ON LMI OPTIMIZATION WITH APPLICATIONS IN CONTROL PART II.1 LMIs IN SYSTEMS CONTROL STATE-SPACE METHODS STABILITY ANALYSIS Didier HENRION www.laas.fr/ henrion henrion@laas.fr Denis ARZELIER www.laas.fr/

More information

The Simplest Semidefinite Programs are Trivial

The Simplest Semidefinite Programs are Trivial The Simplest Semidefinite Programs are Trivial Robert J. Vanderbei Bing Yang Program in Statistics & Operations Research Princeton University Princeton, NJ 08544 January 10, 1994 Technical Report SOR-93-12

More information

Converse Lyapunov theorem and Input-to-State Stability

Converse Lyapunov theorem and Input-to-State Stability Converse Lyapunov theorem and Input-to-State Stability April 6, 2014 1 Converse Lyapunov theorem In the previous lecture, we have discussed few examples of nonlinear control systems and stability concepts

More information

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form: 0.1 N. L. P. Katta G. Murty, IOE 611 Lecture slides Introductory Lecture NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP does not include everything

More information

EE 227A: Convex Optimization and Applications October 14, 2008

EE 227A: Convex Optimization and Applications October 14, 2008 EE 227A: Convex Optimization and Applications October 14, 2008 Lecture 13: SDP Duality Lecturer: Laurent El Ghaoui Reading assignment: Chapter 5 of BV. 13.1 Direct approach 13.1.1 Primal problem Consider

More information

Robust Observer for Uncertain T S model of a Synchronous Machine

Robust Observer for Uncertain T S model of a Synchronous Machine Recent Advances in Circuits Communications Signal Processing Robust Observer for Uncertain T S model of a Synchronous Machine OUAALINE Najat ELALAMI Noureddine Laboratory of Automation Computer Engineering

More information

Controller synthesis for positive systems under l 1-induced performance

Controller synthesis for positive systems under l 1-induced performance Title Controller synthesis for positive systems under l 1-induced performance Author(s) Chen, X; Lam, J; Li, P; Shu, Z Citation The 24th Chinese Control and Decision Conference (CCDC 212), Taiyuan, China,

More information

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4

More information

H State-Feedback Controller Design for Discrete-Time Fuzzy Systems Using Fuzzy Weighting-Dependent Lyapunov Functions

H State-Feedback Controller Design for Discrete-Time Fuzzy Systems Using Fuzzy Weighting-Dependent Lyapunov Functions IEEE TRANSACTIONS ON FUZZY SYSTEMS, VOL 11, NO 2, APRIL 2003 271 H State-Feedback Controller Design for Discrete-Time Fuzzy Systems Using Fuzzy Weighting-Dependent Lyapunov Functions Doo Jin Choi and PooGyeon

More information

Problem 1 Convexity Property

Problem 1 Convexity Property THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 4 Hafez Bazrafshan October 1, 2015 To facilitate checking, questions are color-coded blue and pertinent

More information

Optimality Conditions

Optimality Conditions Chapter 2 Optimality Conditions 2.1 Global and Local Minima for Unconstrained Problems When a minimization problem does not have any constraints, the problem is to find the minimum of the objective function.

More information