Rankone LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the wellknown Lyapunov's matrix inequality about


 Lorena McCormick
 1 years ago
 Views:
Transcription
1 Rankone LMIs and Lyapunov's Inequality Didier Henrion 1;; Gjerrit Meinsma Abstract We describe a new proof of the wellknown Lyapunov's matrix inequality about the location of the eigenvalues of a matrix in some region of the complex plane. The proof makes use of standard facts from quadratic and semidenite programming. Links are established between the Lyapunov matrix, rankone LMIs and the Lagrange multiplier arising in duality theory. Keywords Linear Systems, Stability, LMI. 1 Introduction Let A C nn be a given complex matrix and let 1 a D = fs C : s b b c 1 s < 0g denote a given open region of the complex plane, where Hermitian matrix a b b C c has one strictly negative eigenvalue and one strictly positive eigenvalue, and the star denotes transpose conjugate. In the sequel, the notation P 0 or P 0 (resp. P 0 or P 0) means that matrix P is positive denite (resp. semidenite). The location of the eigenvalues of A can be characterized as follows. 1 Corresponding author. FAX: Laboratoire d'architecture et d'analyse des Systemes, Centre National de la Recherche Scientique, Avenue du Colonel Roche, Toulouse, cedex, France. Institut National des Sciences Appliquees, Complexe Scientique de Rangueil, Toulouse, cedex, France. Faculty of Applied Mathematics, University of Twente, P.O.Box 1, Enschede, 00 AE, The Netherlands. 1
2 Theorem 1 (Lyapunov's Inequality) Matrix A has all its eigenvalues in region D if and only if there is a matrix P = P 0 C nn such that I ap bp I A b 0: (1) P cp A Matrix inequality (1) is referred to as Lyapunov's inequality. Lyapunov's proof of the above theorem { originally developed in the case that D is the open left halfplane, i.e. a = c = 0, b = 1 and inequality (1) becomes A P + P A 0 { relies on the construction of a positive quadratic function whose derivative is negative along the trajectories of an associated dynamical system, see e.g. []. It can be extended to arbitrary regions D via a conformal mapping. Another proof of Theorem 1 can be found in [1, Theorem.19]. Eigenvectors of matrix A are used to show that existence of P implies stability of A, whereas the converse statement is shown via properties of the matrix exponential function. The aim of this note is to give a new, alternative proof of Lyapunov's inequality without referring to stability of the trajectories of a dynamical system or to matrix exponentials. We use elementary concepts from linear algebra, quadratic and semidenite programming. Links are established between the Lyapunov matrix and the Lagrange multiplier arising in duality theory. Relationships with rankone LMIs, the KalmanYakubovich Popov Lemma and (D; G)scaling in analysis are also pointed out. The proof relies on the following important result, proved e.g. as in [9, Lemma ]. Lemma 1 Two column vectors p; q C n with q nonzero satisfy a b [q p] [q p] 0 () b c if and only if p = sq for some s D C. Rankone LMI Problem First we show the equivalence between location of the eigenvalues of A in region D and a rankone LMI optimization problem or a rankone LMI feasibility problem. If s C is an eigenvalue of A, then there exists a nonzero vector q C n such that (A si)q = 0: () Pursuing an idea proposed in [, Chapter 1], it follows that all the eigenvalues of A belong to D if and only if the optimal value of the quadratic optimization problem = min q (A si) (A si)q s:t: s D C q q = 1 ()
3 is strictly positive, where 1 D C = fs C : s a b b c 1 s 0g is the closed region complementary to D in C. Dene the rankone positive semidenite matrix q q X = xx = 0 p p and use the notations A(s)q = [ A I ] x = Ax q = [ I 0 ] x = Qx p = [ 0 I ] x = Px to write inequality () as an LMI in rankone matrix X, namely ax b F (X) = Q P X bx cx Q P 0 () where F is a linear map from C nn to C nn. Using these notations, an alternative formulation of quadratic optimization problem () is given by the following lemma. Lemma The eigenvalues of matrix A belong to region D if and only if > 0 in rankone LMI optimization problem = min Trace A AX s:t: F (X) 0 X = X 0 Trace Q QX = 1 Rank X = 1: () The above rankone LMI problem is an optimization problem. It turns out that we can equivalently state this result via a feasibility problem, following an idea exposed in []. To see this, note that Lemma 1 and equation () imply that A has no eigenvalue in D C if and only if there is no nonzero vector q for which a b [q Aq] [q Aq] 0: () b c The lefthand side of inequality () can alternatively be expressed as aqq + baqq + b qq A + caqq A = I A aqq b qq Now dene the linear map bqq cqq I A : aq b G(Q) = Q I A bq cq I A 0 (8) from C nn to C nn. With Q denoting the nonzero rankone matrix Q = qq we arrive at the following result which is equivalent to Lemma.
4 Lemma The eigenvalues of matrix A belong to region D if and only if there is no solution to the rankone LMI feasibility problem G(Q) 0 Q = Q 0 Trace Q = 1 Rank Q = 1: (9) LMI Problem Now we show that the nonconvex rank constraints in LMI problems () and (9) are actually irrelevant. Let I N = A denote a matrix whose columns span the ndimensional right nullspace of full rowrank matrix A. If s k C is a nondefective eigenvalue of A (i.e. its algebraic multiplicity is equal to its geometric multiplicity) and q k C n is the corresponding eigenvector, then the vector qk x k = s k q k belongs to the right nullspace of matrix A. Similarly, if s k is a defective eigenvalue of A (i.e. its algebraic multiplicity is greater than its geometric multiplicity), then the corresponding chain of linearly independent generalized eigenvectors q k ; q k+1 ; q k+ ; : : : gives rise to vectors x k = qk s k q k x k+1 = q k+1 s k q k+1 + q k also belonging to the right nullspace of A. Let x k+ = V = [x 1 x n ] q k+ s k q k+ + q k+1 (10) denote a matrix built up from all the vectors x i associated with all the eigenvalues s i of A. It follows from the above discussion that the columns of N and V span the same vector space. By denition, vectors q i are linearly independent, thus we can dene linearly independent vectors q i C n such that [q 1 q n ] [q 1 q n ] = I: (11) Following these preliminaries, consider now the following relaxation of rankone LMI problem () = min Trace A AX s:t: F (X) 0 (1) X = X 0 Trace Q QX = 1 where the nonconvex rank constraint has been dropped. Since the nonconvex feasible set in problem () is a subset of the convex feasible set in problem (1), LMI optimization problem (1) is referred to as a convex relaxation of the nonconvex rankone LMI problem (). In relation to the above problem, we can state the following central result.
5 Lemma > 0 in rankone LMI optimization problem () if and only if > 0 in LMI optimization problem (1). Proof The inner product of positive semidenite matrices A A and X is always nonnegative, hence 0. Moreover, the fact that > 0 implies > 0 is trivial since the feasible set in problem () is a subset of the feasible set in problem (1), i.e. it holds. Consequently, in order to show that > 0 implies > 0, the remainder of the proof will consist in proving that = 0 implies = 0. So suppose that X is a positive semidenite matrix such that = 0 in problem (1). Let W be a n r full column rank matrix such that X = W W. By putting matrix A A into Schur form, it can easily be shown that Trace A AW W = 0 implies A AW W = 0. Consequently, the columns of W span a subspace that belongs to the right nullspace of A. In view of the above denition of matrix V, there exists a matrix M such that W = V M. Let m ij denote the entries of positive semidenite matrix MM C nn. For a given index k, it holds either m kk > 0 or m ik = m ki = 0 for all i = 1; : : : ; n. Matrix X is feasible for problem (1) thus F (X) = F (V MM V ) = nx i=1 nx j=1 m ij F (x i x j) 0: (1) Since matrix X cannot be zero by assumption, matrix MM is also nonzero and there exists at least one index k such that m kk > 0. Let x k+l be the last eigenvector in the chain of generalized eigenvectors with eigenvalue s k for which m (k+l)(k+l) is nonzero (note that l = 0 if s k is nondefective). From relations (), (10), (11) and (1) it follows that q k+l F (X)q k+l = m (k+l)(k+l) (a + bs k + b s k + cs k s k) 0: Since m (k+l)(k+l) > 0 we see that s k D C, hence vector x k in equation (10) is such that Trace A Ax k x k = 0 and F (x kx k ) 0 in virtue of Lemma 1. Consequently, matrix x kx k is a solution to rankone LMI problem () such that = 0 and the lemma is proved. The following result is then a straightforward corollary to Lemma : Lemma The eigenvalues of matrix A belong to region D if and only if > 0 in LMI optimization problem (1). Now consider the following relaxation to rankone LMI feasibility problem (9): G(Q) 0 Q = Q 0 Trace; Q = 1 (1) where the rank constraint has been dropped. Using the same kind of arguments as above, we can show the following counterpart to Lemma : Lemma The eigenvalues of matrix A belong to region D if and only if there is no solution to LMI feasibility problem (1).
6 Dual LMI Problem Now we use standard semidenite programming duality results [10] to come up with a more compact formulation of the stability conditions of Lemmas and and prove the Lyapunov's inequality of Theorem 1. Dene the linear map Q F D (P ) = P ap b P bp Q cp P dual to the map introduced in (). It is easy to show that ap = b P Trace F D (P )X = Trace F (X)P: Using standard duality arguments, we now prove that the LMI feasibility problem A A F D (P ) P = P 0 is dual to LMI optimization problem (1). To see this, build the Lagrangian bp cp L(P; X; Y ) = Trace (A A F D (P ))X Trace P Y = Trace A AX + Trace (F (X) Y )P of problem (1) where X = X 0 and Y = Y 0 are Lagrange multiplier matrices. The dual function associated with the Lagrangian reads Trace A AX iff (X) = Y 0 g(x; Y ) = min L(P; X; Y ) = P 1 otherwise: The dual optimization problem, obtained by maximizing dual function g(x; Y ) is therefore LMI optimization problem (1), where the equality constraint Trace Q QX = 1 ensures compactness of the feasible set. The matrix inequalities in problem (1) are strict, hence there is no duality gap and > 0 in LMI optimization problem (1) if and only if LMI problem (1) is feasible. Recall that N denotes a matrix whose columns span the right nullspace of A. Then it follows from the Elimination Lemma [] that feasibility problem (1) can equivalently be written as This is exactly the statement of Theorem 1. Similarly, we can dene G D (P ) = N F D (P )N = N F D (P )N 0 P = P 0: I A ap b P bp cp I A as the linear map from C nn to C nn dual to the linear map G(Q) introduced in (8). It is easy to show that Trace G D (P )Q = Trace G(Q)P: It now follows that nonexistence of a nonzero Q = Q 0 for which G(Q) 0 is equivalent to the existence of P = P 0 for which G D (P ) 0. In other words we proved Theorem 1. (1) (1) (1)
7 Numerical Examples.1 First Example Let A = 1 be a constant matrix with eigenvalues 1 and  and let be the stability region. Primal LMI problem (1) reads = min Trace s:t: D = fs C : s + s < 0g X X = X 0 Trace X = 1: X X 0 With a relative accuracy of 10 8, the LMI Control Toolbox 1.0. for Matlab. [] returns and X = = 0:191 0: : : : : :18180: : : : : : : : : : as the optimum of the above problem. In virtue of Theorem, is strictly positive hence all the eigenvalues of matrix A belong to region D.
8 Dual LMI problem (1) reads 1 P = P 0: 0 P + P 1 C A 1 0 With the help of the LMI Toolbox, we obtained the matrix P = 1:8810 0:108 0:108 0:09190 as a feasible solution for the above problem. On the other hand, LMI problem (1) reads 1 0 Q Q Q = Q 0; Trace Q = 1: This problem is infeasible, which is consistent with the above results and Theorem 1.. Second Example Now let A = 1 be a constant matrix with eigenvalues  and and let be the stability region. D = fs C : s + s < 0g With a relative accuracy of 10 8, the LMI Toolbox returns and X = = 0: : : : : : : : : : : : : : : : : as the optimum of primal problem (1). In virtue of Theorem, some eigenvalues of matrix A do not belong to region D. One can check that X = xx = 0:10 0: : : :10 0: : :190
9 is actually a rankone solution to LMI problem (1). Vector x can be written as q x = sq where q is an eigenvector of matrix A corresponding to the eigenvalue s = D C. One can check that positive semidenite matrix 0: : Q = qq = 0: : is a feasible solution for LMI problems (9) or (1). On the other hand, dual LMI problem (1) is found infeasible, which is consistent with the above results and Theorem 1. Conclusion We have proposed a new proof of Lyapunov's matrix inequality that relies on elementary optimization techniques and linear algebra. Following ideas proposed in [] and [, Chapter 1], we consider the eigenvalue location problem as a mere quadratic optimization problem. Then, the quadratic problem can be formulated as an LMI problem with a nonconvex rank constraint. The Lyapunov matrix can be viewed as a Lagrange multiplier matrix arising when dualizing this problem. In [, Chapter 1, x1..], it is shown that removing the nonconvex rankone constraint leads to a sucient LMI stability condition. Our contribution is in showing in Lemmas and that the LMI conditions are also necessary. In other words, the rank constraint in problems () and (9) are irrelevant as far as eigenvalue location is concerned. In a similar fashion, the eigenvalue location problem can be viewed as a frequencydependent analysis problem with one repeated scalar block si corresponding to the Laplace variable s. The Lyapunov matrix P plays the role of a Dscaling matrix associated with the repeated scalar block, and the irrelevance of the nonconvex rank constraint readily follows from the losslessness of the (D; G)scaling as pointed out in [8]. Equivalence of primal problem (1) and dual problem (1) can also be shown via geometric arguments similar to that used in the proof of the KalmanYakubovichPopov (KYP) Lemma in [9, Theorem 1], in the proof of losslessness of (D; G)scaling [, Lemma.1], in the Sprocedure [11] or also in the generalized Sprocedure proposed in [, Theorem 1]. Our approach is also very similar in spirit to the one pursued in [9] to provide an alternative proof of the KYP Lemma. Note however that in this reference the author considers a version of the KYP Lemma where the Laplace variable s varies on the imaginary axis or the unit circle. This result has been extended to other onedimensional curves of the complex plane such as the real axis [] or a segment on the imaginary axis []. These curves are boundaries of the twodimensional stability regions D considered in the present note. It is therefore expected that we can similarly derive more general versions of the KYP Lemma in twodimensional stability regions. 9
10 Finally, we are currently investigating the application of these techniques to the study of stability of polynomial matrices, twoindeterminate polynomial matrices and uncertain polynomial matrices. Related results will be reported elsewhere. References [1] S. Barnett \Polynomials and Linear Control Systems", Marcel Dekker, New York, 198. [] S. Boyd, L. El Ghaoui, E. Feron and V. Balakrishnan \Linear Matrix Inequalities in System and Control Theory", SIAM Studies in Applied Mathematics, Philadelphia, Pennsylvania, 199. [] L. El Ghaoui and S. I. Niculescu (Editors) \Advances in Linear Matrix Inequality Methods in Control", SIAM Advances in Control and Design, Philadelphia, Pennsylvania, [] T. Iwasaki, G. Meinsma and M. Fu \Generalized Sprocedure and Finite Frequency KYP Lemma", preprint, [] T. Kailath \Linear Systems", Prentice Hall, Englewood Clis, New Jersey, [] The MathWorks, Inc. \LMI Control Toolbox for Matlab", Release 1.0., See the home page [] G. Meinsma, Y. Shrivastava and M. Fu \A Dual Formulation of Mixed and the Losslessness of (D; G)scaling", IEEE Transactions on Automatic Control, Vol., No., pp. 10{10, 199. [8] A. Packard and J. Doyle \The Complex Singular Value", Automatica, Vol. 9, No. 1, pp. 1{109, 199. [9] A. Rantzer \On the KalmanYakubovichPopov Lemma", Systems and Control Letters, Vol. 8, pp. {10, 199. [10] L. Vandenberghe and S. Boyd \Semidenite Programming", SIAM Review, Vol. 8, pp. 9{9, 199. [11] V. A. Yakubovich \The Sprocedure in Nonlinear Control Theory", Vestnik Leningrad University of Mathematics, Vol., pp. {9, 19. In Russian,
I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec  Spring 2010
I.3. LMI DUALITY Didier HENRION henrion@laas.fr EECI Graduate School on Control Supélec  Spring 2010 Primal and dual For primal problem p = inf x g 0 (x) s.t. g i (x) 0 define Lagrangian L(x, z) = g 0
More informationDidier HENRION henrion
POLYNOMIAL METHODS FOR ROBUST CONTROL Didier HENRION www.laas.fr/ henrion henrion@laas.fr Laboratoire d Analyse et d Architecture des Systèmes Centre National de la Recherche Scientifique Université de
More informationLinear Systems with Saturating Controls: An LMI Approach. subject to control saturation. No assumption is made concerning openloop stability and no
Output Feedback Robust Stabilization of Uncertain Linear Systems with Saturating Controls: An LMI Approach Didier Henrion 1 Sophie Tarbouriech 1; Germain Garcia 1; Abstract : The problem of robust controller
More informationROBUST ANALYSIS WITH LINEAR MATRIX INEQUALITIES AND POLYNOMIAL MATRICES. Didier HENRION henrion
GRADUATE COURSE ON POLYNOMIAL METHODS FOR ROBUST CONTROL PART IV.1 ROBUST ANALYSIS WITH LINEAR MATRIX INEQUALITIES AND POLYNOMIAL MATRICES Didier HENRION www.laas.fr/ henrion henrion@laas.fr Airbus assembly
More informationFast Algorithms for SDPs derived from the KalmanYakubovichPopov Lemma
Fast Algorithms for SDPs derived from the KalmanYakubovichPopov Lemma Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 8 September 2003 European Union RTN Summer School on MultiAgent
More informationLecture Note 5: Semidefinite Programming for Stability Analysis
ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State
More informationSemidefinite Programming Duality and Linear Timeinvariant Systems
Semidefinite Programming Duality and Linear Timeinvariant Systems Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 2 July 2004 Workshop on Linear Matrix Inequalities in Control LAASCNRS,
More informationEE 227A: Convex Optimization and Applications October 14, 2008
EE 227A: Convex Optimization and Applications October 14, 2008 Lecture 13: SDP Duality Lecturer: Laurent El Ghaoui Reading assignment: Chapter 5 of BV. 13.1 Direct approach 13.1.1 Primal problem Consider
More informationUC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009
UC Berkeley Department of Electrical Engineering and Computer Science EECS 227A Nonlinear and Convex Optimization Solutions 5 Fall 2009 Reading: Boyd and Vandenberghe, Chapter 5 Solution 5.1 Note that
More informationarzelier
COURSE ON LMI OPTIMIZATION WITH APPLICATIONS IN CONTROL PART II.1 LMIs IN SYSTEMS CONTROL STATESPACE METHODS STABILITY ANALYSIS Didier HENRION www.laas.fr/ henrion henrion@laas.fr Denis ARZELIER www.laas.fr/
More informationStability of linear timevarying systems through quadratically parameterdependent Lyapunov functions
Stability of linear timevarying systems through quadratically parameterdependent Lyapunov functions Vinícius F. Montagner Department of Telematics Pedro L. D. Peres School of Electrical and Computer
More information3. Algebra and Duality
31 Algebra and Duality P. Parrilo and S. Lall, ECC 2003 2003.09.03.01 3. Algebra and Duality Example: nonconvex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone
More information4. Algebra and Duality
41 Algebra and Duality P. Parrilo and S. Lall, CDC 2003 2003.12.07.01 4. Algebra and Duality Example: nonconvex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone
More informationStrong duality in Lasserre s hierarchy for polynomial optimization
Strong duality in Lasserre s hierarchy for polynomial optimization arxiv:1405.7334v1 [math.oc] 28 May 2014 Cédric Josz 1,2, Didier Henrion 3,4,5 Draft of January 24, 2018 Abstract A polynomial optimization
More informationExtreme Abridgment of Boyd and Vandenberghe s Convex Optimization
Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization Compiled by David Rosenberg Abstract Boyd and Vandenberghe s Convex Optimization book is very wellwritten and a pleasure to read. The
More informationA Study of the Duality between Kalman Filters and LQR Problems
Purdue University Purdue epubs Department of Electrical and Computer Engineering Technical Reports Department of Electrical and Computer Engineering 1132016 A Study of the Duality between Kalman Filters
More informationA semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint
Iranian Journal of Operations Research Vol. 2, No. 2, 20, pp. 2934 A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint M. Salahi Semidefinite
More informationLinks Between Robust and Quadratic Stability of. Michael Sebek 4;5. Vladimr Kucera 4;5. Abstract
Links Between Robust and Quadratic Stability of Uncertain DiscreteTime Polynomials 1 Didier Henrion 2;3 Michael Sebek 4;5 Vladimr Kucera 4;5 bstract n uncertain polynomial is robustly stable, or stable
More informationDenis ARZELIER arzelier
COURSE ON LMI OPTIMIZATION WITH APPLICATIONS IN CONTROL PART II.1 LMIs IN SYSTEMS CONTROL STATESPACE METHODS STABILITY ANALYSIS Denis ARZELIER www.laas.fr/ arzelier arzelier@laas.fr 15 Octobre 2008 Statespace
More informationOptimization based robust control
Optimization based robust control Didier Henrion 1,2 Draft of March 27, 2014 Prepared for possible inclusion into The Encyclopedia of Systems and Control edited by John Baillieul and Tariq Samad and published
More informationCONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren
CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren CESAME, Université catholique de Louvain Bâtiment Euler, Avenue G. Lemaître 46 B1348 LouvainlaNeuve,
More informationA new robust delaydependent stability criterion for a class of uncertain systems with delay
A new robust delaydependent stability criterion for a class of uncertain systems with delay Fei Hao Long Wang and Tianguang Chu Abstract A new robust delaydependent stability criterion for a class of
More informationOlivier Bachelier 3. Michael Sebek 4;5. Abstract. polynomial matrix to belong to a given region D of the complex plane.
DStability of Polynomial Matrices 1 Didier Henrion ; Olivier Bachelier Michael Sebek ; Abstract Necessary and sucient conditions are formulated for the zeros of an arbitrary polynomial matrix to belong
More information14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.
CS/ECE/ISyE 524 Introduction to Optimization Spring 2016 17 14. Duality ˆ Upper and lower bounds ˆ General duality ˆ Constraint qualifications ˆ Counterexample ˆ Complementary slackness ˆ Examples ˆ Sensitivity
More informationFrom Convex Optimization to Linear Matrix Inequalities
Dep. of Information Engineering University of Pisa (Italy) From Convex Optimization to Linear Matrix Inequalities eng. Sergio Grammatico grammatico.sergio@gmail.com Class of Identification of Uncertain
More informationHybrid Systems  Lecture n. 3 Lyapunov stability
OUTLINE Focus: stability of equilibrium point Hybrid Systems  Lecture n. 3 Lyapunov stability Maria Prandini DEI  Politecnico di Milano Email: prandini@elet.polimi.it continuous systems decribed by
More informationIntroduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 10 Dr. Ted Ralphs IE406 Lecture 10 1 Reading for This Lecture Bertsimas 4.14.3 IE406 Lecture 10 2 Duality Theory: Motivation Consider the following
More informationThe Simplest Semidefinite Programs are Trivial
The Simplest Semidefinite Programs are Trivial Robert J. Vanderbei Bing Yang Program in Statistics & Operations Research Princeton University Princeton, NJ 08544 January 10, 1994 Technical Report SOR9312
More informationMaximizing the Closed Loop Asymptotic Decay Rate for the TwoMassSpring Control Problem
Maximizing the Closed Loop Asymptotic Decay Rate for the TwoMassSpring Control Problem Didier Henrion 1,2 Michael L. Overton 3 May 12, 2006 Abstract We consider the following problem: find a fixedorder
More informationLMI relaxations in robust control (tutorial)
LM relaxations in robust control tutorial CW Scherer Delft Center for Systems and Control Delft University of Technology Mekelweg 2, 2628 CD Delft, The Netherlands cwscherer@dcsctudelftnl Abstract This
More informationSemidefinite Programming Basics and Applications
Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent
More information3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions
A. LINEAR ALGEBRA. CONVEX SETS 1. Matrices and vectors 1.1 Matrix operations 1.2 The rank of a matrix 2. Systems of linear equations 2.1 Basic solutions 3. Vector spaces 3.1 Linear dependence and independence
More informationLecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.
MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationOn Computing the Worstcase Performance of Lur'e Systems with Uncertain Timeinvariant Delays
Article On Computing the Worstcase Performance of Lur'e Systems with Uncertain Timeinvariant Delays Thapana Nampradit and David Banjerdpongchai* Department of Electrical Engineering, Faculty of Engineering,
More informationFixedOrder Robust H Controller Design with Regional Pole Assignment
SUBMITTED 1 FixedOrder Robust H Controller Design with Regional Pole Assignment Fuwen Yang, Mahbub Gani, and Didier Henrion Abstract In this paper, the problem of designing fixedorder robust H controllers
More informationThe model reduction algorithm proposed is based on an iterative twostep LMI scheme. The convergence of the algorithm is not analyzed but examples sho
Model Reduction from an H 1 /LMI perspective A. Helmersson Department of Electrical Engineering Linkoping University S581 8 Linkoping, Sweden tel: +6 1 816 fax: +6 1 86 email: andersh@isy.liu.se September
More informationSTABILITY OF PLANAR NONLINEAR SWITCHED SYSTEMS
LABORATOIRE INORMATIQUE, SINAUX ET SYSTÈMES DE SOPHIA ANTIPOLIS UMR 6070 STABILITY O PLANAR NONLINEAR SWITCHED SYSTEMS Ugo Boscain, régoire Charlot Projet TOpModel Rapport de recherche ISRN I3S/RR 200407
More information1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations
The Simplex Method Most textbooks in mathematical optimization, especially linear programming, deal with the simplex method. In this note we study the simplex method. It requires basically elementary linear
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationExample: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma
41 Algebra and Duality P. Parrilo and S. Lall 2006.06.07.01 4. Algebra and Duality Example: nonconvex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone of valid
More information5. Duality. Lagrangian
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More information61 The Positivstellensatz P. Parrilo and S. Lall, ECC
61 The Positivstellensatz P. Parrilo and S. Lall, ECC 2003 2003.09.02.10 6. The Positivstellensatz Basic semialgebraic sets Semialgebraic sets TarskiSeidenberg and quantifier elimination Feasibility
More informationInput: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function
Linear programming Input: System of inequalities or equalities over the reals R A linear cost function Output: Value for variables that minimizes cost function Example: Minimize 6x+4y Subject to 3x + 2y
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More information1 Outline Part I: Linear Programming (LP) InteriorPoint Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin
Sensitivity Analysis in LP and SDP Using InteriorPoint Methods E. Alper Yldrm School of Operations Research and Industrial Engineering Cornell University Ithaca, NY joint with Michael J. Todd INFORMS
More informationMAT Linear Algebra Collection of sample exams
MAT 342  Linear Algebra Collection of sample exams Ax. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 20182019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationLinear and nonlinear programming
Linear and nonlinear programming Benjamin Recht March 11, 2005 The Gameplan Constrained Optimization Convexity Duality Applications/Taxonomy 1 Constrained Optimization minimize f(x) subject to g j (x)
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationMarcus Pantoja da Silva 1 and Celso Pascoli Bottura 2. Abstract: Nonlinear systems with timevarying uncertainties
A NEW PROPOSAL FOR H NORM CHARACTERIZATION AND THE OPTIMAL H CONTROL OF NONLINEAR SSTEMS WITH TIMEVARING UNCERTAINTIES WITH KNOWN NORM BOUND AND EXOGENOUS DISTURBANCES Marcus Pantoja da Silva 1 and Celso
More informationChapter 1. Preliminaries
Introduction This dissertation is a reading of chapter 4 in part I of the book : Integer and Combinatorial Optimization by George L. Nemhauser & Laurence A. Wolsey. The chapter elaborates links between
More information1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad
Quadratic Maximization and Semidenite Relaxation Shuzhong Zhang Econometric Institute Erasmus University P.O. Box 1738 3000 DR Rotterdam The Netherlands email: zhang@few.eur.nl fax: +3110408916 August,
More informationConvex Optimization 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski Convex Optimization 1 Many optimization objectives generated
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming
E5295/5B5749 Convex optimization with engineering applications Lecture 5 Convex programming and semidefinite programming A. Forsgren, KTH 1 Lecture 5 Convex optimization 2006/2007 Convex quadratic program
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the ndimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the ndimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationRational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution
Preprints of the 18th FAC World Congress Milano (taly) August 28  September 2, 2011 Rational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution Y
More informationwhere m r, m c and m C are the number of repeated real scalar blocks, repeated complex scalar blocks and full complex blocks, respectively. A. (D; G)
1 Some properties of an upper bound for Gjerrit Meinsma, Yash Shrivastava and Minyue Fu Abstract A convex upper bound of the mixed structured singular value is analyzed. The upper bound is based on a multiplier
More informationExercise Sheet 1.
Exercise Sheet 1 You can download my lecture and exercise sheets at the address http://sami.hust.edu.vn/giangvien/?name=huynt 1) Let A, B be sets. What does the statement "A is not a subset of B " mean?
More informationSpectral Theorem for Selfadjoint Linear Operators
Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;
More informationMIT Algebraic techniques and semidefinite optimization February 14, Lecture 3
MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150  Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationLMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAASCNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009
LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAASCNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix
More informationSTABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS. Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec
STABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec DASCTCUFSC P.O. Box 476, 88040900 Florianópolis, SC,
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationHybrid Systems Course Lyapunov stability
Hybrid Systems Course Lyapunov stability OUTLINE Focus: stability of an equilibrium point continuous systems decribed by ordinary differential equations (brief review) hybrid automata OUTLINE Focus: stability
More information1 Number Systems and Errors 1
Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floatingpoint Representation...........
More informationChapter 5 Eigenvalues and Eigenvectors
Chapter 5 Eigenvalues and Eigenvectors Outline 5.1 Eigenvalues and Eigenvectors 5.2 Diagonalization 5.3 Complex Vector Spaces 2 5.1 Eigenvalues and Eigenvectors Eigenvalue and Eigenvector If A is a n n
More informationCSCI 1951G Optimization Methods in Finance Part 01: Linear Programming
CSCI 1951G Optimization Methods in Finance Part 01: Linear Programming January 26, 2018 1 / 38 Liability/asset cashflow matching problem Recall the formulation of the problem: max w c 1 + p 1 e 1 = 150
More informationDuality. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36725
Duality Barnabas Poczos & Ryan Tibshirani Convex Optimization 10725/36725 1 Duality in linear programs Suppose we want to find lower bound on the optimal value in our convex problem, B min x C f(x) E.g.,
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationConvex Optimization Boyd & Vandenberghe. 5. Duality
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationI. Multiple Choice Questions (Answer any eight)
Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam  Course Instructor : Prashanth L.A. Date : Sep24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY
More informationUniversity of Twente. Faculty of Mathematical Sciences. On stability robustness with respect to LTV uncertainties
Faculty of Mathematical Sciences University of Twente University for Technical and Social Sciences P.O. Box 17 75 AE Enschede The Netherlands Phone: +315348934 Fax: +31534893114 Email: memo@math.utwente.nl
More informationAn Observation on the Positive Real Lemma
Journal of Mathematical Analysis and Applications 255, 48 49 (21) doi:1.16/jmaa.2.7241, available online at http://www.idealibrary.com on An Observation on the Positive Real Lemma Luciano Pandolfi Dipartimento
More informationLinear Matrix Inequality (LMI)
Linear Matrix Inequality (LMI) A linear matrix inequality is an expression of the form where F (x) F 0 + x 1 F 1 + + x m F m > 0 (1) x = (x 1,, x m ) R m, F 0,, F m are real symmetric matrices, and the
More informationLINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM
LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator
More informationDuality. Geoff Gordon & Ryan Tibshirani Optimization /
Duality Geoff Gordon & Ryan Tibshirani Optimization 10725 / 36725 1 Duality in linear programs Suppose we want to find lower bound on the optimal value in our convex problem, B min x C f(x) E.g., consider
More informationStatic Output Feedback Stabilisation with H Performance for a Class of Plants
Static Output Feedback Stabilisation with H Performance for a Class of Plants E. Prempain and I. Postlethwaite Control and Instrumentation Research, Department of Engineering, University of Leicester,
More informationUniversity of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm
University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm Name: The proctor will let you read the following conditions
More informationJune Engineering Department, Stanford University. System Analysis and Synthesis. Linear Matrix Inequalities. Stephen Boyd (E.
Stephen Boyd (E. Feron :::) System Analysis and Synthesis Control Linear Matrix Inequalities via Engineering Department, Stanford University Electrical June 1993 ACC, 1 linear matrix inequalities (LMIs)
More informationIntroduction to Semidefinite Programming I: Basic properties a
Introduction to Semidefinite Programming I: Basic properties and variations on the GoemansWilliamson approximation algorithm for maxcut MFO seminar on Semidefinite Programming May 30, 2010 Semidefinite
More informationOn the application of different numerical methods to obtain nullspaces of polynomial matrices. Part 1: block Toeplitz algorithms.
On the application of different numerical methods to obtain nullspaces of polynomial matrices. Part 1: block Toeplitz algorithms. J.C. Zúñiga and D. Henrion Abstract Four different algorithms are designed
More informationA Review of Linear Programming
A Review of Linear Programming Instructor: Farid Alizadeh IEOR 4600y Spring 2001 February 14, 2001 1 Overview In this note we review the basic properties of linear programming including the primal simplex
More information2nd Symposium on System, Structure and Control, Oaxaca, 2004
263 2nd Symposium on System, Structure and Control, Oaxaca, 2004 A PROJECTIVE ALGORITHM FOR STATIC OUTPUT FEEDBACK STABILIZATION Kaiyang Yang, Robert Orsi and John B. Moore Department of Systems Engineering,
More informationCOURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion
COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS Didier HENRION www.laas.fr/ henrion October 2006 Geometry of LMI sets Given symmetric matrices F i we want to characterize the shape in R n of the LMI set F
More informationAn Exact Stability Analysis Test for SingleParameter. PolynomiallyDependent Linear Systems
An Exact Stability Analysis Test for SingleParameter PolynomiallyDependent Linear Systems P. Tsiotras and P.A. Bliman Abstract We provide a new condition for testing the stability of a singleparameter,
More informationhomogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45
address 12 adjoint matrix 118 alternating 112 alternating 203 angle 159 angle 33 angle 60 area 120 associative 180 augmented matrix 11 axes 5 Axiom of Choice 153 basis 178 basis 210 basis 74 basis test
More informationOptimization Theory. A Concise Introduction. Jiongmin Yong
October 11, 017 16:5 wsbook9x6 Book Title Optimization Theory 01708Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 01708Lecture Notes page Optimization
More informationRobust and Optimal Control, Spring 2015
Robust and Optimal Control, Spring 2015 Instructor: Prof. Masayuki Fujita (S5303B) G. Sum of Squares (SOS) G.1 SOS Program: SOS/PSD and SDP G.2 Duality, valid ineqalities and Cone G.3 Feasibility/Optimization
More informationApplications of Controlled Invariance to the l 1 Optimal Control Problem
Applications of Controlled Invariance to the l 1 Optimal Control Problem Carlos E.T. Dórea and JeanClaude Hennet LAASCNRS 7, Ave. du Colonel Roche, 31077 Toulouse Cédex 4, FRANCE Phone : (+33) 61 33
More informationOptimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers
Optimization for Communications and Networks Poompat Saengudomlert Session 4 Duality and Lagrange Multipliers P Saengudomlert (2015) Optimization Session 4 1 / 14 24 Dual Problems Consider a primal convex
More informationModule 04 Optimization Problems KKT Conditions & Solvers
Module 04 Optimization Problems KKT Conditions & Solvers Ahmad F. Taha EE 5243: Introduction to CyberPhysical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html September
More informationUNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems
UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems Robert M. Freund February 2016 c 2016 Massachusetts Institute of Technology. All rights reserved. 1 1 Introduction
More informationMidterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
Midterm Review Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye (LY, Chapter 14, Appendices) 1 Separating hyperplane
More informationSemidefinite Programming
Semidefinite Programming Notes by Bernd Sturmfels for the lecture on June 26, 208, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra The transition from linear algebra to nonlinear algebra has
More informationw T 1 w T 2. w T n 0 if i j 1 if i = j
Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible 
More informationResearch Article Indefinite LQ Control for DiscreteTime Stochastic Systems via Semidefinite Programming
Mathematical Problems in Engineering Volume 2012, Article ID 674087, 14 pages doi:10.1155/2012/674087 Research Article Indefinite LQ Control for DiscreteTime Stochastic Systems via Semidefinite Programming
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More information