Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

Similar documents
I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Didier HENRION henrion

Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma

Linear Systems with Saturating Controls: An LMI Approach. subject to control saturation. No assumption is made concerning open-loop stability and no

ROBUST ANALYSIS WITH LINEAR MATRIX INEQUALITIES AND POLYNOMIAL MATRICES. Didier HENRION henrion

Lecture Note 5: Semidefinite Programming for Stability Analysis

arzelier

Semidefinite Programming Duality and Linear Time-invariant Systems

EE 227A: Convex Optimization and Applications October 14, 2008

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Strong duality in Lasserre s hierarchy for polynomial optimization

Stability of linear time-varying systems through quadratically parameter-dependent Lyapunov functions

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

4. Algebra and Duality

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

A Study of the Duality between Kalman Filters and LQR Problems

Links Between Robust and Quadratic Stability of. Michael Sebek 4;5. Vladimr Kucera 4;5. Abstract

The Simplest Semidefinite Programs are Trivial

Semidefinite Programming Basics and Applications

CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren

Optimization based robust control

A new robust delay-dependent stability criterion for a class of uncertain systems with delay

Olivier Bachelier 3. Michael Sebek 4;5. Abstract. polynomial matrix to belong to a given region D of the complex plane.

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

From Convex Optimization to Linear Matrix Inequalities

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Hybrid Systems - Lecture n. 3 Lyapunov stability

Maximizing the Closed Loop Asymptotic Decay Rate for the Two-Mass-Spring Control Problem

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

LMI relaxations in robust control (tutorial)

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Foundations of Matrix Analysis

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

On Computing the Worst-case Performance of Lur'e Systems with Uncertain Time-invariant Delays

Fixed-Order Robust H Controller Design with Regional Pole Assignment

Exercise Sheet 1.

The model reduction algorithm proposed is based on an iterative two-step LMI scheme. The convergence of the algorithm is not analyzed but examples sho

STABILITY OF PLANAR NONLINEAR SWITCHED SYSTEMS

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

5. Duality. Lagrangian

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Linear Algebra Massoud Malek

Chapter 5 Eigenvalues and Eigenvectors

Linear Algebra: Matrix Eigenvalue Problems

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

MAT Linear Algebra Collection of sample exams

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Linear and non-linear programming

Linear Matrix Inequality (LMI)

Lecture 7: Positive Semidefinite Matrices

University of Twente. Faculty of Mathematical Sciences. On stability robustness with respect to LTV uncertainties

Marcus Pantoja da Silva 1 and Celso Pascoli Bottura 2. Abstract: Nonlinear systems with time-varying uncertainties

Chapter 1. Preliminaries

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Introduction to Semidefinite Programming I: Basic properties a

Rational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution

Convex Optimization 1

Spectral Theorem for Self-adjoint Linear Operators

where m r, m c and m C are the number of repeated real scalar blocks, repeated complex scalar blocks and full complex blocks, respectively. A. (D; G)-

STABILITY AND STABILIZATION OF A CLASS OF NONLINEAR SYSTEMS WITH SATURATING ACTUATORS. Eugênio B. Castelan,1 Sophie Tarbouriech Isabelle Queinnec

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

1 Number Systems and Errors 1

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Hybrid Systems Course Lyapunov stability

Convex Optimization Boyd & Vandenberghe. 5. Duality

I. Multiple Choice Questions (Answer any eight)

Linear Algebra Review. Vectors

An Observation on the Positive Real Lemma

Dual LMI Approach to Linear Positive System Analysis

Semidefinite Programming

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

Static Output Feedback Stabilisation with H Performance for a Class of Plants

June Engineering Department, Stanford University. System Analysis and Synthesis. Linear Matrix Inequalities. Stephen Boyd (E.

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

MATRIX ALGEBRA. or x = (x 1,..., x n ) R n. y 1 y 2. x 2. x m. y m. y = cos θ 1 = x 1 L x. sin θ 1 = x 2. cos θ 2 = y 1 L y.

On the application of different numerical methods to obtain null-spaces of polynomial matrices. Part 1: block Toeplitz algorithms.

2nd Symposium on System, Structure and Control, Oaxaca, 2004

Linear Algebra: Characteristic Value Problem

An Exact Stability Analysis Test for Single-Parameter. Polynomially-Dependent Linear Systems

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

A Review of Linear Programming

Lecture 14: Optimality Conditions for Conic Problems

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

Applications of Controlled Invariance to the l 1 Optimal Control Problem

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Robust and Optimal Control, Spring 2015

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

Elementary linear algebra

Transcription:

Rank-one LMIs and Lyapunov's Inequality Didier Henrion 1;; Gjerrit Meinsma Abstract We describe a new proof of the well-known Lyapunov's matrix inequality about the location of the eigenvalues of a matrix in some region of the complex plane. The proof makes use of standard facts from quadratic and semi-denite programming. Links are established between the Lyapunov matrix, rank-one LMIs and the Lagrange multiplier arising in duality theory. Keywords Linear Systems, Stability, LMI. 1 Introduction Let A C nn be a given complex matrix and let 1 a D = fs C : s b b c 1 s < 0g denote a given open region of the complex plane, where Hermitian matrix a b b C c has one strictly negative eigenvalue and one strictly positive eigenvalue, and the star denotes transpose conjugate. In the sequel, the notation P 0 or P 0 (resp. P 0 or P 0) means that matrix P is positive denite (resp. semi-denite). The location of the eigenvalues of A can be characterized as follows. 1 Corresponding author. FAX: + 1 9 9. E-mail: henrion@laas.fr Laboratoire d'architecture et d'analyse des Systemes, Centre National de la Recherche Scientique, Avenue du Colonel Roche, Toulouse, cedex, France. Institut National des Sciences Appliquees, Complexe Scientique de Rangueil, Toulouse, cedex, France. Faculty of Applied Mathematics, University of Twente, P.O.Box 1, Enschede, 00 AE, The Netherlands. 1

Theorem 1 (Lyapunov's Inequality) Matrix A has all its eigenvalues in region D if and only if there is a matrix P = P 0 C nn such that I ap bp I A b 0: (1) P cp A Matrix inequality (1) is referred to as Lyapunov's inequality. Lyapunov's proof of the above theorem { originally developed in the case that D is the open left half-plane, i.e. a = c = 0, b = 1 and inequality (1) becomes A P + P A 0 { relies on the construction of a positive quadratic function whose derivative is negative along the trajectories of an associated dynamical system, see e.g. []. It can be extended to arbitrary regions D via a conformal mapping. Another proof of Theorem 1 can be found in [1, Theorem.19]. Eigenvectors of matrix A are used to show that existence of P implies stability of A, whereas the converse statement is shown via properties of the matrix exponential function. The aim of this note is to give a new, alternative proof of Lyapunov's inequality without referring to stability of the trajectories of a dynamical system or to matrix exponentials. We use elementary concepts from linear algebra, quadratic and semi-denite programming. Links are established between the Lyapunov matrix and the Lagrange multiplier arising in duality theory. Relationships with rank-one LMIs, the Kalman-Yakubovich- Popov Lemma and (D; G)-scaling in -analysis are also pointed out. The proof relies on the following important result, proved e.g. as in [9, Lemma ]. Lemma 1 Two column vectors p; q C n with q non-zero satisfy a b [q p] [q p] 0 () b c if and only if p = sq for some s D C. Rank-one LMI Problem First we show the equivalence between location of the eigenvalues of A in region D and a rank-one LMI optimization problem or a rank-one LMI feasibility problem. If s C is an eigenvalue of A, then there exists a non-zero vector q C n such that (A si)q = 0: () Pursuing an idea proposed in [, Chapter 1], it follows that all the eigenvalues of A belong to D if and only if the optimal value of the quadratic optimization problem = min q (A si) (A si)q s:t: s D C q q = 1 ()

is strictly positive, where 1 D C = fs C : s a b b c 1 s 0g is the closed region complementary to D in C. Dene the rank-one positive semidenite matrix q q X = xx = 0 p p and use the notations A(s)q = [ A I ] x = Ax q = [ I 0 ] x = Qx p = [ 0 I ] x = Px to write inequality () as an LMI in rank-one matrix X, namely ax b F (X) = Q P X bx cx Q P 0 () where F is a linear map from C nn to C nn. Using these notations, an alternative formulation of quadratic optimization problem () is given by the following lemma. Lemma The eigenvalues of matrix A belong to region D if and only if > 0 in rank-one LMI optimization problem = min Trace A AX s:t: F (X) 0 X = X 0 Trace Q QX = 1 Rank X = 1: () The above rank-one LMI problem is an optimization problem. It turns out that we can equivalently state this result via a feasibility problem, following an idea exposed in []. To see this, note that Lemma 1 and equation () imply that A has no eigenvalue in D C if and only if there is no non-zero vector q for which a b [q Aq] [q Aq] 0: () b c The left-hand side of inequality () can alternatively be expressed as aqq + baqq + b qq A + caqq A = I A aqq b qq Now dene the linear map bqq cqq I A : aq b G(Q) = Q I A bq cq I A 0 (8) from C nn to C nn. With Q denoting the non-zero rank-one matrix Q = qq we arrive at the following result which is equivalent to Lemma.

Lemma The eigenvalues of matrix A belong to region D if and only if there is no solution to the rank-one LMI feasibility problem G(Q) 0 Q = Q 0 Trace Q = 1 Rank Q = 1: (9) LMI Problem Now we show that the non-convex rank constraints in LMI problems () and (9) are actually irrelevant. Let I N = A denote a matrix whose columns span the n-dimensional right null-space of full row-rank matrix A. If s k C is a non-defective eigenvalue of A (i.e. its algebraic multiplicity is equal to its geometric multiplicity) and q k C n is the corresponding eigenvector, then the vector qk x k = s k q k belongs to the right null-space of matrix A. Similarly, if s k is a defective eigenvalue of A (i.e. its algebraic multiplicity is greater than its geometric multiplicity), then the corresponding chain of linearly independent generalized eigenvectors q k ; q k+1 ; q k+ ; : : : gives rise to vectors x k = qk s k q k x k+1 = q k+1 s k q k+1 + q k also belonging to the right null-space of A. Let x k+ = V = [x 1 x n ] q k+ s k q k+ + q k+1 (10) denote a matrix built up from all the vectors x i associated with all the eigenvalues s i of A. It follows from the above discussion that the columns of N and V span the same vector space. By denition, vectors q i are linearly independent, thus we can dene linearly independent vectors q i C n such that [q 1 q n ] [q 1 q n ] = I: (11) Following these preliminaries, consider now the following relaxation of rank-one LMI problem () = min Trace A AX s:t: F (X) 0 (1) X = X 0 Trace Q QX = 1 where the non-convex rank constraint has been dropped. Since the non-convex feasible set in problem () is a subset of the convex feasible set in problem (1), LMI optimization problem (1) is referred to as a convex relaxation of the non-convex rank-one LMI problem (). In relation to the above problem, we can state the following central result.

Lemma > 0 in rank-one LMI optimization problem () if and only if > 0 in LMI optimization problem (1). Proof The inner product of positive semi-denite matrices A A and X is always nonnegative, hence 0. Moreover, the fact that > 0 implies > 0 is trivial since the feasible set in problem () is a subset of the feasible set in problem (1), i.e. it holds. Consequently, in order to show that > 0 implies > 0, the remainder of the proof will consist in proving that = 0 implies = 0. So suppose that X is a positive semi-denite matrix such that = 0 in problem (1). Let W be a n r full column rank matrix such that X = W W. By putting matrix A A into Schur form, it can easily be shown that Trace A AW W = 0 implies A AW W = 0. Consequently, the columns of W span a subspace that belongs to the right null-space of A. In view of the above denition of matrix V, there exists a matrix M such that W = V M. Let m ij denote the entries of positive semi-denite matrix MM C nn. For a given index k, it holds either m kk > 0 or m ik = m ki = 0 for all i = 1; : : : ; n. Matrix X is feasible for problem (1) thus F (X) = F (V MM V ) = nx i=1 nx j=1 m ij F (x i x j) 0: (1) Since matrix X cannot be zero by assumption, matrix MM is also non-zero and there exists at least one index k such that m kk > 0. Let x k+l be the last eigenvector in the chain of generalized eigenvectors with eigenvalue s k for which m (k+l)(k+l) is non-zero (note that l = 0 if s k is non-defective). From relations (), (10), (11) and (1) it follows that q k+l F (X)q k+l = m (k+l)(k+l) (a + bs k + b s k + cs k s k) 0: Since m (k+l)(k+l) > 0 we see that s k D C, hence vector x k in equation (10) is such that Trace A Ax k x k = 0 and F (x kx k ) 0 in virtue of Lemma 1. Consequently, matrix x kx k is a solution to rank-one LMI problem () such that = 0 and the lemma is proved. The following result is then a straightforward corollary to Lemma : Lemma The eigenvalues of matrix A belong to region D if and only if > 0 in LMI optimization problem (1). Now consider the following relaxation to rank-one LMI feasibility problem (9): G(Q) 0 Q = Q 0 Trace; Q = 1 (1) where the rank constraint has been dropped. Using the same kind of arguments as above, we can show the following counterpart to Lemma : Lemma The eigenvalues of matrix A belong to region D if and only if there is no solution to LMI feasibility problem (1).

Dual LMI Problem Now we use standard semidenite programming duality results [10] to come up with a more compact formulation of the stability conditions of Lemmas and and prove the Lyapunov's inequality of Theorem 1. Dene the linear map Q F D (P ) = P ap b P bp Q cp P dual to the map introduced in (). It is easy to show that ap = b P Trace F D (P )X = Trace F (X)P: Using standard duality arguments, we now prove that the LMI feasibility problem A A F D (P ) P = P 0 is dual to LMI optimization problem (1). To see this, build the Lagrangian bp cp L(P; X; Y ) = Trace (A A F D (P ))X Trace P Y = Trace A AX + Trace (F (X) Y )P of problem (1) where X = X 0 and Y = Y 0 are Lagrange multiplier matrices. The dual function associated with the Lagrangian reads Trace A AX iff (X) = Y 0 g(x; Y ) = min L(P; X; Y ) = P 1 otherwise: The dual optimization problem, obtained by maximizing dual function g(x; Y ) is therefore LMI optimization problem (1), where the equality constraint Trace Q QX = 1 ensures compactness of the feasible set. The matrix inequalities in problem (1) are strict, hence there is no duality gap and > 0 in LMI optimization problem (1) if and only if LMI problem (1) is feasible. Recall that N denotes a matrix whose columns span the right null-space of A. Then it follows from the Elimination Lemma [] that feasibility problem (1) can equivalently be written as This is exactly the statement of Theorem 1. Similarly, we can dene G D (P ) = N F D (P )N = N F D (P )N 0 P = P 0: I A ap b P bp cp I A as the linear map from C nn to C nn dual to the linear map G(Q) introduced in (8). It is easy to show that Trace G D (P )Q = Trace G(Q)P: It now follows that non-existence of a non-zero Q = Q 0 for which G(Q) 0 is equivalent to the existence of P = P 0 for which G D (P ) 0. In other words we proved Theorem 1. (1) (1) (1)

Numerical Examples.1 First Example Let A = 1 be a constant matrix with eigenvalues -1 and - and let be the stability region. Primal LMI problem (1) reads = min Trace s:t: D = fs C : s + s < 0g 1 1 1 X X = X 0 Trace 1 0 + X = 1: X X 0 With a relative accuracy of 10 8, the LMI Control Toolbox 1.0. for Matlab. [] returns and X = = 0:191 0:818198 0:08098 0:000000000:0000000000 0:08098 0:18180:000000000:0000000000 0:000000000:000000000:000000000:0000000000 0:000000000:000000000:000000000:0000000000 as the optimum of the above problem. In virtue of Theorem, is strictly positive hence all the eigenvalues of matrix A belong to region D.

Dual LMI problem (1) reads 1 P = P 0: B @ 0 P + P 1 C A 1 0 With the help of the LMI Toolbox, we obtained the matrix P = 1:8810 0:108 0:108 0:09190 as a feasible solution for the above problem. On the other hand, LMI problem (1) reads 1 0 Q Q 0 1 0 Q = Q 0; Trace Q = 1: This problem is infeasible, which is consistent with the above results and Theorem 1.. Second Example Now let A = 1 be a constant matrix with eigenvalues - and and let be the stability region. D = fs C : s + s < 0g With a relative accuracy of 10 8, the LMI Toolbox returns and X = = 0:0000000000 0:100000000:00000000:00000000:000000000 0:00000000:900000000:000000000 :000000000 0:00000000:000000000 :000000000 :000000000 1:000000000 :000000000 :000000000 :000000000 as the optimum of primal problem (1). In virtue of Theorem, some eigenvalues of matrix A do not belong to region D. One can check that X = xx = 0:10 0:988981 1:8118801 :190 8 0:10 0:988981 1:8118801 :190

is actually a rank-one solution to LMI problem (1). Vector x can be written as q x = sq where q is an eigenvector of matrix A corresponding to the eigenvalue s = D C. One can check that positive semidenite matrix 0:100000000:000000000 Q = qq = 0:00000000:9000000000 is a feasible solution for LMI problems (9) or (1). On the other hand, dual LMI problem (1) is found infeasible, which is consistent with the above results and Theorem 1. Conclusion We have proposed a new proof of Lyapunov's matrix inequality that relies on elementary optimization techniques and linear algebra. Following ideas proposed in [] and [, Chapter 1], we consider the eigenvalue location problem as a mere quadratic optimization problem. Then, the quadratic problem can be formulated as an LMI problem with a nonconvex rank constraint. The Lyapunov matrix can be viewed as a Lagrange multiplier matrix arising when dualizing this problem. In [, Chapter 1, x1..], it is shown that removing the non-convex rank-one constraint leads to a sucient LMI stability condition. Our contribution is in showing in Lemmas and that the LMI conditions are also necessary. In other words, the rank constraint in problems () and (9) are irrelevant as far as eigenvalue location is concerned. In a similar fashion, the eigenvalue location problem can be viewed as a frequencydependent -analysis problem with one repeated scalar block si corresponding to the Laplace variable s. The Lyapunov matrix P plays the role of a D-scaling matrix associated with the repeated scalar block, and the irrelevance of the non-convex rank constraint readily follows from the losslessness of the (D; G)-scaling as pointed out in [8]. Equivalence of primal problem (1) and dual problem (1) can also be shown via geometric arguments similar to that used in the proof of the Kalman-Yakubovich-Popov (KYP) Lemma in [9, Theorem 1], in the proof of losslessness of (D; G)-scaling [, Lemma.1], in the S-procedure [11] or also in the generalized S-procedure proposed in [, Theorem 1]. Our approach is also very similar in spirit to the one pursued in [9] to provide an alternative proof of the KYP Lemma. Note however that in this reference the author considers a version of the KYP Lemma where the Laplace variable s varies on the imaginary axis or the unit circle. This result has been extended to other one-dimensional curves of the complex plane such as the real axis [] or a segment on the imaginary axis []. These curves are boundaries of the two-dimensional stability regions D considered in the present note. It is therefore expected that we can similarly derive more general versions of the KYP Lemma in two-dimensional stability regions. 9

Finally, we are currently investigating the application of these techniques to the study of stability of polynomial matrices, two-indeterminate polynomial matrices and uncertain polynomial matrices. Related results will be reported elsewhere. References [1] S. Barnett \Polynomials and Linear Control Systems", Marcel Dekker, New York, 198. [] S. Boyd, L. El Ghaoui, E. Feron and V. Balakrishnan \Linear Matrix Inequalities in System and Control Theory", SIAM Studies in Applied Mathematics, Philadelphia, Pennsylvania, 199. [] L. El Ghaoui and S. I. Niculescu (Editors) \Advances in Linear Matrix Inequality Methods in Control", SIAM Advances in Control and Design, Philadelphia, Pennsylvania, 1999. [] T. Iwasaki, G. Meinsma and M. Fu \Generalized S-procedure and Finite Frequency KYP Lemma", preprint, 1999. [] T. Kailath \Linear Systems", Prentice Hall, Englewood Clis, New Jersey, 1980. [] The MathWorks, Inc. \LMI Control Toolbox for Matlab", Release 1.0., 1998. See the home page www.mathworks.com. [] G. Meinsma, Y. Shrivastava and M. Fu \A Dual Formulation of Mixed and the Losslessness of (D; G)-scaling", IEEE Transactions on Automatic Control, Vol., No., pp. 10{10, 199. [8] A. Packard and J. Doyle \The Complex Singular Value", Automatica, Vol. 9, No. 1, pp. 1{109, 199. [9] A. Rantzer \On the Kalman-Yakubovich-Popov Lemma", Systems and Control Letters, Vol. 8, pp. {10, 199. [10] L. Vandenberghe and S. Boyd \Semidenite Programming", SIAM Review, Vol. 8, pp. 9{9, 199. [11] V. A. Yakubovich \The S-procedure in Nonlinear Control Theory", Vestnik Leningrad University of Mathematics, Vol., pp. {9, 19. In Russian, 191. 10