Positive semidefinite intervals for matrix pencils.

Similar documents
POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS

Investigations of nonlinear p-y piles and pile groups in soft clay subjected to static loading-distributed parameter sensitivity analysis.

Lecture Note 5: Semidefinite Programming for Stability Analysis

Some inequalities for sum and product of positive semide nite matrices

Absolute value equations

Identifying Redundant Linear Constraints in Systems of Linear Matrix. Inequality Constraints. Shafiu Jibrin

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Uniqueness of the Solutions of Some Completion Problems

Characterization of half-radial matrices

Yongge Tian. China Economics and Management Academy, Central University of Finance and Economics, Beijing , China

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

Matrix Inequalities by Means of Block Matrices 1

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Solution of Linear Equations

Rank and inertia optimizations of two Hermitian quadratic matrix functions subject to restrictions with applications

Numerical Linear Algebra Homework Assignment - Week 2

14.2 QR Factorization with Column Pivoting

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

There are six more problems on the next two pages

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Copositive Plus Matrices

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix

STAT200C: Review of Linear Algebra

A top nine list: Most popular induced matrix norms

Linear Algebra and its Applications

On the Schur Complement of Diagonally Dominant Matrices

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.

Chap 3. Linear Algebra

Solutions of a constrained Hermitian matrix-valued function optimization problem with applications

Global Quadratic Minimization over Bivalent Constraints: Necessary and Sufficient Global Optimality Condition

Lecture notes: Applied linear algebra Part 1. Version 2

Generalized Principal Pivot Transform

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Analytical formulas for calculating extremal ranks and inertias of quadratic matrix-valued functions and their applications

Math113: Linear Algebra. Beifang Chen

Cospectral graphs and the generalized adjacency matrix

Homework 2 Foundations of Computational Math 2 Spring 2019

A Brief Outline of Math 355

CHARACTERIZATIONS. is pd/psd. Possible for all pd/psd matrices! Generating a pd/psd matrix: Choose any B Mn, then

Symmetric Matrices and Eigendecomposition

7. Symmetric Matrices and Quadratic Forms

Moore Penrose inverses and commuting elements of C -algebras

Cheat Sheet for MATH461

AMS526: Numerical Analysis I (Numerical Linear Algebra)

A Note on Simple Nonzero Finite Generalized Singular Values

Some Reviews on Ranks of Upper Triangular Block Matrices over a Skew Field

Math 408 Advanced Linear Algebra

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Primitive Digraphs with Smallest Large Exponent

Numerical Methods for Solving Large Scale Eigenvalue Problems

Linear algebra for computational statistics

1 Last time: least-squares problems

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Ma 227 Review for Systems of DEs

The extreme rays of the 5 5 copositive cone

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Steven J. Leon University of Massachusetts, Dartmouth

I. Multiple Choice Questions (Answer any eight)

The antitriangular factorisation of saddle point matrices

On Euclidean distance matrices

Lecture 7: Positive Semidefinite Matrices

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

A Note on Eigenvalues of Perturbed Hermitian Matrices

AN EXTREMUM PROPERTY OF SUMS OF EIGENVALUES' HELMUT WIELANDT

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Intrinsic products and factorizations of matrices

Linear Algebra Primer

ETNA Kent State University

Permutation transformations of tensors with an application

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz

On the adjacency matrix of a block graph

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

ACI-matrices all of whose completions have the same rank

12. Cholesky factorization

Symmetric Norm Inequalities And Positive Semi-Definite Block-Matrices

MAT Linear Algebra Collection of sample exams

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

Spectral inequalities and equalities involving products of matrices

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.

SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*

An Algorithm for Solving the Convex Feasibility Problem With Linear Matrix Inequality Constraints and an Implementation for Second-Order Cones

arxiv: v3 [math.ra] 10 Jun 2016

Generalized Prandtl-Ishlinskii Hysteresis Model and its Analytical Inverse for Compensation of Hysteresis in Smart Actuators. Mohammed Al Janaideh

~ g-inverses are indeed an integral part of linear algebra and should be treated as such even at an elementary level.

Lecture 18. Ramanujan Graphs continued

Section 3.2. Multiplication of Matrices and Multiplication of Vectors and Matrices

Math 321: Linear Algebra

Application of Theorems of Schur and Albert

Matrix decompositions

Tensor Complementarity Problem and Semi-positive Tensors

0.1 Rational Canonical Forms

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Matrix Theory. A.Holst, V.Ufnarovski

Transcription:

University of Windsor Scholarship at UWindsor Electronic Theses and Dissertations 1-1-2004 Positive semidefinite intervals for matrix pencils. Huiming Song University of Windsor Follow this and additional works at: https://scholar.uwindsor.ca/etd Recommended Citation Song, Huiming, "Positive semidefinite intervals for matrix pencils." (2004). Electronic Theses and Dissertations. 7159. https://scholar.uwindsor.ca/etd/7159 This online database contains the full-text of PhD dissertations and Masters theses of University of Windsor students from 1954 forward. These documents are made available for personal study and research purposes only, in accordance with the Canadian Copyright Act and the Creative Commons license CC BY-NC-ND (Attribution, Non-Commercial, No Derivative Works). Under this license, works must always be attributed to the copyright holder (original author), cannot be used for any commercial purposes, and may not be altered. Any other use would require the permission of the copyright holder. Students may inquire about withdrawing their dissertation and/or thesis from this database. For additional inquiries, please contact the repository administrator via email (scholarship@uwindsor.ca) or by telephone at 519-253-3000ext. 3208.

Positive Semidefinite Intervals for Matrix Pencils by Huiming Song A Thesis Submitted to the Faculty of Graduate Studies and Research through the Department of Mathematics and Statistics in Partial Fulfillment of the Requirements for the Degree of Master of Science at the University of Windsor Windsor, Ontario, Canada 2004 Huiming Song

1*1 Library and Archives Canada Published Heritage Branch 395 Wellington Street OttawaONK1A0N4 Canada Bibliotheque et Archives Canada Direction du Patrimoine de I'edition 395, rue Wellington Ottawa ON K1A 0N4 Canada Your file Votre reference ISBN: 978-0-494-57557-4 Our file Notre re'fe'rence ISBN: 978-0-494-57557-4 NOTICE: The author has granted a nonexclusive license allowing Library and Archives Canada to reproduce, publish, archive, preserve, conserve, communicate to the public by telecommunication or on the Internet, loan, distribute and sell theses worldwide, for commercial or noncommercial purposes, in microform, paper, electronic and/or any other formats. The author retains copyright ownership and moral rights in this thesis. Neither the thesis nor substantial extracts from it may be printed or otherwise reproduced without the author's permission. AVIS: L'auteur a accorde une licence non exclusive permettant a la Bibliotheque et Archives Canada de reproduire, publier, archiver, sauvegarder, conserver, transmettre au public par telecommunication ou par I'lnternet, prefer, distribuer et vendre des theses partout dans le monde, a des fins commerciales ou autres, sur support microforme, papier, electronique et/ou autres formats. L'auteur conserve la propriete du droit d'auteur et des droits moraux qui protege cette these. Ni la these ni des extraits substantiels de celle-ci ne doivent etre imprimes ou autrement reproduits sans son autorisation. In compliance with the Canadian Privacy Act some supporting forms may have been removed from this thesis. While these forms may be included in the document page count, their removal does not represent any loss of content from the thesis. Conformement a la loi canadienne sur la protection de la vie privee, quelques formulaires secondaires ont ete enleves de cette these. Bien que ces formulaires aient inclus dans la pagination, il n'y aura aucun contenu manquant. 1*1 Canada

HI ABSTRACT In this thesis we are concerned with the determination of the values of t for which the resulting matrix A + te is positive semidefinite. That is, we want to find the positive semidefinite interval for the matrix pencil A + te. We first present a new point of view for the case that A is positive definite to obtain the same results as in [Car88] and then use this point of view to determine the positive semidefinite interval for the case that A is negative definite. In both of these cases, the positive semidefinite interval is determined from the eigenvalues of the matrix A~ l E. We then show how to combine our results to obtain the positive semidefinite interval for the case that A is nonsingular but indefinite, and for the case when A is singular and R(E) C R(A). Examples and remarks on implementation are also provided.

IV ACKNOWLEDGEMENTS It is my great pleasure to take this opportunity to express my sincere appreciation to my co-supervisors Dr. Richard J. Caron and Dr. Tim Traynor, who have brought me to the field of Optimization and have taught me a great deal. I owe a great debt of gratitude for their patience, inspiration and friendship. Also I would like express my gratitude to Dr. Abdo Y. Alfakih and Dr. Guoqing Zhang for their guidance and for reading my thesis. The Department of Mathematics and Statistics and the Operational Research (OR) Group have provided an excellent environment for my study, and for that I am thankful. Finally, I thank our graduate secretary Ms. Dina Labelle for all the help that she gave me during my study in Windsor.

CONTENTS Abstract iii Acknowledgements iv 1. Introduction 1 2. Background 3 2.1. Preliminaries 3 2.2. Caron and Gould's results for A positive semidefinite and E of rank one or two 7 2.3. Valiaho's results for determining the inertia of the matrix pencil as a function of the parameter 9 2.4. Caron's results for A positive semidefinite 13 3. The Positive Semidefinite Interval when A is Nonsingular 16 3.1. A is positive definite 16 3.2. A is negative definite 18 3.3. A is nonsingular and indefinite 23 4. The Positive Semidefinite Interval when A is Singular 26 5. Implementation and Examples 29 6. Conclusion and Future Work 36 References 38 Vita Auctoris 39

1 1. INTRODUCTION Let A and E be real symmetric n x n matrices. The (linear) matrix pencil A + te is the function t*-*a + te (t M). We consider the problem of determining the set T such that for t e T, A + te is positive semidefinite. [Recall that a symmetric matrix is positive (negative) semidefnite if all of its eigenvalues are non-negative (non-positive), is positive (negative) definite if all of its eigenvalues are positive (negative), and is indefinite if it has both positive and negative eigenvalues.] Since the set of positive semidefinite matrices is a closed convex cone in the space ofnxn matrices, T is a closed (possibly unbounded) interval. We call T the positive semidefinite interval. One source of interest in this problem is its connection to mathematical programming. Consider first the parametric quadratic program problem ([BC86], [Rit67], [Val85]) mm{(c(t)) T x + -x T C{t)x\(ai(t)) T x < bi(t),i = 1,, m}. If C{t) = A + te, then the quadratic programming is convex if and only if A + te is positive semidefinite. The positive semidefinite interval gives the values of t for which the critical points of the quadratic programming problem are guaranteed to be global minimizers. Consider also the area of semidefinite programming ([WSVOO], [VB96]) where the contraints can be written as the system of

linear matrix inequalities n F*(x) := F (j) 0 + ^ x ^ >: 0, j = 1,,q i=i where FQ and iy 7 are symmetric matrices, and where F^(x) >z 0 means that F^(x) is positive semidefinite. An application of the Coordinate Directions (CD) hit-and-run algorithm (introduced by Telgen [Tel79] and published in [Bon83]) for finding necessary constraints in the system of LMI requires the determination of T. In this thesis we are concerned with the computation of the endpoints of T. Chapter 2 gives some background material, including Caron and Gould's results for A positive semidefinite and E of rank one or two ([CG86]), Valiaho's results for determining the inertia (number of positive, negative,and zero eigenvalues) of the matrix pencil as a function oft ([Val88]) and Caron's results for A positive semidefinite ([Car88]). Chapter 3 discusses the case when A is nonsingular. Explicit expressions of the endpoints of the positive semidefinite interval are given when A is positive definite and when A is negative definite. In these cases, the endpoints are determined from the eigenvalues of A~ l E. Then we can combine these results to obtain the positive semidefinite interval when A is nonsingular but indefinite. Chapter 4 discusses the case when A is singular. Chapter 5 presents remarks on implementation and includes examples. Concluding remarks are given in Chapter 6. 2

3 2. BACKGROUND 2.1. Preliminaries. Throughout this paper, we shall denote the range space of a matrix M by R(M), its null space by N(M), and its rank by r(m). For a real symmetric nxn matrix M, we say M is positive definite (and denote it by M >- 0) if for all x e R n, x ^ 0, x T Mx > 0. We say M is positive semidefinite (and denote it by M >z 0) if for all xew 1 x T Mx > 0. It is well known that M is positive definite if and only if all eigenvalues of M are positive and M is positive semidefinite if and only if all eigenvalues of M are nonnegative. Given a real symmetric nxn matrix M, the Schur decomposition ([GVL83], [HJ90]) of M provides an orthogonal nxn matrix Q such that Q^MQ = ( ^ [J ), (1) where D\ is diagonal and nonsingular, and hence r(d\) = r{m). The inertia of a real symmetric matrix M is an ordered triple consisting of the numbers of positive, negative and zero eigenvalues of the matrix, i.e. the inertia of M is the triple In(M) = (<K(M),V(M),8(M)), where TT(M),V(M),5(M) denote the numbers of positive, negative and zero eigenvalues, respectively, of M. Note that ([Hoh73])

4 if L is nonsingular, then In(LML T ) = In(M). The Schur complement ([Cot74]) M of the nonzero element rrihk in a matrix M with row index set 7 and column index set J, is M = S hk M = [ray -], where i ^ h, j ^ k. Note that M is an (n 1) x (n 1) matrix with rows and columns indexed by 7 \ {/i} and J \ {A;}, respectively. Here the operator Shk is called the pivotal condensation with the pivot rrihk- Note that S kh {S hk M) = Shk{SkhM) provided that both sides are defined. By means of pivotal condensations it is possible to determine the inertia of any real symmetric matrix. If M is of rank r, then, independent of the order of the pivots, it is possible to perform on M exactly r successive pivotal condensations ([ZA66]). Thus, we can compute In(M) by the following algorithm. Algorithm 2.1. ([Cot74]) Computing the inertia of a real symmetric n x n matrix M. Step 0. Set C = M, p = q = 0. Step 1. If C = 0 (possibly vacuous), go to Step 2. Otherwise, (i) Select, if possible, a Chh " 0- Set C < ShhC, and set p +- p + 1, if c hh > 0; q <- q + 1, if c hh < 0. (ii) If every diagonal entry is zero, select a Chk ^ 0. Set C <-- SkhShkC, and set p <- p + 1, q <- q + 1. (hi) Go to Step 1. Step 2. In(M) = (p,q,n- p - q).

5 Throughout the above algorithm, we have In(M) = (p,g,0) + In(C). Given two real symmetric n x n matrices A and E, because of continuity of the eigenvalues (Lemma 2.3), if there exists a t for which A + te is nonsingular, the solutions of the equation det(a + te) = 0 divide the real line into open intervals in which the matrix pencil has constant inertia. We refer to these as intervals of constant inertia. The inertia for each interval is the inertia of A+tE where i is in the interval. When A is nonsingular, if det(a + te) = 0 then det{--i-a~ 1 E) = 0; so \ is an eigenvalue of A~ l E. On the other hand, when E is nonsingular, the solutions of the equation det(^l + te) = 0 are eigenvalues of AE~ l. This gives us two points of view for the determination of T. The continuity of eigenvalues of A + te is crucial in the following chapters. We shall state the more general Wielandt-Hoffman Theorem ([Wil65], page 104), from which one can deduce the continuity of eigenvalues of A + te. Theorem 2.2. If C = A-\- B, where A, J5, C are symmetric matrices having eigenvalues ai,pi,ji, respectively, arranged in nonincreasing order, then n n

For each t, let Aj( ) (1 < i < n) be the eigenvalues of A + te arranged in non-increasing order. Lemma 2.3. For each i = 1,, n, the function \ is continuous. Proof. Let \f (1 < i < n) be the eigenvalues of E arranged in non-increasing order. By Theorem 2.2, we have n (A^)-A^0)) 2 <(t-to) 2 (Af) 2 i=\ %=\ Hence for every z, A^ is continuous. The following Theorem and Corollary give a necessary and sufficient condition to guarantee that a block symmetric matrix is positive semidefinite. Theorem 2.4. f[hay68]j Suppose that A is nonsingular and M A B *, ( \ =U T c)' n 6 Then In(M) = In(A) + In(C - B T A~ l B). Corollary 2.5. Suppose that A >- 0. Then if and only if C-B T A~ l ByO. We are now ready to present some of the known results.

2.2. Caron and Gould's results for A positive semidefinite and E of rank one or two. If A is positive semidefinite and E is of rank one or two, then explicit expressions for the endpoints of T, and a numerically stable method for computing them, are given in [CG86]. We will summarize these results. A symmetric matrix of rank one or two can be written as ±(uu T +-\vv T ) where u and v are linearly independent n-vectors and A = 1, 0 or 1. However, if A + te is positive semidefinite for t [a, b], then A + t{ E) is positive semidefinite for t [ 6, a]. Therefore, we can assume without loss of generality that E = uu T + Xvv T. Let the positive semidefinite interval for A+tE be T = [t, t] 1. The expressions for t and t are derived for different cases according to the choice of E (A = 1, 0 or 1) and the relationship between A, u, and v. (1) A = 0 or 1. In this case, (a) if u G R(A) and v R(A), then 2 t = = = ^ = u T x v T y y/(u T x v T y) 2 + 4(u T y) 2 t +oo, where z,«/gr n are such that Ax = u and Ay = v. (b) if u i R{A) or v < R(A), then 7 t = 0, = +oo. Here, and in the sequel, we will use [a, 6] for {t K : a < t < b}. For example, if a = oo and b R, then [a, b] = ( 00,6].

(2) A = 1. In this case, (a) if u G R(A) and v G R(A), then 2 t = u T x + v T y y/(u T x + v T y) 2 4(v T x) 2 2 t =, u T x -f v T y + y{u T x + v T y) 2 4(i> T x) 2 where a;,j/6l" are such that Ax = it and Ay = v. (b) if u G R{A) and v R(A), then t =, t = 0, u' x where x G R n is such that Ax = u. (c) if w ^ #(A) and v G i?(a), then n - 1 = 0, = v T y'' where y G R n is such that Ay = v. (d) if w ^ i?(a) and v R(A), but v G R (A u), where (A it) denotes the matrix A augmented by u, let x G R n and a G R satisfy Ax + au = v. Then, 1 - a 2 = 0 => t = t == 0; 1 - Q 2 > 0 => t = 0 and 1-a 2 t = (v T au T )x : 1 a 2 < 0 => 1 -a 2 t = -r-== = (i>' au 1 )x and t =0.

9 (e) if u R(A), v R{A) and v R (A u), then t = t = 0. 2.3. Valiaho's results for determining the inertia of the matrix pencil as a function of the parameter. Valiaho ([Val88]) extended the results in [CG86] by presenting a method to determine the inertia of A + te as a function of the parameter t. He first observed that if E is nonsingular, the inertia change points are the values of t equal to the eigenvalues of AE^1. Then, using Algorithm 2.1, he evaluated In (.A + te), for one value t interior to each interval of constant inertia. The positive semidefmite interval is the interval with inertia (TT(M),0,6(M)). If E is singular, the problem is reduced to a nonsingular case of lower dimension. As a preliminary step, by using the Lagrange's reduction ([Hoh73]), we put E into the form where L is nonsingular and D is nonsingular and diagonal. D is of order r(e) and Rewrite L as L = (P Q) so that In(D) = (ir(e)me),0). (2) Note that P is an n x r(e) matrix of full rank. Also, we have ([MS75]) R(A E) = R{A) + R(E) = R(A) + R(P) = R(A P).

10 So rank (A E) = rank (A P), which will be denoted by r. The basic idea is to compute the inertia of the matrix (of order n + r{e)) P(t) =[pt _ r l )-l J Theorem 2.4 gives us that In(C( )) = In(-r 1 D- 1 )+lji{a + P(tD)P T ) = In(-tD) + ln{a + te). From (2), we have U[ [)) ) = Uv(E),n(E),0) + ln(a + te), if t > 0; \{ir(e),v(e),0) + In{A + te), if t < 0. Thus, hi(a + te) = l MC{t)) ~ {v{e) '* {E) ' 0) ' { } lf t>0] \ln{c{t))-(ir(e),v{e),0), if t < 0, So, in order to compute In(A + t^), it is sufficient to compute In(C(t)). We perform Algorithm 2.1 (Schur complement) on C(t) in two phases to obtain In(C(t)). Phase 1: We perform Algorithm 2.1 on C(t), using elements (h,k), h,k < n, as pivots as long as possible. The number of possible operations is r(a). At the end of this phase, we have P = TT(A), q = v(a) :

11 and C of the form Cl(t)= f^pt p 2 _ t -l D -lj> where the zero block is of order n r(a). Phase 2: We continue performing Algorithm 2.1 on Ci(i), applying double operations SkhShk with h < n, k > n, as long as possible. The number of possible double operations is r(pi). So at the end of this phase, we have and C of the form p = ir{a) + r(p 1 ), q = v{a) + r(p 1 ), CM = { 0 E.+t-'A, ) = * _1 ( 0 teh + Ai)' where Ai and E\ are of order r{e) r{p\) and the zero block on the main diagonal is of order n r(a) r{p\). Since the total number of operations in Phase 1 and Phase 2 is r(a) + r(p\), we have r(a) + r(pi) = rank (A P) = rank (A E) = r. So, Therefore, r(p l ) = r-r(a). p = 7r(A) + r(pi) = r - v(a), q = v(a) + r(pi) = r - 7r(A). Moreover, Ai and E\ are of order r(a) + r(e) r and the zero block on the main diagonal of C*2(t) is of order n r.

12 Now we have In(C(*)) = In(C2(*)) + (p,<z,0) Therefore, when t > 0, fln^i + ^i)+ (p,g,n-r), if t > 0; ln(-(ai + ^i)) + (p,g,n-r), if i < 0. In(A + * ) = In(Ai + te x ) + (p - t>( ), g - ir(e), when t < 0, = In(Ai + ^i) n-r) + (r - ^(A) - v(e),r - ir(a) - ir{e),n - r); In(4 + ) = In(-(Ai + tei)) + (p - TT( ), g - u( ), n - r) = In(-(A 1 +^1)) + (r - I>(J4) - 7r(E),r - ir(a) - v(e),n - r). Then we only need to compute In(^4i + te\). Note that the matrix pencil A\ + te\ is of order r(a) + r(e) r. When E is singular, r(a) + r( ) r<n. Thus, we have reduced the dimension of the problem. If E\ is singular, we repeat the process. While Valiaho's method can indeed determine the positive semidefinite interval, in all cases, it is a consequence of having determined all intervals of constant inertia and the inertia of the matrix pencil for some t within each interval. However, if we are only concerned with finding when the matrix pencil is positive semidefinite, Valiaho's algorithm is more than what is necessary, because we are not interested in the matrix

inertia. Note that this is the case in the analysis of linear matrix inequality constraints sets in semidefmite programming problems. In the next section we show how the results in [CG86] can be extended more directly, focussing on the positive semidefintie interval alone. 13 2.4. Caron's results for A positive semidefmite. In the technical report [Car88], Caron extended the results in [CG86] to general E, with A positive semidefmite. In case R{E) C R(A), the positive semidefmite interval is obtained from the eigenvalues of the n x n matrix X satisfying AX = E and R(X) C R(A). In particular, if A is positive definite, then it follows that R(E) C R(A) and X = A~ l E. We summarize this result in the following theorem. Theorem 2.6. Suppose that R(E) C R(A). Let X be an n x n matrix satisfying AX = E and R(X) C R(A). Let 5 n be the largest eigenvalue of X and 5\, the smallest. Then A + te is positive semidefinite if and only if' t G [t, t\ where t = t = oo, +oo, if5 n >0 otherwise if5!<0 otherwise. [Car88] also discussed the case R(E) < R(A). The results, however, are incorrect and we will present a counterexample. To state the claim, we need some notation.

The Schur decomposition of E provides an orthogonal matrix Q such that /A o o \ Q T EQ = 0 0 0 V 0 0 -D 3 J where D\ and D3 are diagonal positive definite matrices. Define the n x n matrices U and V by /Di 0 0\ / 0 0 0 \ U= \ 0 00 and y = 0 0 0. \ 0 0 0 / \ 0 0 D 3 / Let B = Q AQ. Then A +? is positive semidefinite if and only if B + t(f/ - V) is positive semidefinite. Since R(E) % R{A), then R(U - V) R{B) and therefore, either R(U) % R{B) or R(V) % R(B). Now we can state the claim as in Theorem 4.1 of Car88]: If R{U) R{B) and R(V) C #( ), then T = [0, t v ] where fo is as given in Theorem 2.6, with A replaced by B and E replaced by -V. Example 2.7. Let 14 B U V

Then B is positive semidefinite, R(U) R(B), and R(V) C R(B). We have / 1 0 0 B + ([/ - V) = 0 4 1 \ 0 1 1/4 and /0 0 0 \ B-V = 0 2 1. \0 1 1/4/ So B + (t/ y) is positive semidefinite, but B V is not. Let the positive semidefinite intervals for B+{U V) and B V be T and T, respectively. According to Theorem 4.1 in [Car88], T = T. However, in this example, we have 1 G T but 1 ^ T. For the case R(U) C #( ) and i?(f) g i2(b), we can get a similar counterexample. In the next chapter we present new results. 15

16 3. THE POSITIVE SEMIDEFINITE INTERVAL WHEN A is NONSINGULAR In this chapter we assume that A is nonsingular. In this case, det(a-\-te) is a polynomial in t and is not identically zero. Therefore, the equation det(a + te) = 0 (3) has finitely many solutions. We present the results in three sections, according to whether A is positive definite, negative definite, and indefinite. 3.1. A is positive definite. This case has already been solved in [Car88] and summarized in our Theorem 2.6. We shall present a new point of view that will help us present new results. When A is positive definite, the right endpoint i of T is strictly positive. Thus, if i ^ +oo, then A + ie >z 0 and it follows that all eigenvalues of A + ie are nonnegative. Moreover, one of them must be zero. Otherwise, the continuity of the eigenvalues would imply the existance of a 5 > 0 with A + (i + S)E >~ 0, contradicting the fact that t is the right endpoint. It then follows that det(^4 -f ie) = 0. So, i must be a positive solution for (3). In other words, if (3) has no positive solutions, then i = +oo. Similarly, if (3) has no negative solutions, then t = oo. Lemma 3.1. If A is positive definite and equation (3) has positive solutions, then the smallest one is the right endpoint i of the positive semidefinite interval T for A + te. Otherwise, i = +oo.

Proof. We have already shown that if (3) has no positive solutions then i = +00. Let t m i n be the smallest positive solution to (3). We will will first show that A + t m i n E >z 0. Indeed, if A + t min E is not positive semidefinite, then it must have a negative eigenvalue. Let Xi(i) (1 < i < n) be eigenvalues of A + te arranged in non-increasing order. Without loss of generality, assume that \\(t m i n ) < 0. Since A >- 0, Ai(0) > 0. By continuity of eigenvalues, there must be some to G (0, t m i n ) such that Ai( 0 ) = 0. So det(a + t 0 E) = 0. This is a contradiction, because t m in is the smallest positive solution for (3). Therefore, A -I- t m % n E >_ 0. We now show that A + te >- 0, for t G (0,t m ; n ). Indeed, for 17 A + te = t{-a + E) = t[( 7 ] -A + E) + ( 1 --^-)A} = t[- {A + t min E) + (- - )A] >- 0. train ^ tmin It now remains to show that A-\-tE is not positive semidefinite, for t > t m i n. Since det(a + t m i n E) = 0 there exists y G K n, y ^ 0, such that y T (y4 + min i?)2/ = 0. It then follows that y T Ey < 0. So for t > t min, y T {A + te)y = y T (A + t min E)y + (t-t min )y T Ey < 0. Therefore, A + te is not positive semidefinite, for t > t m in-

The next lemma presents the analogous result when equation (3) has negative solutions. Lemma 3.2. If A is positive definite and equation (3) has negative solutions, then the largest one is the left endpoint t of the positive semidefinite interval T for A + te. Otherwise, = oo. Recall that if det(a + te) = 0, then i / 0 and -1/t is an eigenvalue for A~*E. The following theorem follows from Lemma 3.1 and Lemma 3.2. It is the special case of Theorem 2.6 when A is positive definite, since X = A~ l E. Theorem 3.3. Suppose that A is positive definite. Let S n (5\) be the largest (smallest) eigenvalue of A" l E. The positive semidefinite interval for A + te is T = [t, t\ where 18 t = -oo, if5 n >0 otherwise t = +oo, if6i<0 otherwise. 3.2. A is negative definite. We first show that if E is singular or indefinite the interval is empty. Lemma 3.4. Suppose that A is negative definite. If det(e) = 0 or E is indefinite, then T = 0. Proof If det( ') = 0, then there exists y G R n, y ^ 0, such that y T Ey = 0.

19 It then follows that for any t, y T (A + te)y<0. Therefore, when det(e') = 0, A + te is not positive semidefinite for any t. If E is indefinite, then there exist u, v G R n, such that u T Eu > 0, v T Ev < 0. It then follows that for t < 0, u T {A + te)u < 0; and for > 0, v T (A + te)v <0. Since A + 0 = A is negative definite, T = 0. The next result shows that when A is negative definite and E is positive definite, then A + te is positive definite for sufficiently large t. Lemma 3.5. Suppose that A is negative definite and E is positive definite. If max x T ( A)x then A + te y0. x llx =l t > min x 1 Ex 1x11=1 Here, the numerator is the largest eigenvalue of A and the denominator is the smallest eigenvalue of E. It is easy to see this by using the diagonalization of a symmetric matrix.

Proof. For any x G R n, \\x\\ = 1, E y 0 implies :r T i?:e > 0. Then for max x T ( A)x t > \\x\\=i min x T Ex we have \\x\\=l z T (j4 + E)a; = -x T (-A)x + tx T Ex Therefore, A + teyq. max :c T ( A)x II M -I ^ ' > -:T T (-,4):E + = x T^x mm a:' Ex > T ( A)x + max x T ( A)x ll*ll=i > 0. We can now determine the interval T. Lemma 3.6. Suppose that A is negative definite and E is positive definite. Let t max be the largest positive solution for equation (3). Then the positive semidefinite interval for A-\-tE is 20 Proof We first show that A -f t max E y 0. Indeed, if A + t max E is not positive semidefinite, then it must have a negative eigenvalue. By Lemma 3.5, A + te is positive definite for sufficiently large t. So, there exists some t\ G (t max, +oo) such that det(yl + t\e) = 0. This is a contradiction, because t max is the largest positive solution for equation (3). Therefore, A + t max E y 0.

We.now show that A + te >- 0, for t (t max, +00). For t (*moa;,+oo), we have 21 = ([(J-A + E) + (7 - T^-M] = t[- (4 + t ma^) + (- - )4] >- 0. tmax t ^max Finally, we show that A + te is not positive semidefinite, for t < tmax- Since det(a + t max E) = 0, there exists?/ G K n, 2/ 7^ 0, such that Note that E y 0 implies So, for < t max, we have 2/ T #2/ > 0. T/ T (A + te)y = y T {A + t ma^) / + (t - t max )y T Ey < 0, and so A + te is not positive semidefinite for t < t max. The next lemma gives the corresponding result for the case that E is negative definite. Lemma 3.7. Suppose that A is negative definite and E is negative definite. Let t m i n be the smallest negative solution for equation (3). Then the positive semidefinite interval for A-\-tE is D OO, Vmin,

The following theorem gives T for the case of A negative definite. Theorem 3.8. Suppose that A is negative definite. Let S n (5\) be the largest (smallest) eigenvalue of A~ l E. Then the positive semidefinite interval T for A + te is given as follows. (1) T = 0 if A~ l E has a zero eigenvalue. (2)T = if5 1 <0<5 n. (3)T=[- } +oo) if5 n <0.. (4) T =(- ),- ].*/<*! >0. Proof. (1) Since A~*E has a zero eigenvalue, we have det(a -1 ') = 0. Hence det(e) = 0. By Lemma 3.4, T = 0. (2) Let u,v G M n be eigenvectors corresponding to 5\, S n. So A~ l Eu = 6iu, 22 Then, A~ l Ev = 6 n v. u T Eu = 5iu T Au > 0, v T 7v = 5 n v T Av < 0. Therefore, E is indefinite. By Lemma 3.4, T 0. (3) Since n < 0, all eigenvalues of A~ X E are negative. Hence the equation det(>l + t ) = 0 has no negative solutions. It then follows from A is negative definite, that for all < 0, A + te -< 0. That is, for any

23 x er n,x^o, any t < 0, and any t < 0, x T (A + te)x < 0. Thus, for / y ' A Of* T X -TlX a; Ea; > - Letting > oo, we have for any a; G K n, x ^ 0, T :r > 0. That is, ^ 0. On the other hand, 8 n < 0 implies A~ l E has no zero eigenvalue and hence det(j4-1.e) ^ 0. Thus det(i?) ^ 0. Therefore, E y 0. By Lemma 3.6, T=[-i +oo). (4) The proof is similar to (3) and follows from Lemma 3.7. 3.3. A is nonsingular and indefinite. We now combine the results in Section 3.1 and Section 3.2 to obtain the positive semidefinite interval when A is nonsingular but indefinite. The Schur decomposition of A provides an orthogonal n x n matrix Q such that Q T W = (? \ ) (4) where D l5 D2 are diagonal and positive definite. Let «TB «=(5S)-

24 Then Q^A + te )Q =(* t + E P _ D [ E l te2 ). By Theorem 3.3 and Theorem 3.8, we can determine the positive semidefinite intervals for D\-\rtEi and D2 + te2- We denote them by T\ and T 2, respectively. Let 81,,8k be the nonzero eigenvalues of A~ l E and U = j-, i = 1,, k. Define Clearly, T 0 CTCT^Tj. To = {ti\l <i<k, A + UEh 0}. Theorem 3.9. When A is nonsingular and indefinite, the positive semidefinite interval T for A + te is given as follows. (1) Suppose that T 0 = 0. Then T = 0. (2) Suppose that T$ is a singleton {to}. (a) J/Ti n T 2 = {to}, then T = {to}. (b) IfTi n T 2 = [a, b], then T = {t 0 }. (c) IfT 1 C\T 2 = [a, +00) and/or some t* > t 0, A + t* b 0, t/ien T = [to, +00). (d) If T\ n T 2 = [a, +00) and /or some t* > t 0, A + *# is not positive semidefinite, then T = {to}. (e) IfTiC\T 2 = (-00,6] and for some t* < t 0, A + t*e± 0, t/ien T = ( 00, to]. (f) // Ti n T 2 = (-00, b] and for some t* <t 0, A + t*e is not positive semidefinite, then T {to}- (3) Suppose that To has more than one point, say tj 15,U r, where r > 1 and t^ < < U r. (a) // T\ n T 2 = [a, b], then T = [t h, t ir ]. (b) IfT 1 C\T 2 = [a, +00) and for some t* > t ir, A + t*e >z 0, t/ien T = [tj 15 +00).

(c) // T\ n T 2 = [a, +00) and for some t* > t ir, A + t*e is not positive semidefinite, then T = [t^,^]. (d) IfT 1 HT 2 = (-00,6] and for some t* < t h, A+t*E >z 0, then T = ( oo,t ir \. (e) IfTi n T 2 = (-00, b] and for some t* < t h, A + t*e is not positive semidefinite, then T = [t^, $ r ]. Proof. The above results are straightforward. For example, we will prove 3(b). Let T = [t,t\. In this case, if t < +00, then det(a + te) = 0 and,4 + ts >: 0. Thus t G T 0. Since A + t*e b 0 and t is the right endpoint of the positive semidefinite interval, we have t > t*. It then follows from t* > U r that t > U r. This is a contradiction, because U r is the largest element in To. Therefore, t = +00. On the other hand, since T C T\ fl T2, we have t > 00. So det(>l + ) = 0 and A + E h 0. Thus fet 0 and hence t > t h. Since A + t^e y 0 and t is the left endpoint of the positive semidefinite interval, we have that t < t^. So t t^. Therefore, T = [t^, +00). 25 D

26 4. THE POSITIVE SEMIDEFINITE INTERVAL WHEN A is SINGULAR In this chapter, we are concerned with the positive semidefinite interval for the matrix pencil A + te when A is singular. The Schur decomposition of A provides an orthogonal n x n matrix Q such that where D\ is diagonal and nonsingular. Q T AQ=( D J ), (5) Let o T^=( g). Then where i has the dimensions of D\. Since D\ is nonsingular, the positive semidefinite interval for the matrix pencil D\ + te\, which we denote by T\, can be found using Theorem 3.3, Theorem 3.8, or Theorem 3.9. Let T% denote the positive semidefinite interval for the matrix pencil tei- Then, (1) E 2 = 0 =* T 2 = R. (2) 2 >: 0, 2 ^ 0 =» T 2 = [0, +00). (3) E 2^ 0,E 2^ 0^T 2 = (-co,0]. (4) 2 is indefinite => T 2 = {0}.

Note that if A + te is positive semidefinite, then both D\ + te\ and te 2 are positive semidefinite. However, in general, T\ H T 2 is not the positive semidefinite interval for A + te. Example 4.1. Let A = I ) and E = ( ). Then the positive semidefinite interval for A + te is T = {0}. But Ti = T 2 = ( oo,+oo). In the following, let's consider the case R(E) C i?(a), which is equivalent to N(A) C JV( ). Given the Schur decomposition of A in (5), for any nxn matrix W, let 27 Lemma 4.2. i?(w) C i?(^) z/ and only ifw 2 = 0 and W 3 = 0. Proof. Note that i?(w) C R(A) if and only if for some X, i.e. Let AX = W, Q T AQQ T XQ = Q T WQ. «T * «- ( * * ) - Then #(W) C fl(a) if and only if /A 0 \ / X, X 4 \ = [W 1 W A \ \ o oj\x 3 x 2 J- \w 3 w 2 J-' l o o r \w 3 w 2 r

28 Therefore, R(W) C R(A) if and only if W 2 = 0 and W 3 = 0. D In Lemma 4.2, if W is symmetric, then W4 = W3" = 0. By Lemma 4.2, we have for the case R{E) C i?(a), that Hence, 0^ + tw=(^v Bl o)- Since A + te hoif and only if Di+t^bO, we have Theorem 4.3. Let the positive semidefinite interval for D\+tE\ be T\. Then the positive semidefinite interval for A+tE ist = T\. Since D\ is nonsingular, we can use Theorem 3.3, Theorem 3.8, or Theorem 3.9 to determine T\.

29 5. IMPLEMENTATION AND EXAMPLES When A is positive semidefinite, it is necessary to determine whether or not R(E) C R(A). This can be done by checking the consistency of AX = E. One approach is to first determine a decomposition of A using Choleski factorization with symmetric pivoting ([DMBS79]). If r(a) = r, there exists a nonunique permutation matrix P and a triangular matrix R (unique for a given P) such that P T AP = R T R, where D / Rn R12 \ R ={ 0 0 J' and where Rn is a nonsingular upper triangular r x r matrix and #12 is an r x (n - r) matrix. Let X P = P T XP and E P = P T EP. Now, AX = E is equivalent to R T RX P = E P. Set Y = RX P. We first solve R T Y = E P for the matrix Y. This is done as follows. Set 7 YX y 4 A Y = *3 Y2 and E P = ^PI Ep 4 E P 2, E P 2 Then we can rewrite R T Y E P as Since R\\ is nonsingular, Y\ and I4 are uniquely determined by Rj^ = E P1 and R n Y4 = E P 4.

30 But Y3 and Y 2 are undetermined. Furthermore, the equation AX = E has a solution only if Y\ and Y4 satisfy Rj 2 Yi = E P3 and Rj 2 Y 4 = E P2. Next we solve Y = RXp. Let X ( Xpi Xp4 \ \ Xp 3 Xp 2 J Then Y RXp can be rewritten as ( Rn R12 \ ( X P1 X P4 \ f Fi Y A \ \ 0 0 J \ X P3 X P2 J \Y 3 Y 2 J- This implies that Y3 = 0 and Y 2 = 0. Also, Xp% and Xp 2 are arbitrary; and Xp\ and Xp± are the unique solutions to RnXpi = Y\ RnXpz RiiXp^ = Y4 R\ 2 Xp2, respectively. We choose Xp$ = 0 and Xp 2 0. The resulting matrix is X = P ( Xpi XpA ^] P T \ Xps Xp 2 J We can use Householder reduction and the QR method ([GVL83]) to compute eigenvalues of the matrix X. The Householder matrix is of the form H = I - ^uu T, IMI where / is the identity matrix of order n and u G R n. Note that H is symmetric and orthogonal.

Suppose that r(x) = r. Then there exist Householder matrices H\,, H r such that where R is upper triangular. So where Q is orthogonal. H r H\X = R, X = QR, Let X\ = RQ. By Householder reduction, X\ = QiRi, where Q\ is orthogonal and R\ is upper triangular. Let X2 = R\Q\- By Householder reduction, X2 = Q2R2, where Q2 is orthogonal and R2 is upper triangular. In general, suppose that we have Xk = QkRk, where Qk is orthogonal and Rk is upper triangular. Let Xk+i = RkQk- By Householder reduction, Xk+i = Qk+iRk+i, where Qk+i is orthogonal and Rk+i is upper triangular. Since Xk+l = RkQk = all Xk have the same eigenvalues as X. Qk^kQk, Also, as &» +00, Xk tends to an upper triangular matrix. So, for sufficiently large k, the diagonal elements of Xk are approximate values of the eigenvalues of X. 31

Example 5.1. Let A = and 1 0 0 E= I 0-1 0 0 0 0 The Choleski decomposition of A is with and P T AP=R T R 2 0 R=( R Q 1 R * 2 ) = I o 0 3 0 0 p = We see that and 7? - I 2 Rn ~ [ 0 3 J*12 = ( 0 Then E P = T P l EP First we solve the equation R T Y = E P

which is to get and Since both and Y! = 1/2 0 0-1/3 MS Rj 2 Y x = 0 Rj 2 Y A = 0, we have #( ) C i?(a). Set Y 3 = 0 and F 2 = 0. i.e., Next we solve the equation Y = RX P Setting Xp3 and Xp2 to zero, we get and Api - I 0-1/9 Xp4 ' 0 0

34 The resulting matrix X is 1/4 0 0 X = PX P P T = 0-1/9 0 0 0 0 Finally, the eigenvalues of X are 1/4, 1/9, and 0. Therefore, the positive semidefinite interval is [ 4, 9]. In case A is negative semidefinite, A is positive semidefinite and the above procedure applies. The next example shows how Theorem 3.9 can be used in the case when A is indefinite. Example 5.2. Let The matrix pencils and *+**=(;!) + '(S- 0 -D 2 + te 2 = {-l) + t{2) have positive semidefinite intervals T\ = [ 1,1] and T 2 [1/2, +oo), yielding T x nt 2 = [1/2,1].

The eigenvalues of A l E are: So do = 2 2 2 1 ^5 (5, = -- 2 2 2 2 2 2 1 1 v^ Then T 0 = {^ = 1,2,3 and A + UE±0} = {1, Therefore, the positive semidefinite interval for

36 6. CONCLUSION AND FUTURE WORK In this thesis, we present a summary of overview of the results on the positive semideflnite interval of the matrix pencil A + te. Also, for the case that A is positive definite, we present a new point of view (Lemma 3.1 and Lemma 3.2) to get the same result as in [Car88]. For the case that A is negative definite, Theorem 3.8 shows how to determine the positive semideflnite interval from the eigenvalues of the matrix A~ l E. For the case that A is nonsingular but indefinite, we can use Theorem 3.9 to determine the positive semidefinite interval. For the case that A is singular and R(E) C R(A), the Schur decomposition of A provides an orthogonal matrix Q such that Q T {A + te)q f l n l J, where D\ is diagonal and nonsingular. Then the positive semideflnite interval for A + te is exactly the positive semideflnite interval for D\ + te\. We summarize the results with the following table. A PSD PSD PD ND indef. nonsing. singular E rank 1 or 2 R(E) C R(A) arbitrary arbitrary arbitrary R(E) C R(A) PSD interval for A + te [CG86] [Car88] Th3.3 Th3.8 Th3.9 Th 4.3 Answers to the following questions require further research.

(1) How do we determine the positive semidefmite interval when A is singular and R(E) <Z R{A)1 (2) How do we solve the equation det(a + te) 0 when both A and E are singular? In this case det(^4 + te) may be identically zero. 37

38 REFERENCES [BC86] M. J. Best and R. J. Caron, A parameterized Hessian quadratic programming problem, Ann. Oper. Res. 5 (1986), no. 1-4, 373-394. [Bon83] Arnon Boneh, PREDUCE - a probabilistic algorithm for identifying redundancy by a random feasible point generator (RFPG), Redundancy in Mathematical Programming (Mark H. Karwan, Vahid Lotfi, Jan Telgen, and Stanley Zionts, eds.), Springer-Verlag, Berlin, 1983, pp. 108-134. [Car88] R. J. Caron, Positive semidefinite parametric matrices, Windsor Math Report 950 (1988). [CG86] R. J. Caron and N. I. M. Gould, Finding a positive semidefinite interval for a parametric matrix, Linear Algebra Appl. 76 (1986), 19-29. [Cot74] Richard W. Cottle, Manifestations of the Schur complement, Linear Algebra and Appl. 8 (1974), 189-211. [DMBS79] J. J. Dongarra, C. B. Moler, J. R. Bunch, and G. W. Stewart, UNPACK users' guide, Society for Industrial and Applied Mathematics, Philadelphia, 1979. [GVL83] Gene H. Golub and Charles F. Van Loan, Matrix computations, Johns Hopkins Series in the Mathematical Sciences, vol. 3, Johns Hopkins University Press, Baltimore, MD, 1983. [Hay68] Emilie V. Haynsworth, Determination of the inertia of a partitioned Hermitian matrix, Linear Algebra and Appl. 1 (1968), no. 1, 73-81. [HJ90] Roger A. Horn and Charles R. Johnson, Matrix analysis, Cambridge University Press, Cambridge, 1990, Corrected reprint of the 1985 original. [Hoh73j Franz E. Hohn, Elementary matrix algebra, third ed., The Macmillan Co., New York, 1973. MR MR0349697 (50 #2190) [MS75] George Marsaglia and George P. H. Styan, Equalities and inequalities for ranks of matrices, Linear and Multilinear Algebra 2 (1974/75), 269-292. [Rit67] Klaus Ritter, A method for solving nonlinear maximum-problems depending on parameters, Naval Res. Logist. Quart. 14 (1967), 147-162. [Tel79] J. Telgen, The CD algorithm, 1979, Private communication. [Val85] Hannu Valiaho, A unified approach to one-parametric general quadratic programming, Math. Programming 33 (1985), no. 3, 318-338. [Val88] H. Valiaho, Determining the inertia of matrix pencil as a function of the parameter, Linear Algebra Appl. 106 (1988), 245-258. [VB96] L. Vandenberghe and S. Boyd, Semidefinite programming, SIAM Review 38 (1996), 49-95. [Wil65] J. H. Wilkinson, The algebraic eigenvalue problem, Clarendon Press, Oxford, 1965. MR 32 #1894 [WSV00] Henry Wolkowicz, Romesh Saigal, and Lieven Vandenberghe (eds.), Handbook of semidefinite programming, International Series in Operations Research & Management Science, 27, Kluwer Academic Publishers, Boston, MA, 2000. [ZA66] S. I. Zukhovitskiy and L. I. Avdeyeva, Linear and convex programming, Translated from the Russian by Scripta Technica, Inc. Edited by Bernard R. Gelbaum, W. B. Saunders Co., Philadelphia, Pa., 1966.

39 VITA AUCTORIS Huiming Song was born in 1973 in China. He got his B.Sc. in Mathematics in Peking University, China, in 1994. He is currently a candidate for the Master's degree in Mathematics at the University of Windsor and hopes to finish his graduate study in December 2004.