ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION

Similar documents
Block companion matrices, discrete-time block diagonal stability and polynomial matrices

Iterative Solution of a Matrix Riccati Equation Arising in Stochastic Control

The norms can also be characterized in terms of Riccati inequalities.

Operators with numerical range in a closed halfplane

Elementary linear algebra

Singular Value Inequalities for Real and Imaginary Parts of Matrices

w T 1 w T 2. w T n 0 if i j 1 if i = j

The symmetric linear matrix equation

Linear Algebra: Matrix Eigenvalue Problems

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation

On the algebraic Riccati equation

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Discrete Riccati equations and block Toeplitz matrices

Chapter 2 Linear Transformations

arxiv: v1 [math.gr] 8 Nov 2008

Stability and Inertia Theorems for Generalized Lyapunov Equations

STABILITY OF INVARIANT SUBSPACES OF COMMUTING MATRICES We obtain some further results for pairs of commuting matrices. We show that a pair of commutin

SYMMETRIC SUBGROUP ACTIONS ON ISOTROPIC GRASSMANNIANS

Analysis Preliminary Exam Workshop: Hilbert Spaces

Lecture notes: Applied linear algebra Part 1. Version 2

YONGDO LIM. 1. Introduction

Perturbation Theory for Self-Adjoint Operators in Krein spaces

Characterization of half-radial matrices

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

Two Results About The Matrix Exponential

Definite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra

Representation Theory

Diagonal matrix solutions of a discrete-time Lyapunov inequality

NORMS ON SPACE OF MATRICES

On Solving Large Algebraic. Riccati Matrix Equations

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Chapter 3 Transformations

SYMPLECTIC MANIFOLDS, GEOMETRIC QUANTIZATION, AND UNITARY REPRESENTATIONS OF LIE GROUPS. 1. Introduction

DAVIS WIELANDT SHELLS OF NORMAL OPERATORS

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Minimum Polynomials of Linear Transformations

The Kalman-Yakubovich-Popov Lemma for Differential-Algebraic Equations with Applications

Parameterizing orbits in flag varieties

Radon Transforms and the Finite General Linear Groups

A proof of the Jordan normal form theorem

235 Final exam review questions

QUASI-UNIFORMLY POSITIVE OPERATORS IN KREIN SPACE. Denitizable operators in Krein spaces have spectral properties similar to those

ECE 275A Homework #3 Solutions

FACTORING A QUADRATIC OPERATOR AS A PRODUCT OF TWO POSITIVE CONTRACTIONS

Lemma 1.3. The element [X, X] is nonzero.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Structured Condition Numbers of Symmetric Algebraic Riccati Equations

INVESTIGATING THE NUMERICAL RANGE AND Q-NUMERICAL RANGE OF NON SQUARE MATRICES. Aikaterini Aretaki, John Maroulas

APPLIED LINEAR ALGEBRA

L(C G (x) 0 ) c g (x). Proof. Recall C G (x) = {g G xgx 1 = g} and c g (x) = {X g Ad xx = X}. In general, it is obvious that

Positive Solution to a Generalized Lyapunov Equation via a Coupled Fixed Point Theorem in a Metric Space Endowed With a Partial Order

PRODUCT OF OPERATORS AND NUMERICAL RANGE

MATH Linear Algebra

ON THE MATRIX EQUATION XA AX = X P

TROPICAL SCHEME THEORY

REAL RENORMINGS ON COMPLEX BANACH SPACES

Positive definite preserving linear transformations on symmetric matrix spaces

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

KOREKTURY wim11203.tex

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

The Singular Value Decomposition and Least Squares Problems

The Cayley-Hamilton Theorem and the Jordan Decomposition

Lecture 7: Positive Semidefinite Matrices

MATH36001 Generalized Inverses and the SVD 2015

The Hermitian R-symmetric Solutions of the Matrix Equation AXA = B

ON A PROBLEM OF ELEMENTARY DIFFERENTIAL GEOMETRY AND THE NUMBER OF ITS SOLUTIONS

Siegel Moduli Space of Principally Polarized Abelian Manifolds

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ

Abstract. In this article, several matrix norm inequalities are proved by making use of the Hiroshima 2003 result on majorization relations.

Chapter 1. Matrix Algebra

Where is matrix multiplication locally open?

COMMON COMPLEMENTS OF TWO SUBSPACES OF A HILBERT SPACE

A DECOMPOSITION THEOREM FOR FRAMES AND THE FEICHTINGER CONJECTURE

β : V V k, (x, y) x yφ

7. Dimension and Structure.

Lecture Summaries for Linear Algebra M51A

arxiv: v1 [math.ag] 10 Jan 2015

13. Forms and polar spaces

e jk :=< e k, e j >, e[t ] e = E 1 ( e [T ] e ) h E.

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

MAT 5330 Algebraic Geometry: Quiver Varieties

1 Linear Algebra Problems

Definitions for Quizzes

Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 4

A note on the σ-algebra of cylinder sets and all that

Linear algebra and applications to graphs Part 1

Algebra I Fall 2007

ALUTHGE ITERATION IN SEMISIMPLE LIE GROUP. 1. Introduction Given 0 < λ < 1, the λ-aluthge transform of X C n n [4]:

Normality of adjointable module maps

Foundations of Matrix Analysis

ELA

SYMPLECTIC GEOMETRY: LECTURE 5

Functional Analysis Exercise Class

Lipschitz stability of canonical Jordan bases of H-selfadjoint matrices under structure-preserving perturbations

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Chapter 4 Euclid Space

Transcription:

ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION Harald K. Wimmer 1 The set of all negative-semidefinite solutions of the CARE A X + XA + XBB X C C = 0 is homeomorphic to a well defined set of A-invariant subspaces provided that the purely imaginary eigenvalues of A are controllable. Based on that homeomorphism isolated n.s.d. solutions of the CARE are characterized by properties of their kernels. 1. INTRODUCTION In this paper we consider the continuous-time algebraic Riccati equation (CARE) R(X) = A X + XA + XBB X C C = 0, (1.1) where A, B and C are complex matrices of dimensions n n, n p and q n respectively. We focus on the set T = {X R(X) = 0, X 0} of negative-semidefinite solutions and we shall characterize those elements X of T which are isolated (in the topology which T inherits as a subset of the normed space C n n ). Such an isolated X has the property that for a sufficiently small ɛ there is no solution Y T, Y X, with X Y < ɛ. For λ σ(a) let E λ (A) = Ker(A λi) n denote the corresponding generalized eigenspace and let C CA V (A, C) = Ker. CA n 1 1 Research supported by Deutsche Forschungsgemeinschaft (Wi 1219/1-1).

be the unobservable subspace. It is the purpose of this paper to prove the following result. THEOREM 1.1. Assume that (1.1) has a solution X 0. Then T has an isolated element if and only if all purely imaginary eigenvalues of A are B-controllable. A solution X T is isolated in T if and only if its kernel has the following property: If Re λ > 0 and rank A λi n 2 C then either or Ker A λi Ker X = 0 C V (A, C) E λ (A) Ker X. Isolated solutions of quadratic matrix equations were studied for the first time by J. Daughtry [Da]. We note the corresponding result for the ARE A X + XA + XBB X Q = 0, (1.2) where Q = Q. THEOREM 1.2 [RR]. Assume that (A, B) is controllable. Then X is isolated in the set of hermitian solutions of (1.2) if and only if each common eigenvalue of A + BB X and (A + BB X) which is not purely imaginary is an eigenvalue of of geometric multiplicity one. H = A Q BB A Our study is based on results of [W1] which will be reviewed in Section 2. If A has no uncontrollable modes on the imaginary axis then according to [W1] there is a bijection between T and a well-defined set N of A-invariant subspaces. Some facts on the gap metric and on isolated invariant subspaces are put together in Section 3. In Section 4 we show that the bijection between T and N mentioned above is a homeomorphism. We give a proof of Theorem 1.1 in Section 5. The following notation will be used. In the partitions C = C C > = C < C = C > (1.3)

the subscripts refer to real parts such that C = {λ λ C, Re λ 0}, etc. To (1.3) correspond the decompositions C n = E (A) E > (A) = E < (A) E = (A) E > (A) where E (A) = {E λ (A), λ C }, etc. Let InvA denote the lattice of A-invariant subspaces of C n. To the triple (A, B, C) we associate the controllable subspace R(A, B) = Im(B, AB,..., A n 1 B) and the unobservable subspace V (A, C). It will be convenient to define and similarly V = (A, C), V (A, C), etc. R(A, B) + E < (A) is the stabilizable subspace. V (A, C) = V (A, C) E (A) With this notation V (A, C) is the undedectable and 2. DECOMPOSITION AND PARAMETRIZATION OF SOLUTIONS The subsequent theorems which describe the structure of the solution set T are taken from [W1]. According to [G], [GH] there exists a solution X 0 of (1.1) if and only if V (A, C) + R(A, B) + E < (A) = C n. (2.1) Because of V < (A, C) E < (A) the preceding condition (2.1) can be written as Put Then (2.2) is equivalent to V = (A, C) + [V > (A, C) + R(A, B) + E < (A)] = C n. (2.2) U r = V > (A, C) + R(A, B) + E < (A). (2.3) C n = U 0 U r (2.4) for some subspace U 0 V = (A, C). We call (2.4) an LR-decomposition with Riccati part U r and a Lyapunov complement U 0. If C n admits a decomposition (2.4) with nontrivial summands then (1.1) breaks up into a Lyapunov matrix equation and an irreducible Riccati equation. THEOREM 2.1 [W1]. (1) Let C n = U 0 U r be an LR-decomposition. If S = (S 0, S r ) is nonsingular such that Im S 0 = U 0, Im S r = U r, dim U r = n r, then S 1 AS A 0 0, S 1 B =, CS = (0, C r ) (2.5) A r0 A r 0 B r

and σ(a 0 ) C = (2.6) and V > (A r, C r ) + R(A r, B r ) + E < (A r ) = C nr. (2.7) Assume (2.5) (2.7). Then we have X T if and only if X = (S 1 ) X 0 0 S 1 (2.8) 0 X r and X 0 0 satisfies the Lyapunov equation L 0 (X 0 ) = A 0X 0 + X 0 A 0 = 0 and X r is a solution of R r (X r ) = A rx r + X r A r + X r B r B r X r C r C r = 0. (2) Let Π : C n U r be the projection on U r along U 0. Put ρ(x) = XΠ. Then ρ(x) T and (I Π) XΠ = 0. If X is given as in (2.8) then ρ(x) = (S 1 ) 0 0 0 X r S 1. Put S = ρ(t ) such that S contains all solutions of the form (2.8) with Lyapunov part X 0 = 0. Whether a Lyapunov complement U 0 does appear or not depends on purely imaginary eigenvalues of A. LEMMA 2.2 [W1]. Assume T. Then we have S = T if and only if rank(a λi, B) = n for all λ C =. (2.9) In the case where (A, B) is stabilizable the conditions (2.1) and (2.9) are satisfied and we have T and T = S. The following observation will be useful. LEMMA 2.3. For X T we have E = (A + BB X) = V = (A, C).

PROOF. Put A X = A + BB X. Then R(X) = 0 is equivalent to We recall the well known fact that a Lyapunov equation A XX + XA X = XB BX + C C. (2.10) F Y + Y F = D D where σ(f ) C = and Y is semidefinite implies D = 0. Assume E = (A X ) = Im I 0. Then and σ(â1) C =. Let A X = X =  1  12 0  2 1 X21 X 21 X 2 (2.11) (2.12) and C = (C 1 C 2 ) be partitioned accordingly. Then (2.10) yields  1X 1 + X 1  1 = (X 1 X21)BB X 1 X 21 + C 1C 1. Hence we obtain and C 1 = 0 and BB A = X 1 X 21 A 1 A 12 0 A 2 = 0 (2.13) (2.14) with A 1 = Â1. Therefore we have E = (A X ) V = (A, C). To prove the converse inclusion V = (A, C) E = (A X ) (2.15)

assume now that a basis has been chosen such that A and C are in the form (2.14) and C = (0, C 2 ). By a slight abuse of notation take X as in (2.12). Then A 1X 1 + X 1 A 1 = (X 1 X21)BB X 1 X 21. (2.16) As before we can regard (2.16) as a Lyapunov equation and conclude that (2.13) holds. From (2.14) we obtain A X in the form (2.11) with Â1 = A 1 which yields (2.15). subspaces of C n. Define There is an order isomorphism between S and the following system N of A-invariant N = {N N InvA, V (A, C) N V (A, C), N + +R(A, B) + E < (A) = C n }. (2.17) THEOREM 2.4 [W1]. (1) The map γ : S N given by γ(x) = Ker X is a bijection, and both γ and γ 1 are order preserving, i.e. for X, Y S and M, N N the relations X Y and M N imply γ(x) γ(y ) and γ 1 (M) γ 1 (N). (2) For X S we have Ker X = V (A, C) E > (A X ). (2.18) 3. THE GAP METRIC AND ISOLATED INVARIANT SUBSPACES For k IN let C k be endowed with the usual scalar product (x, y) = x y and with the corresponding norm x = (x x) 1/2. The norm of a matrix G C n m (or of a linear map G : C m C n ) is defined accordingly as operator norm, G = max{ Gx, x C m, x = 1}. Let P M denote the orthogonal projection of C n onto a subspace M of C n. The gap Θ(L, M) between two subspaces L and M of C n is defined as Θ(L, M) = P L P M. It is a metric in the set of all subspaces of C n. The following lemma can be used to compute the gap. LEMMA 3.1. Let L and M be subspaces of C n such that 1 dim L = dim M = r < n. Assume L = Im I r 0

and let be unitary such that Then U = U 1 U 12 U 21 U 2 M = Im U 1 and M = Im U 12 U 21 U 2 Θ(L, M) = Θ(L, M ) = U 21 = U 12.. PROOF. We have P L = I 0 (I, 0) and P M = U 1 U 21 (U 1, U 21). From (P L P M ) 2 = P L P L P M P M P L + P M = = 1U1 0 0 U 21 U21 and I U 1 U 1 = I U 1 U 1 = U 21 U 21 follows Θ(L, M) = U 21. Similarly we have Θ(L, M ) = U 12, and P L P M = (I P M ) (I P L ) = P M P L T yields Θ(L, M) = Θ(L, M ). For the proofs of the subsequent results we refer to [GLR]. LEMMA 3.2. Let L, M and V be subspaces of C n. (i) If Θ(L, M) < 1 then dim L = dim M. (ii) Assume L + V = C n. If Θ(L, M) is sufficiently small, then M + V = C n. LEMMA 3.3. Let X C n n be given. Then there exists a constant α > 0 such that for all Y C n n with dim Ker X = dim Ker Y (3.1) we have Θ(Ker X, Ker Y ) α X Y. (3.2)

Let A C n n be given. A subspace M InvA is called isolated if there is an ɛ > 0 such that the only subspace N InvA satisfying Θ(M, N) < ɛ is M itself. THEOREM 3.4. (1) A subspace M InvA is isolated if and only if for each λ σ(a) the primary component M E λ (A) is isolated as an A Eλ (A)-invariant subspace. (2) A subspace M Inv(A Eλ (A)) is isolated if and only either dim Ker(A λi) = 1 or dim Ker(A λi) 2 and M = 0 or M = E λ (A). 4. A HOMEOMORPHISM In this section it will be shown that the map γ : S N of Theorem 2.4 and its inverse are continuous. A technical lemma and a theorem on parameter dependence of least solutions will be needed. LEMMA 4.1. Let be a unitary n n matrix and let X = U = U 1 U 12 U 21 U 2 0 0 0 X 2, X 2 < 0, and Y = Y 1 Y 12 Y 12 Y 2 be hermitian n n matrices which are partitioned conformingly. Assume Then Ker Y = Im U 1 U 21 0, (4.1). (4.2) Θ(Ker X, Ker Y ) = U 21 = U 12. (4.3) Assume furthermore U 12 < 1 2. Then U 2 is nonsingular. Put T = (U 12 U 1 2 ). (4.4) Then Y = T I Y 2 (T I), Y 2 < 0, (4.5)

and we have U 12 T 2 U 12 (4.6) and Y X ( T 2 + T ) Y 2 + Y 2 X 2. (4.7) PROOF. The identity (4.3) is obvious from Lemma 3.1. Now let µ denote the smallest eigenvalue of U 2 U 2. Then U 12U 12 + U 2 U 2 = I (4.8) implies µ = 1 U 12 2 > 3 4. Hence U 2 is nonsingular and U 1 2 2 = 1 µ < 4 3 < 2, which implies T U 12 U 1 2 2 U 12. On the other hand we see from (4.8) that U 2 1 and conclude that U 12 T U 2 T. From (4.2) follows U Y U = diag(0, Λ) for some nonsingular Λ. Hence Y = U 12 U 2 Λ(U 12 U 2 ), and in particular Y 2 = U 2 ΛU2, which yields (4.5). The estimate (4.7) is obtained from Y X = diag(t Y 2 T, 0) + 0 T Y 2 Y 2 T 0 + diag(0, Y 2 X 2 ). THEOREM 4.2 [De],[R]. Let W be the set of all ordered triples (A, D, Q) of complex n n matrices with the following properties: (i) D 0, Q = Q, (A, D) is stabilizable. (ii) There exists a solution X = X of A X + XA + XDX Q = 0. (4.9) Then (4.9) has a least solution X, and X is a continuous function of (A, D, Q) W. We adapt the preceding theorem for our purposes. Consider the assumption

(iii) Q 0, ( A, Q) is detectable. It is well known (see e.g. [K]) that (i) together with (iii) implies (ii). In that case the least solution X is the unique negative-definite solution of (4.9). Theorem 4.2 remains valid if we replace (ii) by (iii). A proof that X depends continuously on the parameters of (4.9) could be given along the lines described in [GL, p. 1465] using the implicit function theorem. THEOREM 4.3. The map γ : S N given by γ(x) = Ker X is a homeomorphism. PROOF. It is not difficult to show that the map X γ(x) = Ker X is continuous. We fix a solution X S. According to Lemma 2.3 we have E = (A X ) = E = (A Y ) for all Y S. Hence if Y is sufficiently close to X then dim E > (A X ) = dim E > (A Y ) and (2.18) yields dim Ker X = dim Ker Y. Condition (3.1) of Lemma 3.3 is satisfied and (3.2) implies continuity of γ. In order to prove that the map γ 1 : N S is continuous we fix a subspace N N and choose an orthonormal basis of C n such that N = Im I 0. (4.10) Then the solution X S with γ(x) = Ker X = N is of the form X = diag(0, X 2 ). Note that N = Ker X InvA and N V (A, C) imply A = A 1 A 12 0 A 2, B = B 1 B 2, C = (0 C 2 ). (4.11) Furthermore N + R(A, B) + E < (A) = C n is equivalent to stabilizability of (A 2, B 2 ), whereas V (A, C) N means that ( A 2, C 2 ) is detectable. The matrix X 2 is the unique negative-definite (and thus the least) solution of R 2 (X 2 ) = A 2X 2 + X 2 A 2 + X 2 B 2 B 2X 2 C 2C 2 = 0. Now let P N and Y S be such that Θ = Θ(P, N) < 1 2 and Ker Y = P. Then Lemma 3.2 implies dim Ker X = dim Ker Y. Hence we can apply Lemma 4.1 and assume that Y and Ker Y are given by (4.1) and (4.2). In particular we have Θ = U 21. From R(Y ) = 0 and (4.11) we obtain (A 12 A 2) Y 21 + (Y 21 Y 2 ) A 12 + Y 2 A 2 +(Y 21 Y 2 ) B 1 (B1 B2) Y 21 C2C 2 = 0. B 2 Y 2 (4.12)

If T = (U 12 U 1 2 ) is defined as in (4.4) then (4.5) yields such that (4.12) can be written as Now set (Y 21 Y 2 ) = Y 2 (T I) (A 2 + T A 12 ) Y 2 + Y 2 (A 2 + T A 12 ) + +Y 2 (B 2 B 2 + T B 2 B 1 + B 1 B 2T + T B 1 B 1T )Y 2 C 2C 2 = 0. Ã 2 = A 2 + T A 12, B2 B 2 = B 2 B 2 + (T B 1 B 2 + B 2 B 1T + T B 1 B 1T ) such that Y 2 < 0 is a solution of R 2 (W 2 ) = Ã 2W 2 + W 2 Ã 2 + W 2 B2 B 2 W 2 C 2C 2 = 0. (4.13) Recall (4.6) which implies that U 12 0 is equivalent to T 0. Hence if Θ = U 12 is sufficiently small then Ã2 and B 2 B 2 are close to A 2 and B 2 B 2, respectively. Hence Y 2 is the least solution of (4.13) and according to Theorem 4.2 the solution Y 2 = Y 2 (T ) is a continuous function of T. Therefore lim Y 2(T ) = X 2. T 0 We conclude from (4.7) that lim γ 1 (N) γ 1 (P ) = 0 if Θ(N, P ) 0. σ(a)}. In particular Furthermore define 5. ISOLATED SOLUTIONS For N InvA define N λ = N E λ (A) such that N λ InvA and N = {N λ λ V λ (A, C) = V (A, C) E λ (A). N λ = {N N InvA, N V λ (A, C), N + [R(A, B) E λ (A)] = E λ (A)}. Then N λ is equivalent to V λ (A, C) N λ. It is not difficult to characterize an element N N by its components N λ. subspaces N λ satisfy LEMMA 5.1 [W2]. Assume N =. Then N InvA is in N if and only if the N λ = V λ (A, C) for all λ C

and N λ N λ for all λ C >. For the proof of Theorem 1.1 we have to determine the isolated elements of N. LEMMA 5.2. For a subspace N N the following statements are equivalent: (i) N is isolated in N. (ii) N is isolated in Inv(A V (A,C) ). (iii) If Re λ > 0 and rank A λi n 2 (5.1) C then either Ker A λi N = 0 (5.2) C or V (A, C) E λ (A) N. (5.3) PROOF. (i) (ii). Put à = A V (A,C) and Ãλ = A Vλ (A,C). Suppose N is not isolated in InvÃ. Consider the decomposition N = N λ. According to Theorem 3.4 there exists an α σ(a) such that N α is not isolated in InvÃα. We have α C > since N λ = V λ (A, C) if λ C. Now replace the component N α in N by a subspace M α InvÃα, M α N α, and set M = {N λ, λ α} M α. The proof of Theorem 3.4 (1) in [GLR, p. 429] shows that for M α sufficiently close to N α we have an estimate Θ(N, M) κθ(n α, M α ) (5.4) where κ is independent of M α. Recall that N α N α has the property N α + [R(A, B) E α (A)] = E α (A). Then Lemma 3.2 implies M α + [R(A, B) E α (A)] = E α (A),

which yields M α N α and M N. From (5.4) we conclude that we can find an M N, M N, arbitrarily close to N. Hence N is not isolated in N. Because of N Invà the implication (ii) (i) is obvious. (ii) (iii). Lemma 5.1 and Theorem 3.4 imply that (ii) holds if and only if for all λ C > the subspace N λ is isolated in InvÃλ. Because of dim Ker(Ãλ λi) = dim Ker A λi C only those eigenvalues λ have to be considered for which (5.1) holds. If dim Ker(Ãλ λi) 2 then according to Theorem 3.4 the subspace N λ is isolated in InvÃλ if and only if N λ = 0 or N λ = V λ (A, C). The first case is equivalent to (5.2) the second one to (5.3). PROOF OF THEOREM 1.1. Assume T. If the controllability condition (2.9) is not satisfied then according to Lemma 2.2 the subspace U r given by (2.3) is not the whole C n. Let X 0 be a solution of (1.1) and let S be as in Theorem 2.1 such that X = (S 1 ) diag(x 0, X r )S 1 and L 0 (X 0 ) = A 0X 0 + X 0 A 0 = 0, and σ(a 0 ) C =. Put Y 0 = X 0 (1 + ɛ). Then L 0 (Y 0 ) = 0. If ɛ is small then Y = (S 1 ) diag(y 0, X r )S 1 is a solution of (1.1) which is close to X. Hence X is not isolated in T. Now assume T and (2.9). Then T = S and N are homeomorphic. Hence a solution X is isolated in T if and only if N = Ker X is isolated in N, and we can apply Lemma 5.2. Note that (2.1) is equivalent to V (A, C) N. If N = then N = V (A, C) is the greatest element of N and we have (5.3) for all λ C. Then ˆX = γ 1 (V (A, C)) is the greatest negative-semidefinite solution of (1.1) and if (2.9) holds ˆX is isolated in T, which completes the proof.

REFERENCES [Da] P. Daughtry Isolated solutions of quadratic matrix equations, Linear Algebra Appl. 21 (1978), 89 94. [De] D.F. Delchamps, A note on the analicity of the Riccati metric, in: Algebraic and Geometric Methods in Linear Systems Theory, Lecture Notes in Applied Mathematics 18, Eds. C.I. Byrnes and C.F. Martin, Amer. Math. Soc., Providence, RI, 1980, pp. 37 41. [GL] P. Gahinet and A.J. Laub, Computable bounds for the sensitivity of the algebraic Riccati equation, SIAM J. Control Optim. 28 (1990), 1461 1480. [G] A.H.W. Geerts, A necessary and sufficient condition for solvability of the linear quadratic control problem without stability, Systems Control Lett. 11 (1988), 47 51. [GH] A.H.W. Geerts and M.L.J. Hautus, The output-stabilizable subspace and linear optimal control, Proceedings of the Internat. Symposium MTNS-89, Vol.II, Eds., M.A. Kaashoek et al., Birkhäuser, Boston, 1990, pp. 113 120. [GLR] I. Gohberg, P. Lancaster and L. Rodman, Invariant Subspaces of Matrices with Applications, Wiley, New York, 1986. [K] V. Kučera, Algebraic Riccati equations: hermitian and definite solutions, in: The Riccati Equation, Eds., S. Bittanti et al., Springer Verlag, Berlin, 1991. pp. 53 88. [RR] A.C.M. Ran and L. Rodman, Stability of invariant maximal semidefinite subspaces II, Linear Algebra Appl. 63 (1984), 133 173. [R] L. Rodman, On extremal solutions of the algebraic Riccati equation, in: Algebraic and Geometric Methods in Linear Systems Theory, Lecture Notes in Applied Mathematics 18, Eds. C.I. Byrnes and C.F. Martin, Amer. Math. Soc., Providence, RI, 1980, pp. 311 327. [W1] H.K. Wimmer, Decomposition and parameterization of semidefinite solutions of the continuous-time algebraic Riccati equation, to appear in SIAM J. Control Optim. [W2] H.K. Wimmer, Lattice properties of sets of semidefinite solutions of continuoustime algebraic Riccati equations, to appear in Automatica J. IFAC. Harald K. Wimmer Mathematisches Institut Universität Würzburg D-97074 Würzburg Germany. AMS subject classifications: 15A24, 47A55.