Yongge Tian. China Economics and Management Academy, Central University of Finance and Economics, Beijing , China
|
|
- Silvester Kelley
- 5 years ago
- Views:
Transcription
1 On global optimizations of the rank and inertia of the matrix function A 1 B 1 XB 1 subject to a pair of matrix equations B 2 XB 2, B XB = A 2, A Yongge Tian China Economics and Management Academy, Central University of Finance and Economics, Beijing 181, China Abstract. For a given linear matrix function A 1 B 1XB 1, where X is a variable Hermitian matrix, this paper derives a group of closed-form formulas for calculating the global maximum and minimum ranks and inertias of the matrix function subject to a pair of consistent matrix equations B 2XB 2 = A 2 and B XB = A. As applications, we give necessary and sufficient conditions for the triple matrix equations B 1XB 1 = A 1, B 2XB 2 = A 2 and B XB = A to have a common Hermitian solution. In addition, we discuss the global optimizations on the rank and inertia of the common Hermitian solution of the pair of matrix equations B 2XB 2 = A 2 and B XB = A. AMS subject classifications: 15A, 15A9, 15A24; 65K1; 65K15 Key words: Matrix function; matrix equation; common Hermitian solution; rank; inertia; maximization; minimization; Löwner partial ordering 1 Introduction Throughout this paper, C m n and C m H stand for the sets of all m n complex matrices and all m m complex Hermitian matrices, respectively. The symbols A T, A, r(a), R(A) and N (A) stand for the transpose, conjugate transpose, rank, range (column space) and null space of a matrix A C m n, respectively; I m denotes the identity matrix of order m; A, B denotes a row block matrix consisting of A and B. We write A > (A ) if A is Hermitian positive definite (nonnegative definite). Two Hermitian matrices A and B of the same size are said to satisfy the inequality A > B (A B) in the Löwner partial ordering if A B is positive definite (nonnegative definite). The Moore Penrose inverse of A C m n, denoted by A, is defined to be the unique solution X satisfying the four matrix equations (i) AXA = A, (ii) XAX = X, (iii) (AX) = AX, (iv) (XA) = XA. If X satisfies (i), it is called a g-inverse of A and is denoted by A. A matrix X is called a Hermitian g-inverse of A C m H, denoted by A, if it satisfies both AXA = A and X = X. Further, the symbols E A and F A stand for the two orthogonal projectors E A = I m AA and F A = I n A A. The ranks of E A and F A are given by r(e A ) = m r(a) and r(f A ) = n r(a). A well-known property of the Moore Penrose inverse is (A ) = (A ). In addition, AA = A A if A = A. We shall repeatedly use them in the latter part of this paper. The inertia of A is defined to be the triplet In(A) = { i + (A), i (A), i (A) }, where i + (A), i (A) and i (A) are the numbers of the positive, negative and zero eigenvalues of A counted with multiplicities, respectively. Both i + (A) and i (A), usually called the partial inertia, can easily be computed by elementary congruence matrix operations. For a Hermitian matrix A, the equality r(a) = i + (A) + i (A) holds. In recent years, some optimization problems on the global maximum and minimum ranks and inertias of the following linear matrix functions (LMFs) A BXB, A BX (BX), A BXC (BXC), (1.1) and their applications were studied, where A is a given complex Hermitian matrix, B and C are given complex matrices, and X is a variable matrix of appropriate size; see 14, 15, 16, 17, 18, 28, 1, 2,. As a generalization, we address in this paper the following optimization problem: Problem 1.1 For a given LMF A 1 B 1 XB 1 and a pair of matrix equations B 2 XB 2, B XB = A 2, A that have a a common Hermitian solution, give formulas for calculating the following global Address: yongge.tian@gmail.com
2 maximum and minimum ranks and inertias max X C n H min X C n H r( A 1 B 1 XB 1 ) s.t. B 2 XB 2, B XB = A 2, A, (1.2) r( A 1 B 1 XB 1 ) s.t. B 2 XB 2, B XB = A 2, A, (1.) max i ± ( A 1 B 1 X ) s.t. B 2 XB2, B XB = A 2, A, (1.4) X C n H min i ± ( A 1 B 1 X ) s.t. B 2 XB2, B XB = A 2, A. (1.5) X C n H We shall use some pure algebraic operations on matrices to derive a group of analytical formulas for calculating the global maximum and minimum values of the objective functions in (1.2) (1.5), and then to present a variety of valuable consequences of these formulas. In particular, we shall use the minimum rank formula for (1.) to derive necessary and sufficient conditions for the triple matrix equations B 1 XB 1 = A 1, B 2 XB 2 = A 2, B XB = A (1.6) to have a common Hermitian solution. The rank and inertia of a Hermitian matrix are two basic concepts in matrix theory for describing the dimension of the row/column vector space and the sign distribution of the eigenvalues of the matrix, which are well understood and are easy to compute by the well-known elementary or congruent matrix operations. These two quantities play an essential role in characterizing algebraic properties of Hermitian matrices. Because the rank and inertia of a matrix are finite nonnegative integers, the global maximum and minimum values of the rank and inertia of a matrix expression always exist no matter what the domains of variable entries in the matrix expression are given. The extremal ranks and inertias of a matrix expression can directly be used to characterize some fundamental algebraic properties of the matrix expression, for example, (I) the maximum and minimum dimensions of the row and column spaces of the matrix expression; nonsingularity of the matrix expression when it is square; (III) solvability of the corresponding matrix equation; (IV) rank, inertia and range invariance of the matrix expression; (V) definiteness of the matrix expression when it is Hermitian; etc. Since variable entries in a matrix function are often regarded as continuous variables in some constrained sets, while the objective functions the rank and inertia of the matrix function take values only from a finite set of nonnegative integers, Hence, (1.2) (1.5) can be regarded as continuous-integer optimization problems subject to equality constraints. This kind of nonsmooth optimization problems cannot be solved by using various optimization methods for solving continuous or discrete cases. There is no rigorous mathematical theory for solving a general rank and inertia optimization problem due to the discontinuity and nonconvexity of rank and inertia of matrix. In fact, it has been realized that rank and inertia optimization problems have deep connections with computational complexity, are regarded as NP-hard in general settings; see, e.g., 1, 2,, 4, 6, 7, 8, 11, 2, 2, 25. The following are some known results for ranks and inertias of matrices and their usefulness, which will be used in the latter part of this paper. Lemma 1.2 (28) Let H be a set consisting of Hermitian matrices over C m H. Then, (a) H has a matrix X > (X < ) if and only if max X H i +(X) = m (b) All X H satisfy X > (X < ) if and only if min X H i +(X) = m (c) H has a matrix X (X ) if and only if min X H i (X) = (d) All X H satisfy X (X ) if and only if max X H i (X) = ( ) max i (X) = m. X H ( ) min i (X) = m. X H ( ) min i +(X) =. X H ( ) max i +( X) =. X H 2
3 Lemma 1. (19) Let A C m n, B C m p and C C q n. Then, the following rank expansion formulas hold r A, B = r(a) + r(e A B) = r(b) + r(e B A), (1.7) A r = r(a) + r(cf C A ) = r(c) + r(af C ), (1.8) A B r = r(b) + r(c) + r(e C B AF C ). (1.9) Three useful rank expansion formulas derived from (1.9) are A B A B r = r(p ) + r C P E P C A B r C A BFQ = r(q) + r C Q A B r C P Q = r(p ) + r(q) + r A BFQ E P C, (1.1), (1.11) We shall use them in Section 2 to simplify ranks of block matrices involving E P and F Q. Lemma 1.4 (28) Let A C m H, B Cm n, D C m n, and let A B A B U = B, V = B. D Then, the inertias of U and V can be expanded as (a) If A, then (b) If A, then (c) If R(B) R(A), then. (1.12) i ± (U) = r(b) + i ± (E B AE B ), (1.1) E i ± (V ) = i ± (A) + i A B ± B E A D B A. (1.14) B i + (U) = r A, B, i (U) = r(b), r(u) = r A, B + r(b). (1.15) i + (U) = r(b), i (U) = r A, B, r(u) = r A, B + r(b). (1.16) i ± (V ) = i ± (A) + i ± ( D B A B ), r(v ) = r(a) + r( D B A B ). (1.17) (d) If R(B) R(A) = {} and R(B ) R(D) = {}, then i ± (V ) = i ± (A) + i ± (D) + r(b), r(v ) = r(a) + 2r(B) + r(d). (1.18) Three expansion formulas derived from (1.1) are A B A B A BF i P ± F P B = i ± B P A BF r(p ), r P F P P B = r B P 2r(P ). (1.19) P We shall use them to simplify the inertias of block Hermitian matrices that involve F P = I P P. Lemma 1.5 Let A C m n and B C m H be given. Then,
4 (a) 5, 9 The matrix equation AXA = B has a solution X C n H if and only if R(B) R(A), or equivalently, AA B = B. (b) 28 Under AA B = B, the general Hermitian solution of AXA = B can be written in the following two forms where U C n H and V Cn n are arbitrary. X = A B(A ) + U A AUA A, (1.2) X = A B(A ) + F A V + V F A, (1.21) More results on properties of solutions of AXA = B can be found in 15, 18. Lemma 1.6 Let A j C mj n, B j C p qj and C j C mj qj be given, j = 1, 2. Then, (a) 24 The pair of matrix equations A 1 XB 1 = C 1 and A 2 XB 2 = C 2 (1.22) have a common solution for X C n p if and only if C 1 A 1 R(C j ) R(A j ), R(Cj ) R(Bj ), r C 2 A 2 A1 = r +r B A 1, B 2, j = 1, 2. (1.2) B 1 B 2 (b) 27 Under (1.27), the general common solution to (1.26) can be written in the following parametric form X = X + F A V 1 + V 2 E B + F A1 V E B2 + F A2 V 4 E, (1.24) A1 where A =, B = B A 1, B 2, and the four matrices V 1,..., V 4 C n p are arbitrary. 2 Lemma 1.7 (17) Let A C m H, B Cm p and C C q m be given, and let A B A C M 1 = B, M 2 =, (1.25) C N = A, B, C A B C A B C, N 1 = B, N 2 =. (1.26) C Then, the global maximum and minimum rank and inertias of A BXC (BXC) are given by where max r A BXC (BXC) = min {r(n), r(n 1 ), r(n 2 )}, (1.27) X C p q min r A BXC (BXC) = 2r(N) + max{ s 1, s 2, s, s 4 }, (1.28) X C p q max i ± A BXC (BXC) = min{i ± (M 1 ), i ± (M 2 )}, (1.29) X C p q min i ± A BXC (BXC) = r(n) + max{ i ± (M 1 ) r(n 1 ), i ± (M 2 ) r(n 2 ) }, (1.) X C p q s 1 = r(m 1 ) 2r(N 1 ), s 2 = r(m 2 ) 2r(N 2 ), s = i + (M 1 ) + i (M 2 ) r(n 1 ) r(n 2 ), s 4 = i + (M 1 ) + i (M 2 ) r(n 1 ) r(n 2 ). In particular, if R(C ) R(B), then { } A C max r A BXC X C (BXC) = min r A, B, r, (1.1) p q C A C min r A BXC A B (BXC) = 2r A, B + r 2r, (1.2) X C p q C C max i ± A BXC (BXC) A C = i ±, (1.) X C p q C min i ± A BXC (BXC) A C A B = r A, B + i ± r. (1.4) X C p q C C 4
5 The matrices Xs that satisfy (1.27) (1.) (namely, the global maximizers and minimizers of the objective rank and inertia functions) are not necessarily unique and their expressions were also given in 17 by using certain simultaneous decomposition of the three given matrices. Observe that the righthand sides of (1.27) (1.) are represented in analytical forms of the ranks and inertias of the five given block matrices, we can easily use them to derive extremal ranks and inertias of some general linear and nonlinear matrix functions. In these cases, combining the rank and inertia formulas obtained with the assertions in Lemma 1.1 may yield various conclusions on algebraic properties of linear and nonlinear matrix functions. 2 The global maximum and minimum ranks and inertias of A 1 B 1 XB 1 subject to a consistent matrix equation Because of the noncommutativity of matrix multiplications, solving matrix equations has been a challenging topic of study in linear algebra. If some matrix equations of appropriate sizes are given together, it is natural to ask whether the equations have a possible common solution. For example, if two Hermitian matrix equations B 1 X 1 = A 1 and B 2 X 2 B2 = A 2, are given, where X 1 and X 2 have the same size, and each of them has a solution. In such a case, it would be of interest to seek relations between the Hermitian solutions of the two equations. As mentioned in the previous section, the existence of solution of a matrix equation can be characterized by the minimum rank of the corresponding matrix expression. Further, relations between two matrix equations can also be characterized by rank/methods. Some recent work on the rank and inertia method in the investigation of Hermitian matrix equations can be found in 14, 15, 16, 17, 28,. In 17, the extremal ranks and inertias of A 1 B 1 X subject to a consistent matrix equation B 2 XB2 = A 2 were studied and the following results were obtained. Lemma 2.1 (17) Let A i C mi H and B i C mi n be given for i = 1, 2, and suppose that the matrix equation B 2 XB2 = A 2 has a Hermitian solution. Also, let A 1 B 1 S 2 = { X C n H B 2 XB2 = A 2 }, M = A 2 B 2. (2.1) B2 Then, the global maximum and minimum rank and inertias of A 1 B 1 XB 1 s.t. X S are given by Hence, max r( A 1 B 1 X ) = min {r A 1, B 1, r(m) 2r(B 2 )}, (2.2) 2 min r( A 1 B 1 X A1 B ) = 2r A 1, B 1 2r 1 2 B2 + r(m), (2.) max i ± ( A 1 B 1 X ) = i ± (M) r(b 2 ), (2.4) 2 min i ± ( A 1 B 1 X A1 B ) = r A 1, B 1 r 1 2 B2 + i ± (M). (2.5) (a) There exists an X C n H such that B 2XB 2 = A 2 and A 1 B 1 XB 1 is nonsingular if and only if r A 1, B 1 = m 1 or r(m) = 2r(B 2 ) + m 1. (b) There exists an X C n H such that B 1X = A 1 and B 2 XB2 = A 2 if and only if R(A 1 ) R(B 1 ) and R(A 2 ) R(B 2 ) and r(m) = 2r. (2.6) B 2 (c) There exists an X C n H such that B 2XB 2 = A 2 and A 1 B 1 XB 1 > if and only if i + (M) = r(b 2 ) + m 1. (d) There exists an X C n H such that B 2XB 2 = A 2 and A 1 B 1 XB 1 < if and only if i (M) = r(b 2 ) + m 1. 5
6 (e) There exists an X C n H such that B 2XB2 = A 2 and A 1 B 1 X if and only if A1 B R(A 2 ) R(B 2 ) and i + (M) = r 1 B2 r A 1, B 1. (2.7) (f) There exists an X C n H such that B 2XB2 = A 2 and A 1 B 1 X if and only if A1 B R(A 2 ) R(B 2 ) and i (M) = r 1 B2 r A 1, B 1. (2.8) Corollary 2.2 Let A i C mi H and B i C mi n be given with B i for i = 1, 2, and suppose that each of the two matrix equations B 1 X = A 1 and B 2 XB2 = A 2 has a Hermitian solution. Also denote the sets of all solutions of B 1 X = A 1 and B 2 XB2 = A 2 by S 1 and S 2 respectively. Then, the global maximum and minimum ranks and inertias of A 1 B 1 X subject to X S 2 are given by where M is as given in (2.1). In particular, max r( A 1 B 1 X ) = min {r(b 1 ), r(m) 2r(B 2 )}, (2.9) 2 min r( A 1 B 1 X ) = r(m) 2r, (2.1) 2 B 2 max i ± ( A 1 B 1 X ) = i ± (M) r(b 2 ), (2.11) 2 min i ± ( A 1 B 1 X ) = i ± (M) r, (2.12) 2 B 2 (a) S 2 S 1 if and only if r(m) = 2r(B 2 ). (b) S 2 = S 1 if and only if r(m) = 2r(B 1 ) = 2r(B 2 ). The above results can be used to derive algebraic properties of the submatrices in a solution to the matrix equation BXB = A. Rewrite BXB = A as X1 X B 1, B 2 2 B 1 X2 X B2 = A, (2.1) where B 1 C m n1, B 2 C m n2, X 1 C n1 H, X 2 C n1 n2 and X C n2 H with n 1 + n 2 = n. We next derive the global maximum and minimum ranks and inertias of the submatrices X 1 and X in a Hermitian solution to (2.1). Note that X 1, X 2, X in (2.1) can be rewritten as X 1 = P 1 XP 1, X 2 = P 1 XP 2, X = P 2 XP 2, (2.14) where P 1 = I n1, and P 2 =, I n2. For convenience, we adopt the following notation for the collections of the submatrices X 1 and X in (2.1): { } T 1 = X 1 C n1 H B X1 X 1, B 2 2 B 1 X2 X B2 = A = {X 1 = P 1 XP1 BXB = A}, (2.15) { } T = X C n2 H B X1 X 1, B 2 2 B 1 X2 X B2 = A = {X = P 2 XP2 BXB = A}. (2.16) Applying Corollary 2.2 to (2.15) and (2.16) gives the following results. Theorem 2. Suppose that the matrix equation (2.1) is consistent, and let T 1 and T be of the forms (2.15) and (2.16). Then, { } A B2 max r(x 1 ) = min n 1, r X 1 T 1 B2 2r(B) + 2n 1, (2.17) A B2 min r(x 1 ) = r X 1 T 1 B2 2r(B 2 ), (2.18) max i ± (X 1 ) = i ± X 1 T 1 A r(a) + n 1, (2.19) min i ± (X 1 ) = i ± X 1 T 1 A r(b 2 ), (2.2) 6
7 and { } A max r(x ) = min n 2, r X T 2r(B) + 2n 2, (2.21) A min r(x ) = r X T 2r(B 1 ), (2.22) C max i ± (X ) = i ± X T A r(a) + n 1 2, (2.2) C min i ± (X ) = i ± X T A r(b 1 1 ). (2.24) Hence, (c) Eq. (2.1) has a solution in which X 1 is nonsingular if and only if r A 2r(A) n 1. (d) The submatrix X 1 in any solution to (2.1) is nonsingular if and only if r A = 2r(B 2 )+n 1. (e) Eq. (2.1) has a solution in which X 1 = if and only if r A = 2r(B 2 ). (f) The submatrix X 1 in any solution to (2.1) satisfies X 1 = if and only if r A = 2r(A) 2n 1. (g) Eq. (2.1) has a solution in which X 1 > (X 1 < ) if and only if ( ) i + A = r(a) i A = r(a). (h) The submatrix X 1 in any solution to (2.1) satisfies X 1 > (X 1 < ) if and only if ( ) i + A = n 1 + r(b 2 ) i A = n 1 + r(b 2 ). (i) Eq. (2.1) has a solution satisfying X 1 (X 1 ) if and only if ( ) i A = r(b 2 ) i + A = r(b 2 ). (j) The submatrix X 1 in any solution to (2.1) satisfies X 1 (X 1 ) if and only if ( ) i A = r(a) n 1 i A = r(a) n 1. (k) The positive signature of X 1 in (2.1) is invariant the negative signature of X 1 (2.1) is invariant R(B 1 ) R(B 2 ) = {} and r(b 1 ) = n 1. The definiteness of the Hermitian solutions of a given consistent matrix equation is attractive topic in matrix theory and applications; see, e.g., 1, 9,, 5. From Corollary 2.2, we now can derive the existences of Hermitian solutions of AXA = B satisfying some inequalities. Corollary 2.4 (1) Let A C m n, B C m H and A 1 C n H be given, and assume that the matrix equation AXA = B has a solution for X C n H and define S = { X Cn H AXA = B}. Then, Hence, max i ±( X A 1 ) = n + i ± ( B AA 1 A ) r(a), (2.25) min i ±( X A 1 ) = i ± ( B AA 1 A ), (2.26) max i ±(X) = n + i ± (B) r(a), (2.27) min i ±(X) = i ± (B). (2.28) 7
8 (a) AXA = B has a solution X > A 1 (X < A 1 ) if and only if i + ( B AA 1 A ) = r(a) (i ( B AA 1 A ) = r(a)). (b) AXA = B has a solution X A 1 (X A 1 ) if and only if B AA 1 A (B AA 1 A ). (c) AXA = B has a solution X > (X < ) if and only if B and r(a) = r(b) (B and r(a) = r(b)). (d) AXA = B has a solution X (X ) if and only if B (B ). The global maximum and minimum ranks and inertias of A B 1 XB 1 subject to a pair of matrix equations We first derive a parametric form for the general common Hermitian solution of the pair of matrix equations in (1.2) (1.5). Lemma.1 (1) Let A i C mi H, B i C mi n be given for i = 2,, and suppose that each of the two matrix equations has a solution, i.e., R(A i ) R(B i ) for i = 2,. Then, B 2 XB 2 = A 2 and B XB = A (.1) (a) The pair of matrix equations have a common Hermitian solution if and only if A B 2 r A B B2 = 2r. (.2) B2 B B (b) Under (.2), the general common Hermitian solution of the pair of equations can be written in the following parametric form X = X + V F B + F B V + F B2 UF B + F B U F B2, (.) B2 where X is a special Hermitian common solution to the pair of equations, B =, and U, V C n n are arbitrary. Substituting (.) into A 1 B 1 XB 1 gives A 1 B 1 XB 1 = A 1 B 1 X B 1 B 1 V F B B 1 B 1 F B V B 1 B 1 F B2 UF B B 1 B 1 F B U F B2 B 1, (.4) which is a matrix expression involving two variable matrices V and U. Thus, the constrained matrix expression in (1.2) is equivalently converted to the unconstrained matrix expression in (.4). To find the global maximum and minimum ranks and inertias of (.4), we need the following result. Lemma.2 Let p(x 1, X 2 ) = A B 1 X 1 C 1 (B 1 X 1 C 1 ) B 2 X 2 C 2 (B 2 X 2 C 2 ), (.5) where A C m H, B i C m pi and C i C qi m are given, and X i C pi qi are variable matrices for i = 1, 2, and assume that Also let N = M = A B2 C 1 C 2 C 1 R(B 2 ) R(B 1 ), R(C 1 ) R(B 1 ), R(C 2 ) R(B 1 ). (.6) A, M C 1 1 = A B 2 C1 C2 A B 2 C, N 1 = B2 1 C2, N 2 = C 1, C 1 C A B 2 C1 A C B2 1 C2, M 2 = C 1. C 1 C B 8
9 Then, the global maximum and minimum ranks and inertias of p(x 1, X 2 ) are given by max r p(x 1, X 2 ) = min{r A, B 1, r(n), r(m 1 ), r(m 2 )}, (.7) X 1 C p 1 q 1, X 2 C p 2 q 2 min r p(x 1, X 2 ) = 2r A, B 1 2r(M) + 2r(N) + max{ s 1, s 2, s, s 4 }, (.8) X 1 C p 1 q 1, X 2 C p 2 q 2 max i ± p(x 1, X 2 ) = min{i ± (M 1 ), i ± (M 2 )}, (.9) X 1 C p 1 q 1, X 2 C p 2 q 2 min i ± p(x 1, X 2 ) = r A, B 1 r(m) + r(n) X 1 C p 1 q 1, X 2 C p 2 q 2 where + max{ i ± (M 1 ) r(n 1 ), i ± (M 2 ) r(n 2 ) }, (.1) s 1 = r(m 1 ) 2r(N 1 ), s 2 = r(m 2 ) 2r(N 2 ), s = i + (M 1 ) + i (M 2 ) r(n 1 ) r(n 2 ), s 4 = i + (M 1 ) + i (M 2 ) r(n 1 ) r(n 2 ). Proof Under (.6), applying Lemma 1.7 to the variable matrix X 1 in (.5) and simplifying, we obtain { } max r p(x 1, X 2 ) = min r A B 2 X 2 C 2 (B 2 X 2 C 2 ) A B2 X, B 1, r 2 C 2 (B 2 X 2 C 2 ) C1 X 1 C 1 { } A B2 X = min r A, B 1, r 2 C 2 (B 2 X 2 C 2 ) C1, (.11) C 1 min r p(x 1, X 2 ) = 2r A B 2 X 2 C 2 (B 2 X 2 C 2 ) A B2 X, B 1 + r 2 C 2 (B 2 X 2 C 2 ) C1 X 1 C 1 A B2 X max i ± p(x 1, X 2 ) = i 2 C 2 (B 2 X 2 C 2 ) C1 ±, (.12) X 1 C 1 A B2 X 2r 2 C 2 (B 2 X 2 C 2 ) B 1 C 1 A B2 X = 2r A, B 1 + r 2 C 2 (B 2 X 2 C 2 ) C1 A 2r, (.1) C 1 C 1 min i ± p(x 1, X 2 ) = r A B 2 X 2 C 2 (B 2 X 2 C 2 ) A B2 X, B 1 + i 2 C 2 (B 2 X 2 C 2 ) C1 ± X 1 C 1 A B2 X r 2 C 2 (B 2 X 2 C 2 ) B 1 C 1 = r A, B 1 + i ± A B2 X 2 C 2 (B 2 X 2 C 2 ) C 1 C 1 Notice that A B2 X 2 C 2 (B 2 X 2 C 2 ) C1 A C = 1 C 1 C 1 Applying Lemma 1.6 to this expression gives r A C 1. (.14) B2 C X 2 C 2, 2 X 2 B2, := q(x 2 ). (.15) max rq(x 2 ) = min { r(n), r(m 1 ), r(m 2 ) }, (.16) X 2 C m p 2 min rq(x 2 ) = 2r(N) + max{ s 1, s 2, s, s 4 }, (.17) X 2 C m p 2 max i ± q(x 2 ) = min{ i ± (M 1 ), i ± (M 2 ) }, (.18) X 2 C m p 2 min i ± q(x 2 ) = r(n) + max{ i ± (M 1 ) r(n 1 ), i ± (M 2 ) r(n 2 ) }, (.19) X 2 C m p 2 where s 1 = r(m 1 ) 2r(N 1 ), s 2 = r(m 2 ) 2r(N 2 ), 9
10 s = i + (M 1 ) + i (M 2 ) r(n 1 ) r(n 2 ), s 4 = i (M 1 ) + i + (M 2 ) r(n 1 ) r(n 2 ). Substituting these results into (.11) (.14) yields (.9) (.2). It is obviously of great importance to be able to give analytical formulas for calculating the global maximum and minimum ranks and inertias of the matrix expression in (.5) under the assumptions in (.7). However, it is not easy to find the global maximum and minimum ranks and inertias ranks and inertias of a general p(x 1, X 2 ) as given in (.5). For convenience of representation, we rewrite (.4) as where A 1 B 1 XB 1 = A G 1 V G 2 (G 1 V G 2 ) G UG 4 (G UG 4 ), (.2) A = A 1 B 1 X B 1, G 1 = B 1, G 2 = F B B 1, G = B 1 F B2, G 4 = F B B 1. (.21) It is easy to verify that the above matrices satisfy the conditions (G 2) R(G 1 ), R(G ) R(G 1 ), R(G 4) R(G 1 ), R(G 2) R(G ), R(G 2) R(G 4). (.22) In this case, applying Lemma.2 to (.22) yields the main results of this section. Theorem. Let A i C mi H equations and B i C mi n be given for i = 1, 2,, and assume that the pair of matrix B 2 XB 2 = A 2 and B XB = A (.2) have a common solution X C n H. Also denote the set of all their common Hermitian solutions by and let A1 B P 1 = 1 B2 B S = {X C n H B 2 XB 2 = A 2, B XB = A }. (.24), P 2 = A 1 B 1 B 1 A 1 B 1 B 1 Q 1 = A B A B, Q 2 = A 2 B B B2 B 1 B2 B Then, (a) The global maximum rank of A 1 B 1 XB 1 subject to (.24) is max r( A 1 B 1 X ) { = min r A 1, B 1, r(q 1 ) r A 1 B 1 A 2 B 2, P = A 1 B 1 A B, (.25) B2 B A 1 B 1 B 1, Q = A B B. B 2 (.26) B2 (b) The global minimum rank of A 1 B 1 XB 1 subject to (.24) is where min r( A 1 B 1 X ) } r(b B 2 ) r(b ), r(p 2 ) 2r(B 2 ), r(p ) 2r(B ). (.27) = 2r A 1, B 1 2r(P 1 ) + 2r(Q 1 ) + max{ r(p 2 ) 2r(Q 2 ), r(p ) 2r(Q ), u 1, u 2 }, (.28) u 1 = i + (P 2 ) + i (P ) r(q 2 ) r(q ), u 2 = i (P 2 ) + i + (P ) r(q 2 ) r(q ). (c) The global maximum partial inertia of A 1 B 1 XB 1 subject to (.24) is max i ±( A 1 B 1 X ) = min {i ± (P 2 ) r(b 2 ), i ± (P ) r(b )}. (.29) 1
11 (d) The global minimum partial inertia of A 1 B 1 XB 1 subject to (.24) is min i ±( A 1 B 1 X ) = r A 1, B 1 r(p 1 ) + r(q 1 ) (.) + max{i ± (P 2 ) r(q 2 ), i ± (P ) r(q )}. (.1) Proof Under (.22), we find by Lemma.2 that max r( A 1 B 1 X ) = max r A G 1V G 2 (G 1 V G 2 ) G UG 4 (G UG 4 ) V, U { A G G = min r A, G 1, r 4 A G, r G G, r A G 4 G 4 min r( A 1 B 1 X ) = min r A G 1V G 2 (G 1 V G 2 ) G UG 4 (G UG 4 ) V, U A G1 A G G = 2r A, G 1 2r + 2r 4 G G }, (.2) + max{s 1, s 2, s, s 4 }, (.) max i ±( A 1 B 1 X ) = max i ± A G 1 V G 2 (G 1 V G 2 ) G UG 4 (G UG 4 ) V, U { } A G A G = min i ± G, i 4 ±, (.4) G 4 min i ±( A 1 B 1 X ) = min i ± A G 1 V G 2 (G 1 V G 2 ) G UG 4 (G UG 4 ) V, U A G1 A G G = r A, G 1 r + r 4 G G + max{t 1, t 2 }, (.5) where A G A G G s 1 = r G 2r 4 G, A G s 2 = r 4 A G G 2r 4, G 4 G 4 A G A G s = i + G + i 4 A G G r 4 G 4 G A G A G s 4 = i G + i 4 A G G + r 4 G 4 G A G A G G t 1 = i ± G r 4 G, A G4 A G G t 2 = i ± G r 4 4 G. 4 A G G r 4, G 4 r A G G 4 G 4, 11
12 Applying (1.1) (1.12) and (1.19), and simplifying by ( B 2 X B 2, B X B ) = ( A 2, A ), elementary matrix operations and congruence matrix operations, we obtain r A, G 1 = r A 1 B 1 X, B 1 = r A 1, B 1, (.6) A G G r 4 A1 B = r 1 X B 1 F B2 B 1 F B G F B A 1 B 1 X B 1 B 1 = r B B r(b) r(b 2) r(b ) B A 1 B 1 B 1 B 1 X B = r B B r(b) r(b 2) r(b ) B A 1 B 1 B 1 = r B2 B B A r(b) r(b 2) r(b ) B A A G1 A1 B r = r 1 X B 1 G F B G A i ± G = r(q 1 ) r(b) r(b 2 ) r(b ), (.7) A1 B = r 1 B r(b) = r(p 1 ) r(b), (.8) A1 B = i 1 X A B 1 F 1 B 1 X B 1 B2 ± F B2 = i ± B2 r(b 2 ) B A 1 B 1 B 1 X B2/2 A 1 B 1 = i ± B2 r(b 2 ) = i ± B B 1 X B2/2 1 B 2 r(b 2 ) B B 2 A 2 = i ± (P 2 ) r(b 2 ), (.9) A G G r 4 A1 B G = r 1 X B 1 F B2 B 1 F B F B2 A 1 B 1 X B 1 B 1 = r B2 B 2r(B 2) r(b ) B A 1 B 1 B 1 B 1 X B2 = r B 2 B 2r(B 2) r(b ) B A 1 B 1 B 1 = r B2 2r(B 2) r(b ) B A 2 B = r(q 2 ) 2r(B 2 ) r(b ). (.4) By a similar approach, we can obtain A G4 A G G i ± G = i 4 ± (P ) r(b ), r 4 = r(q G 4 ) r(b 2 ) 2r(B ). (.41) Substituting (.6) (.41) into (.2) (.5) yields (.27) (.1). Some direct consequences of the previous theorem are given below. 12
13 Corollary.4 Let A i C mi H and B i C mi n be given for i = 1, 2,, and suppose that each pair of B 1 X = A 1, B 2 XB2 = A 2 and B XB = A have a common Hermitian solution. Also let S be of the form (.2). Then, { max r( A 1 B 1 X B2 ) = min r(b 1 ), r(q 1 ) r r(b B 2 ) r(b ), } 2r 2r(B B 2 ), 2r 2r(B 2 B ), (.42) B 1 B 1 B 1 min r( A 1 B 1 X ) = 2r(Q 1 ) 2r B 2 2r B, (.4) B B { } max i ±( A 1 B 1 X ) = min r r(b B 2 ), r r(b 2 B ), (.44) min i ±( A 1 B 1 X ) = r(q 1 ) r B 1 B 2 r B 1 B 1 B, (.45) B B where Q 1 is of the form (.26). Proof Under the given conditions, the ranks/inertias of the block matrices in (.25) and (.26) are given by r(p 1 ) = r(b 1 ) + r, r(p ) = 2r, i ± (P 2 ) = r, i ± (P ) = r, B 1 B 2 B, r(p 2 ) = 2r B 2 B 1 B 1 r(q 2 ) = r B B Hence (.27) (.1) reduce to (.42) (.45). + r B 2 B B 1 B 1, r(q ) = r B B B 2 + r B Corollary.5 Let A i C mi mi H and B i C mi n be given for i = 1, 2,, and suppose that each pair of the triple matrix equations B 1 XB 1 = A 1, B 2 XB 2 = A 2, B XB = A (.46) have a common Hermitian solution. Then, there exists a Hermitian X such that (.46) holds if and only if A 1 B 1 B 1 r A B B 1 B 1 A B = r B + r, B2, B. (.47) B2 B B Proof It follows from (.4). A challenging open problem on the triple matrix equations in (.46) is to give a parametric form for their general common Hermitian solution. Setting B 1 = I n in Theorem. may yield a group of results on the extremal ranks/inertias of A 1 X subject to (.24). In particular, we have the following consequences. Corollary.6 Let A i C mi H and B i C mi n be given for i = 2,, and assume that (.2) has a common solution. Also let S be of the form (.24). Then, (a) The global maximum rank of the solution of (.24) is where max r(x) = min{ n, s 1, s 2, s }, (.48) A B s 1 = 2n + r 2 B2 r r(b A B B 2 ) r(b ), s 2 = 2n + r(a 2 ) 2r(B 2 ), s = 2n + r(a ) 2r(B ).. B 1
14 (b) The global minimum rank of the solution of (.24) is A B min r(x) = 2r 2 + max{ t A B 1, t 2, t, t 4 }, (.49) where A2 B t 1 = r(a 2 ) 2r 2 B t = i + (A 2 ) + i (A ) r t 4 = i (A 2 ) + i + (A ) r B2, t 2 = r(a ) 2r A2 B B2 r, B A B A2 B B2 B r A B A B,. (c) The global maximum partial inertia of the solution of (.24) is max i ±(X) = min{ n + i ± (A 2 ) r(b 2 ), n + i ± (A ) r(b ) }. (.5) (d) The global minimum partial inertia of the solution of (.24) is { min i A B ±(X) = r 2 A2 B + max i A B ± (A 2 ) r 2 B Hence, (e) Eq. (.2) has a solution X > if and only if, i ± (A ) r A 2, A, R(A 2 ) = R(B 2 ), R(A ) = R(B ). (f) All solutions of (.2) satisfy X > if and only if A 2, A and one of (g) Eq. (.2) has a solution X < if and only if r(a 2 ) = r(b 2 ) = n, r(a ) = r(b ) = n. A 2, A, R(A 2 ) = R(B 2 ), R(A ) = R(B ). (h) All solutions of (.2) satisfy X < if and only if A 2, A and one of r(a 2 ) = r(b 2 ) = n, r(a ) = r(b ) = n. (i) Eq. (.2) has a solution X if and only if A B2 A 2, A, R R A B, R R A (j) All solutions of (.2) satisfy X if and only if A 2, A and one of r(b 2 ) = n and r(b ) = n. B2 A2 B 2 A B B (k) Eq. (.2) has a solution X if and only if A B A2 B A 2, A, R R, R R 2. A B A B (l) All solutions of (.2) satisfy X if and only if A 2, A and one of r(b 2 ) = n and r(b ) = n.. }. (.51) 14
15 Proof Set A 1 = and B 1 = I n in Theorem. and simplifying, we obtain (a) (d). Applying Lemma 1.2 to (.49) and (.5), we obtain (e) (l). Rewrite B 2 XB2 = A 2 and B XB = A as X1 X B 21, B 22 2 B 21 X1 X X2 = A X 2, B 1, B 2 2 B 1 X2 = A X, (.52) B 22 where B i1 C mi n1, B i2 C mi n2, i = 2,, X 1 C n1 H, X 2 C n1 n2 and X C n2 H with n 1 + n 2 = n. We next derive the extremal ranks and inertias of the submatrices X 1 and X in a Hermitian solution of (.52). Note that X 1, X 2, X in (.52) can be rewritten as X 1 = P 1 XP 1, X 2 = P 1 XP 2, X = P 2 XP 2, (.5) where P 1 = I n1, and P 2 =, I n2. For convenience, we adopt the following notation for the collections of the submatrices X 1 and X in (.52): S 1 = {X 1 = P 1 XP 1 B 2 XB 2 = A 2, B XB = A, X = X }, (.54) S = {X = P 2 XP 2 B 2 XB 2 = A 2, B XB = A, X = X }. (.55) The maximum and minimum ranks and inertias of the submatrices X 1 and X in (.52) can easily be derived from Theorem.. The details are omitted. If each of the triple matrix equations in (1.6) is not consistent, people may alternatively seek its common approximation solutions under various given optimal criteria. One of the most useful approximation solutions of BXB = A is the least-squares Hermitian solution, which is defined to be a Hermitian matrix X that minimizes the norm of the difference A BXB : A BXB 2 = tr ( A BXB )( A BXB ) = min. (.56) The normal equation corresponding to the norm minimization problem is given by B 2 B BXB B = B AB. (.57) This equation is always consistent. Concerning the common least-squares Hermitian solution of (1.6), we have the following result. Corollary.7 Let A i C mi mi H and B i C mi n be given for i = 1, 2,. Then, triple matrix equations have a common least-squares Hermitian solution, namely, there exists an X C n n H such that if and only if r r B i A ib i B i B i B j A jb j B j B j B i B i B j B j A i B i XB i = min, i = 1, 2,, (.58) A 1 B 1 B 1 B 1 B2A 2 B B2B BA B BB B 1 B2B 2 BB Proof It follows from Lemma.1, Corollary.5 and (.57). = 2r Bi B j, i j, i, j = 1, 2,, (.59) = r B 1 B 1 B + r B 1 B 2. (.6) B B Some further research problems can be proposed. For example, a challenging task is to give the closed-form for the general common solution of B 2 XB2 = A 2 and B XB = A that satisfies X > (<,, ), which is equivalent to solving the inequalities X + V F B + F B V + F B2 UF B + F B U F B2 > (<,, ). (.61) Further, it would be of interest to consider the following optimization problems: (a) Maximize and minimize the rank and inertia of the Hermitian matrix expression A 1 B 1 XB 1 subject to k 1 consistent Hermitian matrix equations ( B 2 XB 2,..., B k XB k ) = ( A 2,..., A k ), and to give necessary and sufficient condition for the set of matrix equations ( B 1 XB 1,..., B k XB k ) = ( A 1,..., A k ) to have a common Hermitian solution. (b) Maximize and minimize the rank and inertia of the Hermitian matrix expression A 1 B 1 XB 1 subject to the matrix inequality B 2 XB 2 A 2. 15
16 4 Conclusions In this paper, we studied the problems of maximizing/minimizing the rank/inertia of the constrained matrix expression in (1.2), and obtained some closed-form formulas for the extremal ranks/inertias of (1.2) by pure algebraic operations of matrices and their generalized inverses. As direct applications, we gave necessary and sufficient conditions for the existence of X satisfying the triple matrix equations in (1.6), as well as some matrix inequalities. Although the problems of maximizing/minimizing the rank/inertia are regarded as NP-hard, the results obtained in this paper and 12, 14, 15, 16, 17, 28, 29,, 2, show that closed-form formulas for the extremal ranks/inertias of some simpler matrix expressions can be derived, while these closed-form formulas can be used to solve some fundamental problems in matrix theory, as mentioned in the beginning of this paper. All the results obtained in these papers are brandnew and beyond our conventional understanding on linear matrix expressions. This series of researches show that for many basic or classic problems like solvability of matrix equations and matrix inequalities, we are still able to develop some new methods and use them to derive a variety of innovative results. References 1 E. Candes and B. Recht, Exact matrix completion via convex optimization, Found. of Comput. Math. 9, , M. Fazel, H. Hindi and S. Boyd, A Rank minimization heuristic with application to minimum order system approximation, In: Proceedings of the 21 American Control Conference, pp , 21. M. Fazel, H. Hindi and S. Boyd, Rank minimization and applications in system theory, In: Proceedings of the 24 American Control Conference pp , J.F. Geelen, Maximum rank matrix completion, Linear Algebra Appl. 288(1999), J. Groß, A note on the general Hermitian solution to AXA = B, Bull. Malays. Math. Soc. (2) 21(1998), N.J.A. Harvey, D.R. Karger and S. Yekhanin, The complexity of matrix completion, In: Proceedings of the Seventeenth Annual ACM-SIAM Symposium on Discrete Algorithm, Association for Computing Machinery, New York, pp , T.M. Hoang and T. Thierauf, The complexity of the inertia, Lecture Notes in Computer Science 2556, Springer, 22, pp T.M. Hoang and T. Thierauf, The complexity of the inertia and some closure properties of GapL, In: Proceedings of the Twentieth Annual IEEE Conference on Computational Complexity, 25, pp C.G. Khatri and S.K. Mitra, Hermitian and nonnegative definite solutions of linear matrix equations, SIAM J. Appl. Math. 1(1976), Y. Kim and M. Mesbahi, On the rank minimization problem, In: Proceedings of the 24 American Control Conference, Boston, pp , M. Laurent, Matrix completion problems, In: Encyclopedia of Optimization (C.A. Floudas and P.M. Pardalos, eds.), Vol. III, Kluwer, pp , Ying Li, F. Zhang, W. Guo and J. Zhao, Solutions with special structure to the linear matrix equation AX = B, Comput. Math. Appl. 61(211), X. Liu and J. Rong, On Hermitian nonnegative-definite solutions to matrix equations, Math. Notes 85(29) Y. Liu and Y. Tian, More on extremal ranks of the matrix expressions A BX ± X B with statistical applications, Numer. Linear Algebra Appl. 15(28), Y. Liu and Y. Tian, Extremal ranks of submatrices in an Hermitian solution to the matrix equation AXA = B with applications, J. Appl. Math. Comput. 2(21), Y. Liu and Y. Tian, A simultaneous decomposition of a matrix triplet with applications, Numer. Linear Algebra Appl. 18(211), Y. Liu and Y. Tian, Max-min problems on the ranks and inertias of the matrix expressions A BXC ± (BXC) with applications, J. Optim. Theory Appl. 148(211), Y. Liu, Y. Tian and Y. Takane, Ranks of Hermitian and skew-hermitian solutions to the matrix equation AXA = B, Linear Algebra Appl. 41(29), G. Marsaglia and G.P.H. Styan, Equalities and inequalities for ranks of matrices, Linear and Multilinear Algebra 2(1974), M. Mahajan and J. Sarma, On the complexity of matrix rank and rigidity, Lecture Notes in Computer Science Vol. 4649, Springer, pp ,
17 21 M. Mesbahi, On the rank minimization problem and its control applications, Systems & Control Letters (1998), M. Mesbahi and G.P. Papavassilopoulos, Solving a class of rank minimization problems via semi-definite programs, with applications to the fixed order output feedback synthesis, In: Proceedings of the American Control Conference, Albuquerque, New Mexico, pp. 77 8, B.K. Natarajan, Sparse approximate solutions to linear systems, SIAM J. Comput. 24(1995), A B. Özgüler and N. Akar, A common solution to a pair of linear matrix equations over a principal ideal domain, Linear Algebra Appl. 144(1991), B. Recht, M. Fazel and P.A. Parrilo, Guaranteed minimum rank solutions to linear matrix equations via nuclear norm minimization, SIAM Review 52(21), G.A.F. Seber, A Matrix Handbook for Statisticians, John Wiley & Sons, Y. Tian, Solvability of two linear matrix equations, Linear and Multilinear Algebra 48(2), Y. Tian, Equalities and inequalities for inertias of Hermitian matrices with applications, Linear Algebra Appl. 4(21), Y. Tian, Rank and inertia of submatrices of the Moore Penrose inverse of a Hermitian matrix, Electron. J. Linear Algebra 2(21), Y. Tian, Completing block Hermitian matrices with maximal and minimal ranks and inertias. Electron. J. Linear Algebra 21(21), Y. Tian, Maximization and minimization of the rank and inertia of the Hermitian matrix expression A BX (BX) with applications, Linear Algebra Appl. 44(211), Y. Tian, On additive decompositions of the Hermitian solutions of the matrix equation AXA = B, Mediter. J. Math., DOI:1.17/s Y. Tian and Y. Liu, Extremal ranks of some symmetric matrix expressions with applications, SIAM J. Matrix Anal. Appl. 28(26), J. Wang, V. Sreeram and W. Liu, The parametrization of the pencil A + BKC with constant rank and its application, Internat. J. Infom. Sys. Sci. 4(28), X. Zhang, The general common Hermitian nonnegative-definite solution to the matrix equations AXA = BB and CXC = DD with applications in statistics, J. Multivariate Anal. 9(25),
Rank and inertia optimizations of two Hermitian quadratic matrix functions subject to restrictions with applications
Rank and inertia optimizations of two Hermitian quadratic matrix functions subject to restrictions with applications Yongge Tian a, Ying Li b,c a China Economics and Management Academy, Central University
More informationOptimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications
Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications Yongge Tian China Economics and Management Academy, Central University of Finance and Economics,
More informationAnalytical formulas for calculating the extremal ranks and inertias of A + BXB when X is a fixed-rank Hermitian matrix
Analytical formulas for calculating the extremal ranks and inertias of A + BXB when X is a fixed-rank Hermitian matrix Yongge Tian CEMA, Central University of Finance and Economics, Beijing 100081, China
More informationSolutions of a constrained Hermitian matrix-valued function optimization problem with applications
Solutions of a constrained Hermitian matrix-valued function optimization problem with applications Yongge Tian CEMA, Central University of Finance and Economics, Beijing 181, China Abstract. Let f(x) =
More informationAnalytical formulas for calculating extremal ranks and inertias of quadratic matrix-valued functions and their applications
Analytical formulas for calculating extremal ranks and inertias of quadratic matrix-valued functions and their applications Yongge Tian CEMA, Central University of Finance and Economics, Beijing 100081,
More informationA revisit to a reverse-order law for generalized inverses of a matrix product and its variations
A revisit to a reverse-order law for generalized inverses of a matrix product and its variations Yongge Tian CEMA, Central University of Finance and Economics, Beijing 100081, China Abstract. For a pair
More informationOn V-orthogonal projectors associated with a semi-norm
On V-orthogonal projectors associated with a semi-norm Short Title: V-orthogonal projectors Yongge Tian a, Yoshio Takane b a School of Economics, Shanghai University of Finance and Economics, Shanghai
More informationELA
Electronic Journal of Linear Algebra ISSN 181-81 A publication of te International Linear Algebra Society ttp://mat.tecnion.ac.il/iic/ela RANK AND INERTIA OF SUBMATRICES OF THE MOORE PENROSE INVERSE OF
More informationRank and inertia of submatrices of the Moore- Penrose inverse of a Hermitian matrix
Electronic Journal of Linear Algebra Volume 2 Volume 2 (21) Article 17 21 Rank and inertia of submatrices of te Moore- Penrose inverse of a Hermitian matrix Yongge Tian yongge.tian@gmail.com Follow tis
More informationThe symmetric minimal rank solution of the matrix equation AX=B and the optimal approximation
Electronic Journal of Linear Algebra Volume 18 Volume 18 (2009 Article 23 2009 The symmetric minimal rank solution of the matrix equation AX=B and the optimal approximation Qing-feng Xiao qfxiao@hnu.cn
More informationMore on generalized inverses of partitioned matrices with Banachiewicz-Schur forms
More on generalized inverses of partitioned matrices wit anaciewicz-scur forms Yongge Tian a,, Yosio Takane b a Cina Economics and Management cademy, Central University of Finance and Economics, eijing,
More informationMultiplicative Perturbation Bounds of the Group Inverse and Oblique Projection
Filomat 30: 06, 37 375 DOI 0.98/FIL67M Published by Faculty of Sciences Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Multiplicative Perturbation Bounds of the Group
More informationRe-nnd solutions of the matrix equation AXB = C
Re-nnd solutions of the matrix equation AXB = C Dragana S. Cvetković-Ilić Abstract In this article we consider Re-nnd solutions of the equation AXB = C with respect to X, where A, B, C are given matrices.
More informationThe Hermitian R-symmetric Solutions of the Matrix Equation AXA = B
International Journal of Algebra, Vol. 6, 0, no. 9, 903-9 The Hermitian R-symmetric Solutions of the Matrix Equation AXA = B Qingfeng Xiao Department of Basic Dongguan olytechnic Dongguan 53808, China
More informationMoore Penrose inverses and commuting elements of C -algebras
Moore Penrose inverses and commuting elements of C -algebras Julio Benítez Abstract Let a be an element of a C -algebra A satisfying aa = a a, where a is the Moore Penrose inverse of a and let b A. We
More informationA new algebraic analysis to linear mixed models
A new algebraic analysis to linear mixed models Yongge Tian China Economics and Management Academy, Central University of Finance and Economics, Beijing 100081, China Abstract. This article presents a
More informationLinear Algebra and its Applications
Linear Algebra and its Applications 433 (2010) 476 482 Contents lists available at ScienceDirect Linear Algebra and its Applications journal homepage: www.elsevier.com/locate/laa Nonsingularity of the
More informationSPECIAL FORMS OF GENERALIZED INVERSES OF ROW BLOCK MATRICES YONGGE TIAN
Electronic Journal of Linear lgebra ISSN 1081-3810 publication of the International Linear lgebra Society EL SPECIL FORMS OF GENERLIZED INVERSES OF ROW BLOCK MTRICES YONGGE TIN bstract. Given a row block
More informationON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES
olume 10 2009, Issue 2, Article 41, 10 pp. ON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES HANYU LI, HU YANG, AND HUA SHAO COLLEGE OF MATHEMATICS AND PHYSICS CHONGQING UNIERSITY
More informationResearch Article Constrained Solutions of a System of Matrix Equations
Journal of Applied Mathematics Volume 2012, Article ID 471573, 19 pages doi:10.1155/2012/471573 Research Article Constrained Solutions of a System of Matrix Equations Qing-Wen Wang 1 and Juan Yu 1, 2 1
More informationON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES
ON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES HANYU LI, HU YANG College of Mathematics and Physics Chongqing University Chongqing, 400030, P.R. China EMail: lihy.hy@gmail.com,
More informationSome inequalities for sum and product of positive semide nite matrices
Linear Algebra and its Applications 293 (1999) 39±49 www.elsevier.com/locate/laa Some inequalities for sum and product of positive semide nite matrices Bo-Ying Wang a,1,2, Bo-Yan Xi a, Fuzhen Zhang b,
More informationThe reflexive re-nonnegative definite solution to a quaternion matrix equation
Electronic Journal of Linear Algebra Volume 17 Volume 17 28 Article 8 28 The reflexive re-nonnegative definite solution to a quaternion matrix equation Qing-Wen Wang wqw858@yahoo.com.cn Fei Zhang Follow
More informationMatrix Inequalities by Means of Block Matrices 1
Mathematical Inequalities & Applications, Vol. 4, No. 4, 200, pp. 48-490. Matrix Inequalities by Means of Block Matrices Fuzhen Zhang 2 Department of Math, Science and Technology Nova Southeastern University,
More informationNew insights into best linear unbiased estimation and the optimality of least-squares
Journal of Multivariate Analysis 97 (2006) 575 585 www.elsevier.com/locate/jmva New insights into best linear unbiased estimation and the optimality of least-squares Mario Faliva, Maria Grazia Zoia Istituto
More informationThe Drazin inverses of products and differences of orthogonal projections
J Math Anal Appl 335 7 64 71 wwwelseviercom/locate/jmaa The Drazin inverses of products and differences of orthogonal projections Chun Yuan Deng School of Mathematics Science, South China Normal University,
More informationOn Sums of Conjugate Secondary Range k-hermitian Matrices
Thai Journal of Mathematics Volume 10 (2012) Number 1 : 195 202 www.math.science.cmu.ac.th/thaijournal Online ISSN 1686-0209 On Sums of Conjugate Secondary Range k-hermitian Matrices S. Krishnamoorthy,
More informationGeneralized Principal Pivot Transform
Generalized Principal Pivot Transform M. Rajesh Kannan and R. B. Bapat Indian Statistical Institute New Delhi, 110016, India Abstract The generalized principal pivot transform is a generalization of the
More informationThe Skew-Symmetric Ortho-Symmetric Solutions of the Matrix Equations A XA = D
International Journal of Algebra, Vol. 5, 2011, no. 30, 1489-1504 The Skew-Symmetric Ortho-Symmetric Solutions of the Matrix Equations A XA = D D. Krishnaswamy Department of Mathematics Annamalai University
More informationDiagonal and Monomial Solutions of the Matrix Equation AXB = C
Iranian Journal of Mathematical Sciences and Informatics Vol. 9, No. 1 (2014), pp 31-42 Diagonal and Monomial Solutions of the Matrix Equation AXB = C Massoud Aman Department of Mathematics, Faculty of
More informationThe equalities of ordinary least-squares estimators and best linear unbiased estimators for the restricted linear model
The equalities of ordinary least-squares estimators and best linear unbiased estimators for the restricted linear model Yongge Tian a and Douglas P. Wiens b a School of Economics, Shanghai University of
More informationOperators with Compatible Ranges
Filomat : (7), 579 585 https://doiorg/98/fil7579d Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://wwwpmfniacrs/filomat Operators with Compatible Ranges
More informationELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n
Electronic Journal of Linear Algebra ISSN 08-380 Volume 22, pp. 52-538, May 20 THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE WEI-WEI XU, LI-XIA CAI, AND WEN LI Abstract. In this
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationarxiv: v1 [math.ra] 28 Jan 2016
The Moore-Penrose inverse in rings with involution arxiv:1601.07685v1 [math.ra] 28 Jan 2016 Sanzhang Xu and Jianlong Chen Department of Mathematics, Southeast University, Nanjing 210096, China Abstract:
More informationNonsingularity and group invertibility of linear combinations of two k-potent matrices
Nonsingularity and group invertibility of linear combinations of two k-potent matrices Julio Benítez a Xiaoji Liu b Tongping Zhu c a Departamento de Matemática Aplicada, Instituto de Matemática Multidisciplinar,
More informationMATH36001 Generalized Inverses and the SVD 2015
MATH36001 Generalized Inverses and the SVD 201 1 Generalized Inverses of Matrices A matrix has an inverse only if it is square and nonsingular. However there are theoretical and practical applications
More informationSome results on the reverse order law in rings with involution
Some results on the reverse order law in rings with involution Dijana Mosić and Dragan S. Djordjević Abstract We investigate some necessary and sufficient conditions for the hybrid reverse order law (ab)
More informationTensor Complementarity Problem and Semi-positive Tensors
DOI 10.1007/s10957-015-0800-2 Tensor Complementarity Problem and Semi-positive Tensors Yisheng Song 1 Liqun Qi 2 Received: 14 February 2015 / Accepted: 17 August 2015 Springer Science+Business Media New
More informationA note on the equality of the BLUPs for new observations under two linear models
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 14, 2010 A note on the equality of the BLUPs for new observations under two linear models Stephen J Haslett and Simo Puntanen Abstract
More informationThe DMP Inverse for Rectangular Matrices
Filomat 31:19 (2017, 6015 6019 https://doi.org/10.2298/fil1719015m Published by Faculty of Sciences Mathematics, University of Niš, Serbia Available at: http://.pmf.ni.ac.rs/filomat The DMP Inverse for
More informationRank equalities for idempotent and involutory matrices
Linear Algebra and its Applications 335 (2001) 101 117 www.elsevier.com/locate/laa Rank equalities for idempotent and involutory matrices Yongge Tian a, George P.H. Styan a, b, a Department of Mathematics
More informationSpectral inequalities and equalities involving products of matrices
Spectral inequalities and equalities involving products of matrices Chi-Kwong Li 1 Department of Mathematics, College of William & Mary, Williamsburg, Virginia 23187 (ckli@math.wm.edu) Yiu-Tung Poon Department
More informationof a Two-Operator Product 1
Applied Mathematical Sciences, Vol. 7, 2013, no. 130, 6465-6474 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.39501 Reverse Order Law for {1, 3}-Inverse of a Two-Operator Product 1 XUE
More informationMOORE-PENROSE INVERSE IN AN INDEFINITE INNER PRODUCT SPACE
J. Appl. Math. & Computing Vol. 19(2005), No. 1-2, pp. 297-310 MOORE-PENROSE INVERSE IN AN INDEFINITE INNER PRODUCT SPACE K. KAMARAJ AND K. C. SIVAKUMAR Abstract. The concept of the Moore-Penrose inverse
More informationOn equality and proportionality of ordinary least squares, weighted least squares and best linear unbiased estimators in the general linear model
Statistics & Probability Letters 76 (2006) 1265 1272 www.elsevier.com/locate/stapro On equality and proportionality of ordinary least squares, weighted least squares and best linear unbiased estimators
More informationOn the simplest expression of the perturbed Moore Penrose metric generalized inverse
Annals of the University of Bucharest (mathematical series) 4 (LXII) (2013), 433 446 On the simplest expression of the perturbed Moore Penrose metric generalized inverse Jianbing Cao and Yifeng Xue Communicated
More informationThe Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation
The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation Zheng-jian Bai Abstract In this paper, we first consider the inverse
More informationGeneralized Schur complements of matrices and compound matrices
Electronic Journal of Linear Algebra Volume 2 Volume 2 (200 Article 3 200 Generalized Schur complements of matrices and compound matrices Jianzhou Liu Rong Huang Follow this and additional wors at: http://repository.uwyo.edu/ela
More informationOn the Moore-Penrose and the Drazin inverse of two projections on Hilbert space
On the Moore-Penrose and the Drazin inverse of two projections on Hilbert space Sonja Radosavljević and Dragan SDjordjević March 13, 2012 Abstract For two given orthogonal, generalized or hypergeneralized
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationChapter 1. Matrix Algebra
ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface
More information1 Linear Algebra Problems
Linear Algebra Problems. Let A be the conjugate transpose of the complex matrix A; i.e., A = A t : A is said to be Hermitian if A = A; real symmetric if A is real and A t = A; skew-hermitian if A = A and
More informationResearch Article Some Results on Characterizations of Matrix Partial Orderings
Applied Mathematics, Article ID 408457, 6 pages http://dx.doi.org/10.1155/2014/408457 Research Article Some Results on Characterizations of Matrix Partial Orderings Hongxing Wang and Jin Xu Department
More informationRigid Geometric Transformations
Rigid Geometric Transformations Carlo Tomasi This note is a quick refresher of the geometry of rigid transformations in three-dimensional space, expressed in Cartesian coordinates. 1 Cartesian Coordinates
More informationarxiv: v1 [math.ra] 16 Nov 2016
Vanishing Pseudo Schur Complements, Reverse Order Laws, Absorption Laws and Inheritance Properties Kavita Bisht arxiv:1611.05442v1 [math.ra] 16 Nov 2016 Department of Mathematics Indian Institute of Technology
More informationMAT 2037 LINEAR ALGEBRA I web:
MAT 237 LINEAR ALGEBRA I 2625 Dokuz Eylül University, Faculty of Science, Department of Mathematics web: Instructor: Engin Mermut http://kisideuedutr/enginmermut/ HOMEWORK 2 MATRIX ALGEBRA Textbook: Linear
More informationWeaker assumptions for convergence of extended block Kaczmarz and Jacobi projection algorithms
DOI: 10.1515/auom-2017-0004 An. Şt. Univ. Ovidius Constanţa Vol. 25(1),2017, 49 60 Weaker assumptions for convergence of extended block Kaczmarz and Jacobi projection algorithms Doina Carp, Ioana Pomparău,
More informationA system of matrix equations and a linear matrix equation over arbitrary regular rings with identity
Linear Algebra and its Applications 384 2004) 43 54 www.elsevier.com/locate/laa A system of matrix equations and a linear matrix equation over arbitrary regular rings with identity Qing-Wen Wang Department
More informationOn some linear combinations of hypergeneralized projectors
Linear Algebra and its Applications 413 (2006) 264 273 www.elsevier.com/locate/laa On some linear combinations of hypergeneralized projectors Jerzy K. Baksalary a, Oskar Maria Baksalary b,, Jürgen Groß
More informationJournal of Inequalities in Pure and Applied Mathematics
Journal of Inequalities in Pure and Applied Mathematics http://jipam.vu.edu.au/ Volume 7, Issue 1, Article 34, 2006 MATRIX EQUALITIES AND INEQUALITIES INVOLVING KHATRI-RAO AND TRACY-SINGH SUMS ZEYAD AL
More informationECE 275A Homework #3 Solutions
ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =
More informationISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION
ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION Harald K. Wimmer 1 The set of all negative-semidefinite solutions of the CARE A X + XA + XBB X C C = 0 is homeomorphic
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationOn the Hermitian solutions of the
Journal of Applied Mathematics & Bioinformatics vol.1 no.2 2011 109-129 ISSN: 1792-7625 (print) 1792-8850 (online) International Scientific Press 2011 On the Hermitian solutions of the matrix equation
More informationThe Moore-Penrose inverse of differences and products of projectors in a ring with involution
The Moore-Penrose inverse of differences and products of projectors in a ring with involution Huihui ZHU [1], Jianlong CHEN [1], Pedro PATRÍCIO [2] Abstract: In this paper, we study the Moore-Penrose inverses
More informationarxiv: v1 [math.ra] 21 Jul 2013
Projections and Idempotents in -reducing Rings Involving the Moore-Penrose Inverse arxiv:1307.5528v1 [math.ra] 21 Jul 2013 Xiaoxiang Zhang, Shuangshuang Zhang, Jianlong Chen, Long Wang Department of Mathematics,
More informationELA ON A SCHUR COMPLEMENT INEQUALITY FOR THE HADAMARD PRODUCT OF CERTAIN TOTALLY NONNEGATIVE MATRICES
ON A SCHUR COMPLEMENT INEQUALITY FOR THE HADAMARD PRODUCT OF CERTAIN TOTALLY NONNEGATIVE MATRICES ZHONGPENG YANG AND XIAOXIA FENG Abstract. Under the entrywise dominance partial ordering, T.L. Markham
More informationarxiv: v1 [math.ra] 14 Apr 2018
Three it representations of the core-ep inverse Mengmeng Zhou a, Jianlong Chen b,, Tingting Li c, Dingguo Wang d arxiv:180.006v1 [math.ra] 1 Apr 018 a School of Mathematics, Southeast University, Nanjing,
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationOPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY
published in IMA Journal of Numerical Analysis (IMAJNA), Vol. 23, 1-9, 23. OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY SIEGFRIED M. RUMP Abstract. In this note we give lower
More information~ g-inverses are indeed an integral part of linear algebra and should be treated as such even at an elementary level.
Existence of Generalized Inverse: Ten Proofs and Some Remarks R B Bapat Introduction The theory of g-inverses has seen a substantial growth over the past few decades. It is an area of great theoretical
More informationSome results on matrix partial orderings and reverse order law
Electronic Journal of Linear Algebra Volume 20 Volume 20 2010 Article 19 2010 Some results on matrix partial orderings and reverse order law Julio Benitez jbenitez@mat.upv.es Xiaoji Liu Jin Zhong Follow
More informationOn Pseudo SCHUR Complements in an EP Matrix
International Journal of Scientific Innovative Mathematical Research (IJSIMR) Volume, Issue, February 15, PP 79-89 ISSN 47-7X (Print) & ISSN 47-4 (Online) wwwarcjournalsorg On Pseudo SCHUR Complements
More informationPOSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS
POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS RICHARD J. CARON, HUIMING SONG, AND TIM TRAYNOR Abstract. Let A and E be real symmetric matrices. In this paper we are concerned with the determination
More informationarxiv: v1 [math.ra] 24 Aug 2016
Characterizations and representations of core and dual core inverses arxiv:1608.06779v1 [math.ra] 24 Aug 2016 Jianlong Chen [1], Huihui Zhu [1,2], Pedro Patrício [2,3], Yulin Zhang [2,3] Abstract: In this
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More informationMATRIX COMPLETION AND TENSOR RANK
MATRIX COMPLETION AND TENSOR RANK HARM DERKSEN Abstract. In this paper, we show that the low rank matrix completion problem can be reduced to the problem of finding the rank of a certain tensor. arxiv:1302.2639v2
More informationYimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract
Linear Algebra and its Applications 49 (006) 765 77 wwwelseviercom/locate/laa Relative perturbation bounds for the eigenvalues of diagonalizable and singular matrices Application of perturbation theory
More informationAn iterative method for the skew-symmetric solution and the optimal approximate solution of the matrix equation AXB =C
Journal of Computational and Applied Mathematics 1 008) 31 44 www.elsevier.com/locate/cam An iterative method for the skew-symmetric solution and the optimal approximate solution of the matrix equation
More informationResearch Article On the Hermitian R-Conjugate Solution of a System of Matrix Equations
Applied Mathematics Volume 01, Article ID 398085, 14 pages doi:10.1155/01/398085 Research Article On the Hermitian R-Conjugate Solution of a System of Matrix Equations Chang-Zhou Dong, 1 Qing-Wen Wang,
More informationGroup inverse for the block matrix with two identical subblocks over skew fields
Electronic Journal of Linear Algebra Volume 21 Volume 21 2010 Article 7 2010 Group inverse for the block matrix with two identical subblocks over skew fields Jiemei Zhao Changjiang Bu Follow this and additional
More informationTHE INVERSE PROBLEM OF CENTROSYMMETRIC MATRICES WITH A SUBMATRIX CONSTRAINT 1) 1. Introduction
Journal of Computational Mathematics, Vol22, No4, 2004, 535 544 THE INVERSE PROBLEM OF CENTROSYMMETRIC MATRICES WITH A SUBMATRIX CONSTRAINT 1 Zhen-yun Peng Department of Mathematics, Hunan University of
More informationOn the solvability of an equation involving the Smarandache function and Euler function
Scientia Magna Vol. 008), No., 9-33 On the solvability of an equation involving the Smarandache function and Euler function Weiguo Duan and Yanrong Xue Department of Mathematics, Northwest University,
More informationOn Euclidean distance matrices
On Euclidean distance matrices R. Balaji and R. B. Bapat Indian Statistical Institute, New Delhi, 110016 November 19, 2006 Abstract If A is a real symmetric matrix and P is an orthogonal projection onto
More informationSubset selection for matrices
Linear Algebra its Applications 422 (2007) 349 359 www.elsevier.com/locate/laa Subset selection for matrices F.R. de Hoog a, R.M.M. Mattheij b, a CSIRO Mathematical Information Sciences, P.O. ox 664, Canberra,
More informationRigid Geometric Transformations
Rigid Geometric Transformations Carlo Tomasi This note is a quick refresher of the geometry of rigid transformations in three-dimensional space, expressed in Cartesian coordinates. 1 Cartesian Coordinates
More informationarxiv: v4 [math.oc] 26 May 2009
Characterization of the oblique projector U(VU) V with application to constrained least squares Aleš Černý Cass Business School, City University London arxiv:0809.4500v4 [math.oc] 26 May 2009 Abstract
More informationTrace inequalities for positive semidefinite matrices with centrosymmetric structure
Zhao et al Journal of Inequalities pplications 1, 1:6 http://wwwjournalofinequalitiesapplicationscom/content/1/1/6 RESERCH Trace inequalities for positive semidefinite matrices with centrosymmetric structure
More informationExact Low-rank Matrix Recovery via Nonconvex M p -Minimization
Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:
More informationCOLUMN RANKS AND THEIR PRESERVERS OF GENERAL BOOLEAN MATRICES
J. Korean Math. Soc. 32 (995), No. 3, pp. 53 540 COLUMN RANKS AND THEIR PRESERVERS OF GENERAL BOOLEAN MATRICES SEOK-ZUN SONG AND SANG -GU LEE ABSTRACT. We show the extent of the difference between semiring
More informationChapter Two Elements of Linear Algebra
Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to
More informationMIT Algebraic techniques and semidefinite optimization February 14, Lecture 3
MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications
More informationOn a matrix result in comparison of linear experiments
Linear Algebra and its Applications 32 (2000) 32 325 www.elsevier.com/locate/laa On a matrix result in comparison of linear experiments Czesław Stȩpniak, Institute of Mathematics, Pedagogical University
More informationarxiv: v1 [cs.na] 7 Jul 2017
AE regularity of interval matrices Milan Hladík October 4, 2018 arxiv:1707.02102v1 [cs.na] 7 Jul 2017 Abstract Consider a linear system of equations with interval coefficients, and each interval coefficient
More informationELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES
Volume 22, pp. 480-489, May 20 THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES XUZHOU CHEN AND JUN JI Abstract. In this paper, we study the Moore-Penrose inverse
More informationIntrinsic products and factorizations of matrices
Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 5 3 www.elsevier.com/locate/laa Intrinsic products and factorizations of matrices Miroslav Fiedler Academy of Sciences
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationThe generalized Schur complement in group inverses and (k + 1)-potent matrices
The generalized Schur complement in group inverses and (k + 1)-potent matrices Julio Benítez Néstor Thome Abstract In this paper, two facts related to the generalized Schur complement are studied. The
More informationRobust Farkas Lemma for Uncertain Linear Systems with Applications
Robust Farkas Lemma for Uncertain Linear Systems with Applications V. Jeyakumar and G. Li Revised Version: July 8, 2010 Abstract We present a robust Farkas lemma, which provides a new generalization of
More information