The skew-symmetric orthogonal solutions of the matrix equation AX = B

Similar documents
Research Article Constrained Solutions of a System of Matrix Equations

The reflexive and anti-reflexive solutions of the matrix equation A H XB =C

The Hermitian R-symmetric Solutions of the Matrix Equation AXA = B

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation

The Skew-Symmetric Ortho-Symmetric Solutions of the Matrix Equations A XA = D

THE INVERSE PROBLEM OF CENTROSYMMETRIC MATRICES WITH A SUBMATRIX CONSTRAINT 1) 1. Introduction

Elementary linear algebra

An iterative method for the skew-symmetric solution and the optimal approximate solution of the matrix equation AXB =C

AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES

Absolute value equations

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

The symmetric minimal rank solution of the matrix equation AX=B and the optimal approximation

Linear Algebra Massoud Malek

Intrinsic products and factorizations of matrices

Subset selection for matrices

Chapter 1. Matrix Algebra

ELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n

Linear Algebra Review. Vectors

1 Linear Algebra Problems

arxiv: v1 [math.na] 1 Sep 2018

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Background Mathematics (2/2) 1. David Barber

Review of Linear Algebra

Foundations of Matrix Analysis

1 Singular Value Decomposition and Principal Component

14 Singular Value Decomposition

Linear Algebra: Matrix Eigenvalue Problems

ON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

SCHUR IDEALS AND HOMOMORPHISMS OF THE SEMIDEFINITE CONE

Lecture 7: Positive Semidefinite Matrices

AMS526: Numerical Analysis I (Numerical Linear Algebra)

MATH36001 Generalized Inverses and the SVD 2015

Lecture 1 Review: Linear models have the form (in matrix notation) Y = Xβ + ε,

Inverse Eigenvalue Problems and Their Associated Approximation Problems for Matrices with J-(Skew) Centrosymmetry

On the solvability of an equation involving the Smarandache function and Euler function

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Linear Algebra Lecture Notes-II

Phys 201. Matrices and Determinants

Lecture Summaries for Linear Algebra M51A

Improved Newton s method with exact line searches to solve quadratic matrix equation

Simultaneous Diagonalization of Positive Semi-definite Matrices

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Some inequalities for sum and product of positive semide nite matrices

Positive definite preserving linear transformations on symmetric matrix spaces

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

ON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Knowledge Discovery and Data Mining 1 (VO) ( )

Interval solutions for interval algebraic equations

Linear Algebra and its Applications

POSITIVE MAP AS DIFFERENCE OF TWO COMPLETELY POSITIVE OR SUPER-POSITIVE MAPS

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Miscellaneous Results, Solving Equations, and Generalized Inverses. opyright c 2012 Dan Nettleton (Iowa State University) Statistics / 51

MATH 425-Spring 2010 HOMEWORK ASSIGNMENTS

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear Algebra 2 Spectral Notes

Is there a Small Skew Cayley Transform with Zero Diagonal?

NORMS ON SPACE OF MATRICES

The minimum rank of matrices and the equivalence class graph

The Singular Value Decomposition

A property of orthogonal projectors

A property concerning the Hadamard powers of inverse M-matrices

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

arxiv: v3 [math.ra] 22 Aug 2014

Re-nnd solutions of the matrix equation AXB = C

This can be accomplished by left matrix multiplication as follows: I

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

ECE 275A Homework #3 Solutions

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Multiplicative Perturbation Bounds of the Group Inverse and Oblique Projection

Eigenvalues and Eigenvectors

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

CLASSIFICATION OF COMPLETELY POSITIVE MAPS 1. INTRODUCTION

Scientific Computing: Dense Linear Systems

Quantum Computing Lecture 2. Review of Linear Algebra

Moore Penrose inverses and commuting elements of C -algebras

Lecture notes: Applied linear algebra Part 1. Version 2

Equivalence constants for certain matrix norms II

Generalized Principal Pivot Transform

STAT200C: Review of Linear Algebra

Math Matrix Algebra

On Euclidean distance matrices

Review problems for MA 54, Fall 2004.

EE731 Lecture Notes: Matrix Computations for Signal Processing

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

CS 143 Linear Algebra Review

ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES

MAT 2037 LINEAR ALGEBRA I web:

Sums of diagonalizable matrices

Analytical formulas for calculating extremal ranks and inertias of quadratic matrix-valued functions and their applications

Sub-Stiefel Procrustes problem. Krystyna Ziętak

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

On the Eigenstructure of Hermitian Toeplitz Matrices with Prescribed Eigenpairs

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

The Laplacian spectrum of a mixed graph

Transcription:

Linear Algebra and its Applications 402 (2005) 303 318 www.elsevier.com/locate/laa The skew-symmetric orthogonal solutions of the matrix equation AX = B Chunjun Meng, Xiyan Hu, Lei Zhang College of Mathematics and Econometrics, Hunan University, Changsha 410082, PR China Received 15 January 2004; accepted 18 January 2005 Available online 16 March 2005 Submitted by R. Byers Abstract An n n real matrix X is said to be a skew-symmetric orthogonal matrix if X T = X and X T X = I. Using the special form of the C S decomposition of an orthogonal matrix with skew-symmetric k k leading principal submatrix, this paper establishes the necessary and sufficient conditions for the existence of and the expressions for the skew-symmetric orthogonal solutions of the matrix equation AX = B. In addition, in corresponding solution set of the equation, the explicit expression of the nearest matrix to a given matrix in the Frobenius norm have been provided. Furthermore, the Procrustes problem of skew-symmetric orthogonal matrices is considered and the formula solutions are provided. Finally an algorithm is proposed for solving the first and third problems. Numerical experiments show that it is feasible. 2005 Elsevier Inc. All rights reserved. AMS classification: 65F15; 65F20 Keywords: Skew-symmetric orthogonal matrix; Leading principal submatrix; C S decomposition; The matrix nearness problem; The least-square solutions The work was supported by the National Natural Science Foundation of China (10171031) and by Science Fund of Hunan University (521101767). Corresponding author. E-mail address: chunjunm@hnu.cn (C. Meng). 0024-3795/$ - see front matter 2005 Elsevier Inc. All rights reserved. doi:10.1016/j.laa.2005.01.022

304 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 1. Introduction We first introduce some notations to be used. Let R m n denote the set of real m n matrices, and OR n n, SR n n, ASR n n denote the sets of orthogonal n n matrices, real n n symmetric matrices, real n n skew-symmetric matrices respectively, I n the identity matrix of order n. A T means the transpose of the real matrix A. We define the inner product in space R m n, (A, B) = trace(b T A) = mi=1 nj=1 a ij b ij A,B R n m.thenr m n is a Hilbert inner product space. The norm of a matrix generated by the inner product is Frobenius norm, denoted by. Skew-symmetric orthogonal matrices play an important role in numerical analysis and numerical solution of partial differential equation. For example, the symplectic transformation can preserve the Hamiltonian structure, ( and an) matrix S is symplectic matrix if and only if SJS T 0 In = J,whereJ =. There the I n 0 matrix J is a skew-symmetric orthogonal matrix. It is pointed out that the order of a skew-symmetric orthogonal matrix is necessarily even number. So we denote the set of skew-symmetric orthogonal matrices by SSOR 2m 2m. As another example for skew-symmetric orthogonal matrices, we can see that if P OR m m, then the 0 P matrix P T SSOR 0 2m 2m. In this paper we consider the skew-symmetric orthogonal solutions of the matrix equation AX = B, (1) where A and B are given matrices in R n 2m. This problem has important geometry meanings. If X SSOR 2m 2m satisfies Eq. (1), then there exists an orthogonal transformation with skew-symmetric structure that can transform the row space of the matrix A onto that of B. And an orthogonal transformation is widely applied in numerical analysis because of its numerical stability, hence the problem we will discuss here is useful both theoretically and practically. We also consider the matrix nearness problem min X X, (2) X S X where X is a given matrix in R 2m 2m and S X is the solution set of Eq. (1). In Section 3 we will prove that problem (1) is solvable whenever the given matrices A and B satisfy two equations. However, the data matrices A and B are often perturbed by lots of means, such as observing error, model error, rounding error etc. In many cases the changed A and B do not meet the solvability conditions, which makes problem (1) have no solution. Hence we consider the least-square solutions of problem (1). That means to find a X SSOR 2m 2m to minimize the distance between AX and B,i.e. min AX B. (3) X SSOR2m 2m

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 305 The well-known equation (1) with the unknown matrix X being symmetric, skewsymmetric, symmetric positive semidefinite, bisymmetric and orthogonal were studied (see, for instance, [1 8]). All in this paper, the necessary and sufficient conditions for the existence of and the expressions for the solution of the equation were given by using the structure properties of matrices in required subset and the singular value decomposition of the matrix. But the skew-symmetric orthogonal solutions of Eq. (1) have not been considered yet, which motivates us to study the problem. Another motivation to solve the skew-symmetric orthogonal solutions of Eq. (1) is that it will be probably helpful for checking if the matrix A is generalized Hamiltonian matrix. The matrix A is called to be generalized Hamiltonian iff there exists a skew-symmetric orthogonal matrix Q such that AQ is symmetric. If the skew-symmetric orthogonal matrix X solves Eq. (1) whereb is symmetric matrix then we can claim that the matrix A is generalized Hamiltonian. To see the applications of generalized Hamiltonian matrix, we refer the readers to the paper [9]. The matrix nearness problem (2), that is, finding the nearest matrix in the solution setofeq.(1) to a given matrix in Frobenius norm, was initially proposed in the process of test or recovery of linear systems due to incomplete dates or revising given dates. A preliminary estimate X of the unknown matrix X can be obtained by the experimental observation values and the information of statistical distribution. The other form of the matrix nearness problem (2), min A A A subject to AX = B, where X and B are given matrices, A is given matrix and A is a symmetric, bisymmetric or symmetric positive semidefinite matrix were discussed (see [10,11] and references therein). Problem (3) is a special case of the Procrustes problem. Schonemann studied the orthogonal Procrustes problem in the paper [12] and obtained a explicit solution. The symmetric Procrustes problem was researched by Higham in the paper [13], where the general solutions of the problem and the algorithm to compute it are presented. Problem (3) we will discuss here may be called the skew-symmetric orthogonal Procrustes problem and be the extension of the research of Procrustes problem. The paper is organized as follows: in Section 2, the special form of C S decomposition of an orthogonal matrix with a skew-symmetric k k leading principal submatrix will be introduced. In Section 3, the necessary and sufficient conditions for the existence of and the expressions for the skew-symmetric orthogonal solutions of Eq. (1) will be derived. Section 4 will provide the expressions of the solution of the matrix nearness problem (2). In Section 5, the formula of solutions of problem (3)

306 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 is presented and in Section 6 some examples are given to illustrate the main results obtained in the paper. 2. The C S decomposition of an orthogonal matrix with a skew-symmetric k k leading principal submatrix Let X OR 2m 2m, partition X as X11 X X = 12, (4) X 21 X 22 where X11 T = X 11,i.e.X 11 ASR k k. Using the skew-symmetric structure of k k leading principal submatrix of the orthogonal matrix X, we prove that the C S decomposition of X has special form described as the following theorem. Theorem 1. If X OR 2m 2m, and partitioned as (4), X 11 ASR k k, then the C S decomposition of X can be expressed as T D1 0 X11 X 12 D1 0 Σ11 Σ = 12, (5) 0 D 2 X 21 X 22 0 R 2 Σ 21 Σ 22 here, D 1 OR k k,d 2,R 2 OR (2m k) (2m k), I 0 Σ 11 = 0 C 0, Σ 12 = 0 S 0, 0 I 0 I 0 0 Σ 21 = 0 S 0, Σ 22 = 0 C T 0, I 0 where I = diag( I 1,..., I r ), I 1 = = I 0 1 r =, C = diag(c 1 0 1,...,C l ), 0 ci si 0 C i =,i= 1,...,l. S = diag(s c i 0 1,...,S l ), S i =,S 0 s i TS i+ Ci TC i = i I 2,i= 1,...,l. 2r + 2l = rank(x 11 ), Σ 21 = Σ T 12,k 2r = rank(x 21). Proof. Suppose X OR 2m 2m, and partitioned as (4), X 11 ASR k k.asx 11 is skew-symmetric, and the k k leading principle submatrix of the orthogonal matrix X, thenifλ is one of eigenvalue of X 11,wemusthavethatλ is a pure imaginary number and the absolute value of λ, denoted by λ, is not larger than 1, i.e. λ 1. Assume the spectral decomposition of the real skew-symmetric matrix X 11 is written as I D1 T X 11D 1 = Σ 11 = 0 C 0, (6) 0

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 307 where D 1 OR k k, I = diag( I 1,..., I r ), I 1 = = I 0 1 r =. C = 1 0 diag(c 1,...,C l ), C i = rank(x 11 ). ( 0 ci c i 0 ), 0 < c i < 1, i = 1,...,l and 2(r + l) = T D1 0 X11 Consider the matrix D 0 I X 1, it remains column orthogonal. Let 21 C X 21 D 1 = (W 1,W 2 ), W 1 R (2m k) 2r, W 2 R (2m k) (k 2r) 0, Ĉ = R (k 2r) (k 2r),then Therefore, T D1 0 X11 D 0 I X 1 = 21 D T 1 X 11 D 1 = X 21 D 1 I 0 0 Ĉ W 1 W 2. (7) W 1 = 0, W2 T W 2 = I k 2r Ĉ T Ĉ. (8) Write s i = 1 ci 2, i = r + 1,...,r + l. s 2r+2l+1 = =s k = 1, and let then ( 1 Z 2 = W 2 diag, s r+1 Z T 2 Z 2 = I. 1 s r+1,..., 1 s r+l, 1 s r+l,..., 1 s k ), (9) And we see that Z 2 R (2m k) (k 2r) has orthonormal columns, hence the matrix Z 2 can be expanded into an orthogonal matrix D 2, i.e. D 2 = (Z 1,Z 2 ) OR (2m k) (2m k). From (8) and(9), we obtain Z T 1 W 2 = Z T 1 Z 2 diag(s r+1,...,s k ) = 0 (10) and ( 1 Z2 T W 2 = diag,..., 1 ) W T S 0 2 s r+1 s W 2 =. (11) k 0 I here, S = diag(s r+1,s r+1,...,s r+l,s r+l ), s i = 1 ci 2, i = r + 1,...,r + l. Therefore, we get T Z T D 2 X21 D 1 = 1 (0 ) 0 W2 = Z2 T 0 S 0. (12) I

308 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 Applying the same method, we can also find an orthogonal matrix R 2 OR (2m k) (2m k) such that 0 D1 T X 12R 2 = 0 S 0. (13) I Now we consider the matrix D T 1 0 X11 X 12 D1 0 0 D 2 T, X 21 X 22 0 R 2 it is still orthogonal matrix, and we have, from (6), (12)and(13), that I 0 0 C S 0 D1 0 X11 X 12 D1 0 = 0 I 0 D 2 X 21 X 22 0 R 2 0 X 44 X 45 X 46. 0 S 0 X 54 X 55 X 56 I X 64 X 65 X 66 (14) It can deduce that X 64 = X 65 = X 66 = X 56 = X 46 = 0. And the CS T + SX55 T = 0 implies that X 55 = (S 1 CS T ) T = C T, X 45 = X 54 = 0, X 44 OR 2r 2r. (15) Let D 2 = D 2 diag(x 44,I), the expression (5) can be obtained. Thus the proof is completed. 3. The solvable conditions for problem (1) At first, we assume that there exists a matrix X SSOR 2m 2m such that Eq. (1) holds, then we get and BB T = AXX T A T = AA T (16) AB T = AX T A T = AXA T = BA T. (17) Therefore Eqs. (16) and(17) are necessary for the solvability of problem (1), we will prove in the following paper that they is also sufficient for the solvability.

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 309 Theorem 2. Given A, B R n 2m, there is a X SSOR 2m 2m such that Eq. (1) holds if and only if (a) BB T = AA T ; (b) AB T = BA T. When these conditions are satisfied, the solutions of problem (1) can be expressed as I 0 X = Ṽ Q T, (18) 0 G where Ṽ, Q T OR 2m 2m have relation to given matrices A and B. G SSOR 2r 2r is arbitrary. In order to prove the sufficiency and obtain the expression of solutions, we introduce some important lemmas. Lemma 1. Given A, B R n 2m, if A, B satisfy the condition (a) of theorem 2, then the singular value decomposition of A, B can be described as Σ 0 A = U V T Σ 0, B = U Q T, (19) where U = U 1 U 2 OR n n,v=(v 1 V 2 ), Q = (Q 1 Q 2 ) OR 2m 2m,U 1 R n k,p 1,Q 1 R 2m k,k= rank(a), and Σ R k k > 0 is diagonal, its diagonal elements are nonzero singular value of A or B. Proof. Suppose the singular value decomposition of A be written as the fore part of (19), from condition (a), we have BB T Σ 2 0 = U U T. It implies that B T U 2 = 0, Σ 1 U1 T BBT U 1 Σ 1 = I k. (20) Hence we may find Q 2 such that Q = (B T U 1 Σ 1 Q 2 ) OR 2m 2m.Using(20), we obtain that U T Σ 0 BQ =, that ends our proof. Lemma 2. Given A, B R n 2m, there exists an orthogonal matrix X such that Eq. (1) holds if and only if (a) BB T = AA T.

310 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 Let the singular value decomposition of A, B be (19), then the orthogonal solutions of Eq. (1) can be described as Ik 0 X = V Q T, (21) 0 P where P OR (2m k) (2m k) is arbitrary. Proof. Substitute (19) into Eq. (1), we get Σ 0 V T Σ 0 XQ =. (22) Write V T XQ = X = Using (22) and(23), it can be deduced that Ik 0 X =. 0 P Therefore we have proved the lemma. X 11 X 12, X 11 R X 21 X k k. (23) 22 Proof of Theorem 2. Since the given matrices A and B satisfy the condition (a), we can assume, from the Lemma 1, that the singular value decomposition of A, B be (19). Morever A and B satisfy the condition (b), then Eq. (1) has orthogonal solutions with the form (21). Substitute (19) into the condition (b), we get V1 T Q 1 = Q T 1 V 1. (24) Note that X T = X is equivalent to V T I 0 I 0 Q = 0 P 0 P T Q T V. Comparing the corresponding blocks, with V = ( V 1 ) ( V 2, Q = Q1 ) Q 2,we have Q T 1 V 1 = V1 T Q 1, (25) V T 1 Q 2P = Q T 1 V 2, (26) V2 T Q 2P = P T Q T 2 V 2. (27) And (25) already stands from (24). Then we intend to find the common orthogonal solutions of (26) and(27). Firstly, consider (26), we observe that ( Q T 1 V 2)( Q T 1 V 2) T = V1 T Q 2(V1 T Q 2) T. Then we know, from the Lemma, that Eq. (26) has orthogonal solutions P.

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 311 Consider the orthogonal matrix V T Q,thep p leading principal submatrix of which is skew-symmetric from (24). Applying Theorem 1, weget Q T 1 V 2 = D 1 Σ 12 D T 2, VT 1 Q 2 = D 1 Σ 12 R T 2. (28) And the orthogonal solutions P of (26) can be expressed as G P = R 2 0 I 0 D2 T, (29) I where, G OR 2r 2r is arbitrary. Since V T 2 Q 2 = D 2 Σ 22 R T 2. (30) Using (27), (29) and(30), we obtain that G T = G, (31) that means G SSOR 2r 2r,let ˆQ = Q diag(i, R 2 ), ˆV = U diag(i, D 2 ),weget the skew-symmetric orthogonal solutions of Eq. (1) can be described as I 0 0 X = ˆV 0 G 0 ˆQ. I I 0 0 Let E be the permutation matrix and E = I, partitioned with accordance 0 I 0 I to the 0 G 0, write Ṽ = ˆVE, Q = ˆQE. Therefore, the solutions of I problem (1)be(18), so the proof is finished. 4. The solutions of problem (2) We first consider the following matrix nearness problem. Given C R 2m 2m, seek a matrix P SSOR 2m 2m such that C P =min. (32) For 2 is a continuous function, and the subset SSOR 2m 2m is closed and bounded in R 2m 2m, the problem (32) must have solutions. But the set SSOR 2m 2m is not convex set, then problem (32) has possibly not only one solution.

312 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 Next we will derive the expression of solutions for problem (32). From the properties of Frobenius norm, one know that C P 2 = (C + C T )/2 2 + (C C T )/2 P 2. Hence C P =min is equivalent to (C C T )/2 P 2 = min. Let the spectral decomposition of the skew-symmetric matrix Ĉ = (C C T )/2 be Λ 0 Ĉ = D D T, (33) 0 λi here Λ = diag(λ 1,...,Λ k ), Λ i =,λ λ i 0 i > 0, i = 1,...,k,2k = rank(ĉ), D T D = DD T = I 2m. With the help of (33) and the definition of Frobenius norm, we have ( ) (C C T )/2 P 2 = 2m + Ĉ 2 Λ 0 2 trace D T PD. As D T PDis still skew-symmetric orthogonal matrix, we obtain that the solutions of problem (32) can be expressed as I 0 E 0 (Ĉ) = D D T, (34) 0 H where I = diag( I 1,..., I k ), is arbitrary. I 1 = = I k = 0 1, H SSOR 1 0 (2m 2k) (2m 2k) Theorem 3. Suppose the spectral decomposition of the skew-symmetric matrix Ĉ = (C C T )/2 be (33), then the solution set of problem (32) can be written as (34). Because the solution set of problem (1) with expression (18) is closed and bounded but not convex, the solution of (2) exists and is not unique if the conditions of Theorem 1 are satisfied. Given X R 2m 2m,using(18), we get X X 2 = I 0 Ṽ T X 2 Q 0 G. Write Ṽ = Ṽ 1 Ṽ 2 ), Q = Q 1 Q 2, partitioned according to the matrix I 0, then we observe that X X 0 G =min is equivalent to min G SSOR 2r 2r G Ṽ T 2 X Q 2 2.

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 313 Let C = Ṽ T 2 X Q 2 and Ĉ = (C C T )/2, from Theorem 3, the solution set of can be expressed as (34). Theorem 4. Given X R 2m 2m, and A, B R n 2m, satisfying the conditions of Theorem 1, the solution set of problem (1) be (18), write Ṽ = Ṽ 1 Ṽ 2, Q = I 0 Q 1 Q 2, partitioned according to the matrix, and let C = Ṽ 0 G 2 TX Q 2, Ĉ = (C C T )/2, then if and only if I 0 X = Ṽ Q T, G E 0 G 0 (Ĉ), the minimum of (2) can be reached. 5. The solutions of problem (3) Consider (3), from the definition of Frobenius norm and X SSOR 2m 2m,itis obvious to see that AX B 2 = A 2 + B 2 2 trace(b T AX). (35) Furthermore, 2 trace(b T AX) = trace((b T A A T B)X). Denote the skew-symmetric matrix (B T A A T B) by Ĉ. Assume the spectral decomposition of Ĉ is the form with (33), then we know that 2 trace(b T AX) 2 k i=1 λ i and the equality holds if and only if X E 0 (Ĉ). Therefore we obtain the following theorem. Theorem 5. Given A, B R n 2m, denote (A T B B T A) by Ĉ. Suppose the spectral decomposition of Ĉ be (33), then the solution set of problem (3) can be described as (34). 6. Numerical algorithm and several examples The following algorithm can be used to solve problem (1) and problem (3).

314 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 Algorithm Step 1. Input A, B R n 2m, compute the matrix Ĉ = B T A A T B. Step 2. Compute the spectral decomposition of Ĉ with the form (33). Step 3. Compute the solutions of problem (1) orproblem(3) according to (34). Example 1. Suppose A, B R 6 6 and 0.4326 0 0.5883 0.0956 0 0.3999 1.6656 0 2.1832 0.8323.6900 A = 0.1253 0 0.1364 0.2944.8156 0.2877.1139 1.3362.7119, 1.1465 0 1.0668 0.7143 0 1.2902 1.1909.0593 1.6236.6686 0.4326 1.1909 0.1867.2944 0.8580 1.6656 1.1892 0.7258 0 1.3362 1.2540 B = 0.1253 0.0376 0.5883.7143 4.5937 0.2877 2.3273 2.1832 0 1.6236 1.4410. 1.1465 1.1746 8.1364 0 0.6918 10.5711 0.9996 0.3075 3.5983 0 3.3160 1.4605 We can easily see that the given A and B do not satisfy the solvable conditions of problem (1). Hence we tryto seek an orthogonalskew-symmetric matrixtominimize AX B F. and Solution: Applying the algorithm, we get the following the result:.2438 0.5722 0.0414 0.6712 0.4011 0.2438 0 0.4495 0.700.1406 0.4783 X = 0.5722 0.4495.1823 0.2205 0.6234 0.0414 0.7000 0.1823.6059 0.3284 0.6712 0.1406 0.2205 0.6059 0 0.3375 0.4011 0.4783 0.6234 0.3284 0.3375 0 min AX B =13.8180, X SSOR6 6 e1 = X T X I 6 =3.0778 10 15, e2 = X T + X =2.3572 10 16. Remark 1 (1) Problem (3) has a unique solution if and only if the matrix (B T A A T B) is nonsingular matrix. Example 2 is just the case.

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 315 (2) When we use the stable algorithms to do the steps 1 3, we find, by computing, that if the perturbations A, B are small, the solution perturbation of X generally remains small. We can also apply the algorithm to solve problem (1). The following example is in the case. We will compare the solutions computed by our algorithm with that computed by MATLAB procedure X = A \ B. Example 2. Let U OR 10 10,V OR 6 6,and 0.1596 0.0509 0.2463 0.2674 0.7374 0.5076 0.1521 0.0878 0.0164 0.0856 0.6145 0.3328 0.1347 0.2243 0.1530 0.1895 0.1963 0.4945 0.2287 0.2200 0.0462 0.2204 0.3013 0.3451 0.3426 0.2862 0.4306 0.0963 0.4878 0.3278 0.1061 0.7877 0.2406 0.1628 0.1581 0.3124 0.0491 0.2623 0.237.1840 U = 0.423.0038 0.0706 0.5678 0.1016 0.0638 0.0542 0.6484 0.2311 0.0503, 0.4394 0.0062 0.019.3128 0.1905 0.5435 0.1599 0.1469 0.5251 0.2326 0.4388 0.3427 0.0059 0.3729 0.1518 0.2029 0.3649 0.2417 0.5415 0.0444 0.0139 0.0232 0.5788 0.3033 0.1411 0.1630 0.6060 0.3769 0.128.0120 0.1208 0.0482 0.6009 0.2822 0.448.3518 0.3537 0.1649 0.0637 0.2482 0.0644 0.3118 0.2666 0.0418 0.016.1898 0.3107 0.0056 0.1023 0.8261 0.429.0694 0.8423 0.1112 0.099.2820 0.7955 0.3643 0.3107 0.1399 0.2577 0.2280 V = 0.3563 0.4967 0.3053 0.6472 0.3212 0.1053 0.1873 0.1510 0.0143 0.4317 0.8507 0.1785. 0.1182 0.5765 0.2271 0.5446 0.2944 0.4677 0.0845 0.5105 0.2215 0.2572 0.1012 0.7790 S1 Assume S 1 = diag(2, 1,e/1,e/2,e/3,e/4), S = with e = 10,e = 1,e = 0 4 6 10 1,e = 10 2,...,e = 10 13, A = USV T. Let X SSOR 6 6,.3093 0.6385 0.1968 0.3916 0.5519 0.3093 0.0000 0.1361 0.1062 0.6594 0.6632 X = 0.6385 0.1361 0.0000 0.6755 0.3349 0.0730 0.1968 0.1062 0.6755.5211 0.4713 0.3916 0.6594 0.3349 0.5211 0 0.1679 0.5519 0.6632 0.073.4713 0.1679 0.0000 and B = AX. Clearly, these given matrices A and B are consistent with a skewsymmetric orthogonal solution X. For such matrices A and B, we first seek the skewsymmetric orthogonal solution of AX = B applying our algorithm, then we compute the solution of AX = B using MATLAB procedure X = A\B. The entries in the following table are with exponent and with the mantissas which absolute value are less than one. And we only report their exponents.

316 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 The results 1 of Example 2 e e 1 e 2 e 3 e 4 ee 1 ee 2 ee 3 ee 4 10 13 10 15 10 14 10 15 10 14 10 15 10 2 10 1 10 1 10 12 10 15 10 14 10 15 10 14 10 14 10 2 10 2 10 2 10 11 10 14 10 14 10 15 10 14 10 15 10 4 10 3 10 3 10 10 10 14 10 14 10 15 10 14 10 15 10 5 10 4 10 4 10 9 10 14 10 14 10 15 10 14 10 14 10 5 10 5 10 5 10 8 10 14 10 14 10 15 10 14 10 14 10 6 10 6 10 6 10 7 10 15 10 14 10 16 10 14 10 15 10 7 10 7 10 7 10 6 10 14 10 14 10 15 10 14 10 15 10 8 10 8 10 8 10 5 10 14 10 14 10 15 10 14 10 14 10 9 10 9 10 9 10 4 10 14 10 14 10 15 10 14 10 15 10 10 10 10 10 10 10 3 10 14 10 14 10 15 10 14 10 14 10 11 10 11 10 11 10 2 10 15 10 14 10 15 10 14 10 14 10 12 10 12 10 12 10 1 10 14 10 14 10 15 10 14 10 16 10 13 10 13 10 13 1 10 15 10 15 10 16 10 15 10 15 10 14 10 14 10 14 10 10 14 10 14 10 15 10 14 10 14 10 14 10 14 10 14 here, X i s denote the solutions computed by our algorithm and X i s are the solutions by MATLAB procedure X = A\B. And e 1 = AX i B, e 3 = X i + X T i, ee 1 = A X i B, e 2 = X i X, e 4 = X T i X i + I, ee 2 = X i X, ee 3 = X i + X T i, ee 4 = X T i X i + I. Remark 2 (1) The results show that the matrices X i computed by our algorithm can solve well problem (1) in 14 correct significant digits. At the same time the MATLAB procedure X = A\B can also solve well the equation AX = B with the solutions X i. (2) When e = 10, 1,...,10 4, both Ĉ = B T A A T B and A are full rank and the solution of AX = B is unique, so X i,x i,x are very nearness. In this case we suggest that one should prefer to choose MATLAB procedure X = A\B to solve problem (1) for it is simple. But when some of eigenvalues of the matrix A become nearer and nearer to zero, we find that the solution X i s have gradually lost the skewsymmetric and orthogonality. While the solution X i computed by our algorithm still preserve the property of skew-symmetry and orthogonality well. Therefore when A

C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 317 has small eigenvalues near to zero, the MATLAB procedure X = A\B cannot settle problem (1) any longer while our algorithm can do it well. (3) It is pointed out that AX = B may be ill-conditioned linear equation or least squares problem but a well conditioned skew-symmetric-orthogonal-solutionproblem. An extreme example is A = ( 1 0 ), B = ( 0 1 ). The linear equation or least squares ( problem ) has condition. However, since the matrix Ĉ = B T A A T 0 1 B =, it is well-conditioned for skew-symmetric-orthogonalsolution-problem. 1 0 7. Conclusion In this paper, we have discussed the following three problems: Problem 1. Given matrices A, B R n 2m, find the skew-symmetric orthogonal solutions of the matrix equation AX = B. Problem 2. Given X R 2m 2m, find a matrix X S X, which denotes the solution set of problem 1, such that min X S X X X. Problem 3. Given matrices A, B R n 2m, find a skew-symmetric orthogonal matrix (SSOR 2m 2m )X such that min AX B. X SSOR2m 2m By applying the special form of the C S decomposition of an orthogonal matrix with skew-symmetric k k leading principal submatrix, we have obtained necessary and sufficient conditions for the solvability of problem 1 and the general forms of solutions for these three problems. Furthermore, we propose an algorithm to solve problems (1) and(3), which can preserve both orthogonality and skew-symmetry in the presence of rounding error, and compute the solutions of both problem (1) and problem (3). Numerical examples show that it is feasible. Acknowledgments We would like to thank the referee for his comments and suggestions that significantly improve the presentation of the work.

318 C. Meng et al. / Linear Algebra and its Applications 402 (2005) 303 318 References [1] W.J. Vetter, Vector structures and solutions of linear matrix equations, Linear Algebra Appl. 10 (1975) 181 188. [2] J.R. Magnus, H. Neudecker, The commutation matrix: some properties and applications, Ann. Statist. 7 (1979) 381 394. [3] J.R. Magnus, H. Neudecker, The elimination matrix: some lemmas and applications, SIAM J. Algebr. Discrete Methods 1 (1980) 422 449. [4] F.J.H. Don, On the symmetric solutions of a linear matrix equation, Linear Algebra Appl. 93 (1987) 1 7. [5] H. Dai, On the symmetric solutions of linear matrix equations, Linear Algebra Appl. 131 (1990) 1 7. [6] E.W.C. Heney, Introduction to Approximation Theory, McGraw-Hill, New York, 1966. [7] D. Xie, X. Hu, L. Zhang, The solvability conditions for inverse eigenvalue problem of anti-bisymmetric matrices, J. Comput. Math. (China) 20 (3) (2002) 245 256. [8] L. Zhang, D. Xie, The least-square solution of subspace revolution, J. Hunan Univ. 19 (1) (1992) 115 120. [9] Z. Zhang, X. Hu, L. Zhang, The solvability for the inverse eigenvalue problem of Hermitian generalized Hamiltonian matrices, Inverse Problems 18 (2002) 1369 1376. [10] D. Xie, X. Hu, L. Zhang, The inverse problem for bisymmetric matrices on a linear manifold, Math. Numer. Sinica 2 (2000) 129 138. [11] S.Q. Zhou, H. Dai, The Algebraic Inverse Eigenvalue Problem, Henan Science and Technology Press, Zhengzhou, China, 1991 (in Chinese). [12] P.H. Schonemann, A generalized solution of the orthogonal Procrustes problem, Psychomctrika 31 (1966) 1 10. [13] N.J. Higham, The symmetric Procrustes problem, BIT 28 (1988) 133 143. [14] G.H. Golub, C.F. Van Loan, Matrix Computation, third ed., The Johns Hopkins University Press, USA, 1996.