On the -Sylvester Equation AX ± X B = C

Similar documents
Numerical Methods I Non-Square and Sparse Linear Systems

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Contour integral solutions of Sylvester-type matrix equations

Numerical Methods - Numerical Linear Algebra

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Matrix Theory. A.Holst, V.Ufnarovski

Iterative Methods. Splitting Methods

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

JACOBI S ITERATION METHOD

Definite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra

Math 489AB Exercises for Chapter 2 Fall Section 2.3

MIT Final Exam Solutions, Spring 2017

Extra Problems: Chapter 1

Math 313 Chapter 1 Review

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

MATH 425-Spring 2010 HOMEWORK ASSIGNMENTS

Nonlinear palindromic eigenvalue problems and their numerical solution

Numerical Linear Algebra

A fast randomized algorithm for approximating an SVD of a matrix

EIGENVALUE PROBLEMS (EVP)

Review problems for MA 54, Fall 2004.

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

1 Sylvester equations

Numerical Methods in Matrix Computations

Lab 1: Iterative Methods for Solving Linear Systems

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

w T 1 w T 2. w T n 0 if i j 1 if i = j

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Introduction to Mathematical Programming

Solving large scale eigenvalue problems

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

Solvability of Linear Matrix Equations in a Symmetric Matrix Variable

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Homework 2 Foundations of Computational Math 2 Spring 2019

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

AM205: Assignment 2. i=1

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Section 8.3 Partial Fraction Decomposition

G1110 & 852G1 Numerical Linear Algebra

Orthogonal Transformations

Algebra C Numerical Linear Algebra Sample Exam Problems

MA2501 Numerical Methods Spring 2015

Lecture 4: Linear Algebra 1

Linear algebra and applications to graphs Part 1

Perturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop

3 QR factorization revisited

Foundations of Matrix Analysis

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Notes on Mathematics

CSL361 Problem set 4: Basic linear algebra

Linear Algebraic Equations

Linear Algebra Massoud Malek

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

Gaussian Elimination without/with Pivoting and Cholesky Decomposition

Index. for generalized eigenvalue problem, butterfly form, 211

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Least-Squares Systems and The QR factorization

Notes on Eigenvalues, Singular Values and QR

MATH 3511 Lecture 1. Solving Linear Systems 1

Linear Algebra 1 Exam 2 Solutions 7/14/3

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

The solution of the equation AX + X B = 0

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017

MTH 464: Computational Linear Algebra

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

S N. hochdimensionaler Lyapunov- und Sylvestergleichungen. Peter Benner. Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz

ECE133A Applied Numerical Computing Additional Lecture Notes

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 4

Numerical Methods I Eigenvalue Problems

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Exploiting off-diagonal rank structures in the solution of linear matrix equations

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Scientific Computing: Dense Linear Systems

CAAM 454/554: Stationary Iterative Methods

Maths for Signals and Systems Linear Algebra in Engineering

Cheat Sheet for MATH461

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Solving Linear Systems of Equations

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Numerical Linear Algebra

Math 577 Assignment 7

4. Linear transformations as a vector space 17

Symmetric matrices and dot products

The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Linear Least-Squares Data Fitting

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Applied Linear Algebra in Geoscience Using MATLAB

Process Model Formulation and Solution, 3E4

1 Solutions to selected problems

Consider the following example of a linear system:

Linear algebra for computational statistics

The Singular Value Decomposition

Transcription:

Department of Applied Mathematics National Chiao Tung University, Taiwan A joint work with E. K.-W. Chu and W.-W. Lin Oct., 20, 2008

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

In [Braden1998], the Lyapunov-like linear matrix equation A X + X A = B, A, X C m n (m n) with ( ) = ( ) T was considered using generalized inverses. Applications occur in Hamiltonian mechanics.

In [Braden1998], the Lyapunov-like linear matrix equation A X + X A = B, A, X C m n (m n) with ( ) = ( ) T was considered using generalized inverses. Applications occur in Hamiltonian mechanics. In this talk, we consider the -Sylvester equation AX ± X B = C, A, B, X C n n. (1.1) This includes the special cases of the T-Sylvester equation when = T and the H-Sylvester equation when = H.

Some related equations of (1.1), e.g., AXB ± X = C, AXB ± CX D = E, AX ± X A = C, AX ± YB = C, AXB ± CYD = E, AXA ± BYB = C and AXB ± (AXB) = C will also be studied.

Some related equations of (1.1), e.g., AXB ± X = C, AXB ± CX D = E, AX ± X A = C, AX ± YB = C, AXB ± CYD = E, AXA ± BYB = C and AXB ± (AXB) = C will also be studied. Our tools include the (generalized and periodic) Schur, (generalized) singular value and QR decompositions.

Some related equations of (1.1), e.g., AXB ± X = C, AXB ± CX D = E, AX ± X A = C, AX ± YB = C, AXB ± CYD = E, AXA ± BYB = C and AXB ± (AXB) = C will also be studied. Our tools include the (generalized and periodic) Schur, (generalized) singular value and QR decompositions. An interesting application, for the -Sylvester equation (1.1) AX ± X B = C, A, B, X C n n arises from the eigensolution of the palindromic linearization [Chu2007]

[ ] (λz + Z A B )x = 0, Z = C 2n 2n. C D

[ ] (λz + Z A B )x = 0, Z = C 2n 2n. C D Appying congruence, we have [ ] [ In 0 (λz + Z In X ) ] = X I n 0 I n [ λa + A λ(ax + B) + (XA + C) ] λ(xa + C) + (AX + B) λr(x ) + R(X ) with R(X ) XAX + XB + CX + D.

[ ] (λz + Z A B )x = 0, Z = C 2n 2n. C D Appying congruence, we have [ ] [ In 0 (λz + Z In X ) ] = X I n 0 I n [ λa + A λ(ax + B) + (XA + C) ] λ(xa + C) + (AX + B) λr(x ) + R(X ) with R(X ) XAX + XB + CX + D. How solve the -Riccati equation? R(X ) = 0.

If we can solve R(X ) = 0, then the palindromic linearization can then be square-rooted. We then have to solve the generalized eigenvalue problem for the pencil λ(ax + B) + (XA + C), with the reciprocal eigenvalues in λ(xa + C) + (AX + B) obtained for free.

If we can solve R(X ) = 0, then the palindromic linearization can then be square-rooted. We then have to solve the generalized eigenvalue problem for the pencil λ(ax + B) + (XA + C), with the reciprocal eigenvalues in λ(xa + C) + (AX + B) obtained for free. Solving the -Riccati equation is of course as difficult as the original eigenvalue problem of λz + Z. The usual invariance/deflating subspace approach for Riccati equations leads back to the original difficult eigenvalue problem.

If we can solve R(X ) = 0, then the palindromic linearization can then be square-rooted. We then have to solve the generalized eigenvalue problem for the pencil λ(ax + B) + (XA + C), with the reciprocal eigenvalues in λ(xa + C) + (AX + B) obtained for free. Solving the -Riccati equation is of course as difficult as the original eigenvalue problem of λz + Z. The usual invariance/deflating subspace approach for Riccati equations leads back to the original difficult eigenvalue problem. the application of Newton s method lead to the iterative process δx k+1 (AX k + B) + (X ka + C)δX k+1 = R(X k) which is a -Sylvester equation for δx k+1.

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

Recalled the -Sylvester equation (1.1) AX ± X B = C, A, B, X C n n. With the Kronecker product, (1.1) can be written as P vec(x ) = vec(c), P I A ± (B I )E, (2.1) where X (:, 1) A B [a ij B] C n2 n 2 X (:, 2), vec(x ). 1 Cn2. X (:, n)

And, E e i ej e j ei., e i = 1. 1 i,j n. Note that E vec(x ) = vec(x T ), E(A B)E = B A. The matrix operator on the left-hand-side of (2.1) is n 2 n 2.

And, E 1 i,j n e i ej e j ei., e i = 1.. Note that E vec(x ) = vec(x T ), E(A B)E = B A. The matrix operator on the left-hand-side of (2.1) is n 2 n 2. The application of Gaussian elimination and the like will be inefficient.

And, E 1 i,j n e i ej e j ei., e i = 1.. Note that E vec(x ) = vec(x T ), E(A B)E = B A. The matrix operator on the left-hand-side of (2.1) is n 2 n 2. The application of Gaussian elimination and the like will be inefficient. The approach ignores the structure of the original problem.

Another approach will be to transform (1.1) by some unitary P and Q, so that (1.1) becomes: For = T : PAQ Q T XP T ± PX T Q Q T B T P T = PCP T. (2.2) Let (Q H A H P H, Q H B H P H ) be in (upper-triangular) generalized Schur form.

Another approach will be to transform (1.1) by some unitary P and Q, so that (1.1) becomes: For = T : PAQ Q T XP T ± PX T Q Q T B T P T = PCP T. (2.2) For = H: PAQ Q H XP H ± PX H Q Q H B H P H = PCP H. (2.3) Let (Q H A H P H, Q H B H P H ) be in (upper-triangular) generalized Schur form.

The transformed equations in (2.2) and (2.3) then have the form [ a11 0 T ] [ x11 x ] [ 12 x ± 11 x ] [ 21 b 11 b ] a 21 A 22 x 21 X 22 x 12 X22 21 0 B22 (2.4) [ c11 c = 12 ]. c 21 C 22 Multiply the matrices out, we have

a 11 x 11 ± b11x 11 = c 11, (2.5) a 11 x12 ± x21b 22 = c12 x11b 21, (2.6) A 22 x 21 ± b11x 12 = c 21 x 11 a 21, (2.7) A 22 X 22 ± X22B 22 = C 22 C 22 a 21 x12 x 12 b21.(2.8)

a 11 x 11 ± b 11x 11 = c 11, (2.5) a 11 x 12 ± x 21B 22 = c 12 x 11b 21, (2.6) A 22 x 21 ± b 11x 12 = c 21 x 11 a 21, (2.7) A 22 X 22 ± X 22B 22 = C 22 C 22 a 21 x 12 x 12 b 21.(2.8) From (2.5) for = T, we have (a 11 ± b 11 )x 11 = c 11. (2.9) Let λ 1 a 11 /b 11 σ(a, B). The solvability condition of the above equation is a 11 ± b 11 0 λ 1 1. (2.10)

From (2.5) when = H, we have a 11 x 11 ± b 11 x 11 = c 11. (2.11) To solve (2.11) is to write it together with its complex conjugate in the composite form [ a 11 ±b11 ] [ ] [ ] x11 c11 ±b 11 a11 x11 = c11.

From (2.5) when = H, we have a 11 x 11 ± b 11 x 11 = c 11. (2.11) To solve (2.11) is to write it together with its complex conjugate in the composite form [ a 11 ±b11 ] [ ] [ ] x11 c11 ±b 11 a11 x11 = c11. The determinant of the matrix operator in above: d = a 11 2 b 11 2 0 λ 11 1. (2.12) requiring that no eigenvalue λ σ(a, B) lies on the unit circle.

From (2.6) and (2.7), we obtain [ a11 I ±B ] [ ] 22 x12 ±b11 I A = 22 x 21 [ c12 c 21 ] [ c12 c 21 ] + x 11 [ b21 With a 11 = b 11 = 0, x 11 will be undetermined. However, σ(a, B) = C and this case will be excluded by (2.16). If a 11 0, (2.13) is then equivalent to [ ] a [ ] [ ] [ 11 I ±B 22 x12 c12 0 A 22 b 11 a 11 B 22 x 21 = ĉ 21 The solvability condition of (2.13) and (2.14) is det Ã22 0, à 22 A 22 b 11 a11 B 22 a 21 c 12 ]. (2.13) c 21 b 11 a c 11 12 (2.14) or that λ and λ cannot be in σ(a, B) together. Note that à 22 is still lower-triangular, just like A or B. ].

If b 11 0, (2.13) is equivalent to [ ] 0 B 22 a 11 [ ] b11 A 22 x12 b11 I ±A = 22 x 21 [ ĉ12 ] [ ± c 12 a 11 b c 11 21 ± c 21 ± c 21 (2.15) with an identical solvability condition (2.16). Lastly, (2.8) is of the same form as (1.1) but of smaller size. We summarize the solvability condition for (1.1) in the following theorem: Theorem1 The -Sylvester equation (1.1): ] AX ± X B = C, A, B C n n is uniquely solvable if and only if the condition:

λ σ(a, B) λ σ(a, B) (2.16) is satisfied. Here, the convention that 0 and are mutually reciprocal is followed. The process in this subsection is summarized below: (with BS denoting back-substitution) Algorithm SSylvester (For the unique solution of AX ± X B = C; A, B, C, X C n n.) Compute the lower-triangular generalized Schur form (PAQ, PBQ) using QZ. Store (PAQ, PBQ, PCP ) in (A, B, C). Solve (2.9) for = T, or (2.11) for = H; if fail, exit.

If n = 1 or a 11 2 + b 11 2 tolerance, exit. If a 11 b 11, then if Ã22 A 22 b 11 a11 B 22 has any negligible diagonal elements, then exit. Else if B 22 B 22 a 11 b11 A 22 has any negligible diagonal elements, then exit. Else compute x 21 = (c.f. (2.15)). B 1 22 ĉ12 by BS, x 12 = (± c 21 A 22 x 21 ) /b 11 Apply Algorithm TSylvester to A 22 X 22 ± X 22 B 22 = C 22, n n 1. Output X QX P for = T, or X QXP for = H. End of algorithm

Cost Let the operation count of the Algorithm SSylvester be f (n) complex flops, we assume that A, B are lower-triangular matrices. (After a QZ procedure (66n 3 )) Inverting Ã22 A 22 b 11 a11 B 22 or B 22 B 22 a 11 b11 A 22 ( 1 2 n2 flops). 1 Computing x 21 = B 22 ĉ12 and x 12 = (± c 21 A 22 x 21 ) /b11 (n2 flops). Forming C 22 = C 22 a 21 x 12 x 12a 21 (2n2 ). Thus f (n) f (n 1) + 7 2 n2, ignoring O(n) terms. This implies that f (n) 7 6 n3 and the total operation count for Algorithm SSylvester is 67 1 6 n3 complex flops, ignoring O(n 2 ) terms.

Solvability Condition The solvability condition of AX ± X B = C is obviously identical to that of P vec(x ) = vec(c), P = I A ± (B I )E. However, E reshuffles the columns of B I, making the analysis of the matrix operator P difficult.

Solvability Condition The solvability condition of AX ± X B = C is obviously identical to that of P vec(x ) = vec(c), P = I A ± (B I )E. However, E reshuffles the columns of B I, making the analysis of the matrix operator P difficult. Consider the trivial example when n = 2, A = [a ij ] and B = [b ij ] are lower-triangle matrices, we have P = a 11 ± b 11 ±b 21 a 11 ±b 22 a 21 ±b 11 a 22 a 21 ± b 21 a 22.

The eigenvalues of the corresponding P are λ ii = a ii ± b ii (i = 1, 2) and those of the middle block W 12 where [ a W ij ii ±bjj ±bii a jj The characteristic polynomial of W ij, identical to that for W ji, is λ 2 (a ii + a jj )λ + det W ij with det W ij = a ii a jj bii b jj, and the eigenvalues are λ Wij = 1 2 ]. [ ] a ii + a jj ± (a ii a jj ) 2 + 4bii b jj. Note that some λ ii or λ Wij = 0 if and only if (2.16) in Theorem 1 is violated.

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

Example1 In MATLAB commands: Â = tril(randn(n), 1) + diag(a), ˆB = tril(randn(n), 1) + diag(b) and C = randn(n), where a, b R n. To guarantee condition (2.16), let b = randn(n, 1), a = 2b. In Table1, we list the CPU time ratios, corresponding residuals and their ratios, with increasing dimensions n = 16, 25, 30, 35, 40. Note that the operation counts for the SSA and KRP methods are approximately 67n 3 and 2 3 n6 flops respectively (the latter for the LU decomposition of the n 2 n 2 matrix in (2.1).

t KRP Table1: Results for Example1 Res(KRP) Res(ASS) n t ASS Res(ASS) Res(KRP) 16 1.00e+00 1.8527e-17 2.1490e-17 1.16 25 1.31e+01 2.3065e-17 2.8686e-17 1.24 30 2.61e+01 3.1126e-18 5.7367e-18 2.20 35 6.48e+01 7.0992e-18 1.2392e-17 1.75 40 1.05e+02 1.7654e-18 6.4930e-18 3.68

The results in Table1 show that the advantage of ASS over KRP in CPU time grows rapidly as n increases, as predicted by the operation counts.

The results in Table1 show that the advantage of ASS over KRP in CPU time grows rapidly as n increases, as predicted by the operation counts. Even with better management of sparsity or parallellism, the O(n 6 ) operation count makes the KRP approach uncompetitive even for moderate size n. The residuals from ASS is also better than that from KRP, as (2.1) is solved by Gaussian elimination in an unstructured way.

Example2 We let a = [α + ɛ, β], b = [β, α], where α, β are two randomly numbers greater than 1, with the spectral set σ(a, B) = { α+ε β, β α }, and λ 1λ 2 1 = ε α. Judging from (2.16), (1.1) has worsening condition when ε decreases. We report a comparison of absolute residuals for the ASS and KRP approaches for ε = 10 1, 10 3, 10 5, 10 7 and 10 9 in Table2. The results show that if (2.1) is solved by Guassian elimination, its residual will be larger than that for ASS especially for smaller ε. Note that the size of X reflects partially the condition of (1.1). The KRP approach copes less well than the ASS approach for an ill-conditioned problem.

Table2: Results for Example2 ɛ Res(ASS) Res(KRP) Res(KRP) Res(ASS) O( X ) 1.0e-1 2.0673e-15 2.4547e-15 1.19 10 1 1.0e-3 8.6726e-13 4.3279e-13 0.50 10 3 1.0e-5 2.3447e-12 2.4063e-12 1.03 10 3 1.0e-7 5.9628e-10 1.1786e-09 1.98 10 6 1.0e-9 5.8632e-08 3.4069e-07 5.81 10 8

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

Generalized -Sylvester equation I The more general version of the -Sylvester equation I: AXB ± X = C with A, B, X C m n is uniquely solvable if and only if the condition: λ σ(ab) λ σ(ab) is satisfied.

Generalized -Sylvester equation II The more general version of the -Sylvester equation II: AXB ± CX D = E is uniquely solvable if and only if the condition: is satisfied. λ 1, λ 2 σ(a, C); λ 3, λ 4 σ(d, B) λ 1 λ 2 λ 3 λ 4

Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4 Related Equations 5 Conclusions

Conclusions We have considered the solution of the -Sylvester equation which has not been fully investigated before. The solvability conditions have been derived and algorithms have been proposed.

Conclusions We have considered the solution of the -Sylvester equation which has not been fully investigated before. The solvability conditions have been derived and algorithms have been proposed. It is interesting and exciting that the above the second X in (1.1) makes the equation behave very differently. 1. For the ordinary continuous-time Sylvester equation AX ± XB = C, the solvability condition is: σ(a) σ( B) = φ. 2. For the ordinary discrete-time Sylvester equation X ± AXB = C, the the solvability condition is: if λ σ(a), then λ 1 σ(b). (1.1) looks like a Sylvester equation associated with continuous-time but (2.16) is satisfied when σ(a, B) in totally inside the unit circle, hinting at a discrete-time type of stability behaviour.

Thank you for your attention!