Is there a Small Skew Cayley Transform with Zero Diagonal?

Similar documents
Is there a small skew Cayley transform with zero diagonal?

Linear Algebra: Matrix Eigenvalue Problems

Diagonalizing Matrices

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Lecture 10: Eigenvectors and eigenvalues (Numerical Recipes, Chapter 11)

Numerical Linear Algebra Homework Assignment - Week 2

Lecture 8 : Eigenvalues and Eigenvectors

Math 108b: Notes on the Spectral Theorem

The skew-symmetric orthogonal solutions of the matrix equation AX = B

Introduction. Chapter One

Mathematical Methods for Engineers and Scientists 1

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Absolute value equations

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

AMS526: Numerical Analysis I (Numerical Linear Algebra)

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Linear Algebra 2 Spectral Notes

Do MATLAB s lu( ), inv( ), / and \ have Failure Modes?

Chap 3. Linear Algebra

Eigenvalue PDFs. Peter Forrester, M&S, University of Melbourne

Chapter 3 Transformations

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Foundations of Matrix Analysis

Spectral Theorem for Self-adjoint Linear Operators

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Introduction to Matrices

Usually, when we first formulate a problem in mathematics, we use the most familiar

NOTES ON BILINEAR FORMS

Detailed Proof of The PerronFrobenius Theorem

Equivalent Conditions on Conjugate Unitary Matrices

Numerical Methods - Numerical Linear Algebra

Algebra I Fall 2007

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

FFTs in Graphics and Vision. Groups and Representations

Eigenvalues and eigenvectors

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math. H110 Answers for Final Examination December 21, :56 pm

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Matrix Factorization and Analysis

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...

Mathematical Methods wk 2: Linear Operators

and let s calculate the image of some vectors under the transformation T.

Eigenpairs and Similarity Transformations

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Quantum Computing Lecture 2. Review of Linear Algebra

Constructive Aspects for the Generalized Orthogonal Group. Steven Christopher Fuller

Linear Algebra Massoud Malek

GENERAL ARTICLE Realm of Matrices

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Matrices A brief introduction

Eigenvectors and Hermitian Operators

EECS 275 Matrix Computation

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

Symmetric and anti symmetric matrices

A PRIMER ON SESQUILINEAR FORMS

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Theorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k

Iterative methods for Linear System

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Chapter One. Introduction

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Numerical Linear Algebra

Solutions to Problems for Math. H90 Issued 27 Aug. 2007

Linear algebra 2. Yoav Zemel. March 1, 2012

Introduction to Iterative Solvers of Linear Systems

1 Matrices and vector spaces

Lecture 6 Positive Definite Matrices

We wish to solve a system of N simultaneous linear algebraic equations for the N unknowns x 1, x 2,...,x N, that are expressed in the general form

Chapter XI Novanion rings

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

Remarks on the Cayley Representation of Orthogonal Matrices and on Perturbing the Diagonal of a Matrix to Make it Invertible

Archive of past papers, solutions and homeworks for. MATH 224, Linear Algebra 2, Spring 2013, Laurence Barker

Eigenvalue and Eigenvector Problems

Chapter 8. Rigid transformations

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Chapter 6. Eigenvalues. Josef Leydold Mathematical Methods WS 2018/19 6 Eigenvalues 1 / 45

Quantum Theory and Group Representations

Algebra C Numerical Linear Algebra Sample Exam Problems

1 Matrices and Systems of Linear Equations. a 1n a 2n

Geometry of the Special Unitary Group

FFTs in Graphics and Vision. The Laplace Operator

Math 489AB Exercises for Chapter 2 Fall Section 2.3

5.6. PSEUDOINVERSES 101. A H w.

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Singular Value Decompsition

Algebra Workshops 10 and 11

Lecture Note 1: Background

ELEMENTARY MATRIX ALGEBRA

CS 246 Review of Linear Algebra 01/17/19

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Math 489AB Exercises for Chapter 1 Fall Section 1.0

LINEAR ALGEBRA REVIEW

Transcription:

Is there a Small Skew Cayley Transform with Zero Diagonal? Abstract The eigenvectors of an Hermitian matrix H are the columns of some complex unitary matrix Q. For any diagonal unitary matrix Ω the columns of Q Ω are eigenvectors too. Among all such Q Ω at least one has a skew-hermitian Cayley transform S := (I+Q Ω) 1 (I Q Ω) with just zeros on its diagonal. Why? The proof is unobvious, as is the further observation that Ω may also be so chosen that no element of this S need exceed 1 in magnitude. Thus, plausible constraints, easy to satisfy by perturbations of complex eigenvectors when an Hermitian matrix H is perturbed infinitesimally, can be satisfied for discrete perturbations too. And if H is real symmetric, Q real orthogonal and Ω restricted to diagonals of ±1 s, then, as Evan O Dorney [2014] has proved recently, at least one real skewsymmetric S must have no element bigger than 1 in magnitude. Full text posted at http://www.cs.berkeley.edu/~wkahan/skcayley.pdf Prof. W. Kahan Page 1/15

Hermitian Eigenproblem Hermitian Matrix H = H H = H T complex conjugate transpose. Real Eigenvalues v 1 v 2 v 3 v n sorted and put into a column vector v := [ v 1, v 2, v 3,, v n ] T Corresponding eigenvector columns q 1, q 2, q 3,, q n need not be determined uniquely but can always be chosen to constitute columns of a Unitary matrix Q satisfying H Q = Q Diag(v) and Q H = Q 1. Q Ω is also an eigenvector matrix for every unitary diagonal matrix Ω = Ω 1. Familiar special case: Real symmetric H = H T, real orthogonal Q = Q 1 T. Q Ω is also an eigenvector matrix for every diagonal matrix Ω =Diag(±1 s). Prof. W. Kahan Page 2/15

Perturbed Hermitian Eigenproblem Given Hermitian Matrix H = H o + H for small H. Suppose H o has known eigenvalue column v o and eigenvector matrix Q o. Then eigenvalue column v of H must be close to v o : v v o H. But no eigenvector matrix Q of H need be near Q o unless H is rather smaller than gaps between adjacent eigenvalues v j of H, or of H o. Cautionary Examples: For every tiny nonzero θ, no matter how tiny, 1 + θ 0 0 1 θ 1 θ 0 0 1 + θ H = has eigenvectors rotated through π/2 from H o =. 1 θ θ 1 1 θ 0 0 1 + θ H = has eigenvectors rotated through π/4 from H o =. See Parlett s book and papers by C. Davis & W. Kahan, and by Paige & Wei, on rotations of eigenspaces. Still, how are tiny perturbations of eigenvector matrices to be represented? Prof. W. Kahan Page 3/15

Infinitesimally Perturbed Unitary Matrices Say Q = Q 1 H is perturbed to Q + dq = (Q + dq) 1 H ; then O = di = d(q H Q) = dq H Q + Q H dq, so dq = 2Q ds for some Skew-Hermitian ds = ds H, and Q + dq = Q (I 2dS). This is what brings skew-hermitian matrices to our attention. Discretely Perturbed Unitary Matrices Say Q = Q 1 H is perturbed to a nearby Q + Q = (Q + Q) 1 H ; then either Q + Q = Q e 2 Z for some small skew-hermitian Z, or Q + Q = Q (I + S) 1 (I S) for a small skew-hermitian S. ~~~~~~~~~~~ This is what brings the Cayley Transform to our attention. Prof. W. Kahan Page 4/15

What is a Cayley Transform $(z)? $(z) is an analytic function of a complex variable z on the Riemann sphere, Closed by one point at. 1) It is Involutary : $($(z)) = z. so $ must be Bilinear Rational. 2) It swaps Invert Negate : $( z) = 1/$(z) and so $(1/z) = $(z). 1 z z+ 1 Inference: Only two choices for $(z), ----------- or -----------. Our choice is 1 + z z 1 $ maps 1 z $(z) := -----------, chosen so that $(0) = 1. 1+ z Real Axis Real Axis, Imaginary Axis Unit Circle, Right Half-Plane Unit Disk, Real Orthogonal Matrix Q = Q 1 T Real Skew-Symmetric S = S T, Complex Unitary Matrix Q = Q 1 H Complex Skew-Hermitian S = S H. Prof. W. Kahan Page 5/15

Evading the Cayley Transform s Pole $(B) := (I + B) 1 (I B) Unitary Q = $(S) = Q 1 H Skew-Hermitian S = $(Q) = S H provided det(i + Q) 0 S is finite. Every unitary Q has eigenvalues all with magnitude 1 ; but no Cayley transform Q = $(S) has 1 as an eigenvalue. Will this exclude any eigenvectors? No : Lemma: If Q is unitary and if I+Q is singular, then reversing signs of aptly chosen columns of Q will make I+Q nonsingular and provide a finite skew Cayley transform S = $(Q). Proof: Any of many simple computations. The earliest I know appeared in 1960 ; see Exs. 7-11, pp. 92-3 in 4 of Ch. 6 of Richard Bellman s book Introduction to Matrix Analysis (2d. ed. 1970, McGraw-Hill). Or see pp. 2-3 of ~wkahan/skcayley.pdf. Henceforth take det(i + Q) 0 for granted. Prof. W. Kahan Page 6/15

Back to Perturbed Hermitian Eigenproblem Given Hermitian Matrix H = H o + H for small H. Suppose H o has known eigenvalue column v o and eigenvector matrix Q o. W.L.O.G, exposition is simplified by taking the eigenvectors of H o as a new orthonormal coordinate system, so that H o = Diag(v o ). Now we wish to solve H Q = Q Diag(v) and Q H Q = I ( ) for a sorted eigenvalue column v near v o, and a unitary Q not far from I. Substituting Q = $(S) into ( ) transforms it into a slightly less nonlinear (I+S) H (I S) = (I S) Diag(v) (I+S) and S H = S ( ) If all h jk /(h jj h kk ) for j k are so small that 3rd-order S (H Diag(H)) S will be negligible, then equations ( ) have simple approximate solutions 1 v Diag(H) and s jk -- h jk /(h jj h kk ) for j k. 2 Diagonal elements s jj can be arbitrary imaginaries but small lest 3rd-order terms be not negligible. Forcing s jj := 0 seems plausible. But if done when off-diagonal elements are too big to yield acceptable simple approximations to v and S, can ( ) still be solved for v and small S with diag(s) = o? Prof. W. Kahan Page 7/15

Do a sorted eigenvalue column v and a skew S both satisfying (I+S) H (I S) = (I S) Diag(v) (I+S) and S H = S ( ) always exist with diag(s) = o and S not too big? If so, then Q := $(S). Why might we wish to compute v and S, and then Q? Iterative Refinement. The usual way to enhance the accuracy of solutions v and Q of H Q = Q Diag(v) and Q H Q = I ( ) when H is almost diagonal is Jacobi Iteration. It converges quadratically if programmed in a straightforward way, cubically if programmed in a tricky way made doubly tricky if available parallelism is to be exploited too. See its treatment in Golub & Van Loan s book, and recent papers by Drmac & Veselic. If the simple solution of ( ) is adequate, it converges cubically and is easy to parallelize. Sometimes the simple solution is inadequate, and then we seek a better solution of ( ) by some slightly more complicated method. S should not be too big lest Cayley transform Q := (I+S) 1 (I S) be too inaccurate. Thus is the question at the top of this page motivated. Prof. W. Kahan Page 8/15

Do a sorted eigenvalue column v and a skew S both satisfying (I+S) H (I S) = (I S) Diag(v) (I+S) and S H = S ( ) always exist with diag(s) = o and S not too big? YES in the Complex Case, when S can be complex skew-hermitian. And then at least one such S has diag(s) = o and all s jk 1. This Existence Theorem is proved. How best to find that S is not yet known. In the Real Case, when a real H = H T entails a real skew-symmetric S = S T, every diag(s) = o ; and that some such S has all s jk 1 too has been proved recently by Evan O Dorney [2014]. What follows will be first some examples, and then an outline of the Existence Theorem s proof. In what follows, one of the unitary or real orthogonal eigenvector matrices of H is G, and all other eigenvector matrices Q := G Ω of H are generated by letting diagonal matrix Ω runs through all diagonal unitary matrices Ω = e ı Diag(x) with real columns x, or real diagonals Ω = Diag([ ±1, ±1, ±1,, ±1 ]) in the Real Case. Prof. W. Kahan Page 9/15

A 3-by-3 Example Real orthogonal G := $( ) = /13. Q := G Ω ; S := $(Q). diag(s) = o for six diagonal matrices Ω. Four of them are real, namely Ω := I, Diag([ 1, 1, 1 ]), Diag([ 1, 1, 1 ]), and Diag([ 1, 1, 1 ]). Typical of the last three is $(G Diag([ 1, 1, 1 ])) =. = 3/2. The two complex unitary diagonals Ω are scalars Ω := ( 5 ± 12ı) I/13. For them $(G Ω) = 0 2 2 2 0 2 2 2 0 0 1 3ι 1 3ι 1 3ι 0 1 3ι 1 3ι 1 3ι 0 3 4 12 12 3 4 4 12 3 0 1 1 -- 2 1 0 1 1 -- 1 2 0 /4 and its complex conjugate resp. Note that its every element is strictly smaller than 1 in magnitude though still = 3/2. Allowing Q and S to be complex lets S have smaller elements. Prof. W. Kahan Page 10/15

An Extreme n-by-n Example 0 1 0 1 Real orthogonal G :=. Let Ω run through all unitary diagonals with det(ω) 1. These include 2 n 1 real orthogonal diagonals of ±1 s with an even number of 1 s. Every such Ω has det(ω) = 1. Every Q := G Ω is unitary; 2 n 1 are orthogonal. For all such Ω, every off-diagonal element of S := $(Q) has magnitude 2/ 1 + det(ω). It is minimized when det(ω) = +1 ; only then is diag(s) = o, and then every off-diagonal element of S has magnitude 1. This happens in all the Real Cases. For a detailed explanation see /~wkahan/skcayley.pdf. 0 1 0 1 ( 1) n 1 0 Prof. W. Kahan Page 11/15

The Existence Theorem Given a unitary matrix G (of eigenvectors of an Hermitian matrix H ) let Ω run through unitary diagonal matrices, so Q := G Ω is unitary too, (also a matrix of eigenvectors of that Hermitian matrix H ) and let S := $(Q) be the skew-hermitian Cayley transform of Q = $(S). Then diag(s) = o for at least one such S, and its every element has s jk 1. Existence Proof: Among all such Q = G Ω the one(s) nearest the identity I, in a peculiar sense defined hereunder, must turn out to have the desired kind of S = $(Q). The peculiar gauge of nearness of a unitary Q to I is (Q) := log(det((i+q H ) (I+Q)/4)) = log(det( I + $(Q) H $(Q) )). (Q) > 0 for every unitary Q except (I) = 0 and (Q) = + when det(i+q) = 0. What remains of the proof is a characterization of every unitary Q = G Ω that minimizes (Q). For this we need the first two derivatives of. Prof. W. Kahan Page 12/15

How to Derive Derivatives, with respect to a real column-vector x, of (Q) := log(det((i+q H ) (I+Q)/4)) = log(det( log(det((2i+q 1 +Q)/4)) when unitary Q = G Ω = G e ı Diag(x). We shall abbreviate Diag(x) =: X, and then the Differential dx := Diag(dx). Tools: Ω = e ı X has dω = de ı X = Ω e ı dx since diagonals dx and X commute. d(b 1 ) = B 1 db B 1. Jacobi s formula d log(det(b)) = trace(b 1 db). For a derivation see /~wkahan/mathh110/jacobi.pdf. trace(b C) = trace(c B). Using these tools we find first that d (B) = trace($(b) B 1 db) in general, and then that d (G Ω) = ı diag($(g Ω)) T dx, so (G Ω)/ x = ı diag($(g Ω)) T = ı diag(s) T. This must vanish at the minimum (and any other extremum) of (G Ω), so at least one Ω makes S := $(G Ω) have diag(s) = o, as claimed. Prof. W. Kahan Page 13/15

The second derivative of (G e ı Diag(x) ) is representable by a symmetric Hessian matrix M of second partial derivatives that figures in ( 2 (G Ω)/ x 2 ) x dx = dx T M x. For any fixed x a lengthy computation of ( 2 (G Ω)/ x 2 ) x dx = d( (G Ω)/ x) x = d(ı diag($(g Ω)) T ) x = yields Hessian M = (I + S 2 )/2 in which S = $(G Ω) and S 2 is obtained from S elementwise by substituting s jk 2 for every element s jk. At the minimum of (G e ı Diag(x) ) its Hessian M = (I + S 2 )/2 must be positive (semi)definite, and this implies that every s jk 2 1 since diag(s) = o. Thus is the Existence Theorem s second claim confirmed. And the extreme n-by-n example shows that the upper bound 1 is achievable. END of Existence proof. No such proof can work in the Real Case when H is real symmetric, its eigenvector matrix G is real, and Ω is restricted to real orthogonal diagonals. These constitute a discrete set, not a continuum, so derivatives don t matter. Prof. W. Kahan Page 14/15

Conclusion: Perturbing a complex Hermitian matrix H changes its unitary matrix Q of eigenvectors to a perturbed unitary Q (I+S) 1 (I S) in which the skew- Hermitian S = S H can always be chosen to be small ( no element bigger than 1 in magnitude ) and to have only zeros on its diagonal. When H is real symmetric and Q is real orthogonal and S is restricted to be real skewsymmetric, Evan O Dorney [2014] has proved that S can always be chosen to have no element bigger in magnitude than 1. But how to construct such a small skew S efficiently and infallibly is not known yet. Citations W. Kahan [2006] Is there a small skew Cayley transform with zero diagonal? pp. 335-341 of Linear Algebra & Appl. 417 (2-3). Updated posting at www.cs.berkeley.edu/~wkahan/skcayley.pdf Evan O.Dorney [2014] Minimizing the Cayley trasform of an orthogonal matrix by multiplying by signature matrices pp. 97-103 of Linear Algebra & Appl. 448. Evan found this existence proof in 2010 while still an undergraduate at U.C. Berkeley. Prof. W. Kahan Page 15/15