Remarks on the Cayley Representation of Orthogonal Matrices and on Perturbing the Diagonal of a Matrix to Make it Invertible

Similar documents
MATH 1210 Assignment 4 Solutions 16R-T1

Numerical Linear Algebra Homework Assignment - Week 2

Math 315: Linear Algebra Solutions to Assignment 7

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

0.1 Rational Canonical Forms

and let s calculate the image of some vectors under the transformation T.

MATH 1553, C. JANKOWSKI MIDTERM 3

Evaluating Determinants by Row Reduction

88 CHAPTER 3. SYMMETRIES

18.S34 linear algebra problems (2007)

Math 108b: Notes on the Spectral Theorem

Topics in linear algebra

Components and change of basis

Counting Matrices Over a Finite Field With All Eigenvalues in the Field

NOTES FOR LINEAR ALGEBRA 133

det(ka) = k n det A.

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

Linear Systems and Matrices

Math 113 Homework 5. Bowei Liu, Chao Li. Fall 2013

A linear algebra proof of the fundamental theorem of algebra

A linear algebra proof of the fundamental theorem of algebra

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

Properties of the Determinant Function

MATH 369 Linear Algebra

The Cartan Dieudonné Theorem

MATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants.

MATH Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product.

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier

MATH 223A NOTES 2011 LIE ALGEBRAS 35

be any ring homomorphism and let s S be any element of S. Then there is a unique ring homomorphism

Math Matrix Algebra

Introduction to Computational Manifolds and Applications

Conceptual Questions for Review

Exercise Set 7.2. Skills

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

REU 2007 Apprentice Class Lecture 8

8 Square matrices continued: Determinants

Introduction to Matrix Algebra

Determinants. Recall that the 2 2 matrix a b c d. is invertible if

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

Econ Slides from Lecture 7

REPRESENTATION THEORY WEEK 7

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

No books, notes, any calculator, or electronic devices are allowed on this exam. Show all of your steps in each answer to receive a full credit.

Therefore, A and B have the same characteristic polynomial and hence, the same eigenvalues.

4. Determinants.

A matrix A is invertible i det(a) 6= 0.

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

Operators with numerical range in a closed halfplane

4.2. ORTHOGONALITY 161

= W z1 + W z2 and W z1 z 2

Chapter 5: Matrices. Daniel Chan. Semester UNSW. Daniel Chan (UNSW) Chapter 5: Matrices Semester / 33

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Diagonalizing Matrices

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Extra Problems for Math 2050 Linear Algebra I

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics

CSL361 Problem set 4: Basic linear algebra

Linear Algebra Primer

Lecture 19: Polar and singular value decompositions; generalized eigenspaces; the decomposition theorem (1)

Math Computation Test 1 September 26 th, 2016 Debate: Computation vs. Theory Whatever wins, it ll be Huuuge!

Chapter SSM: Linear Algebra Section Fails to be invertible; since det = 6 6 = Invertible; since det = = 2.

SPRING OF 2008 D. DETERMINANTS

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Solution to Homework 1

Notes on the matrix exponential

Schur s Triangularization Theorem. Math 422

NOTES ON LINEAR ALGEBRA. 1. Determinants

Determinants Chapter 3 of Lay

Math 304 Fall 2018 Exam 3 Solutions 1. (18 Points, 3 Pts each part) Let A, B, C, D be square matrices of the same size such that

1 Matrices and Systems of Linear Equations. a 1n a 2n

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

12. Perturbed Matrices

Calculating determinants for larger matrices

Final Exam. Linear Algebra Summer 2011 Math S2010X (3) Corrin Clarkson. August 10th, Solutions

3 Matrix Algebra. 3.1 Operations on matrices

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math Linear Algebra Final Exam Review Sheet

MAT 2037 LINEAR ALGEBRA I web:

Symmetric and anti symmetric matrices

Math 346 Notes on Linear Algebra

JORDAN NORMAL FORM. Contents Introduction 1 Jordan Normal Form 1 Conclusion 5 References 5

(Inv) Computing Invariant Factors Math 683L (Summer 2003)

Singular Value Decomposition (SVD) and Polar Form

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

Don t forget to think of a matrix as a map (a.k.a. A Linear Algebra Primer)

1 :: Mathematical notation

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A =

Math 215 HW #9 Solutions

MATH2210 Notebook 2 Spring 2018

1 Planar rotations. Math Abstract Linear Algebra Fall 2011, section E1 Orthogonal matrices and rotations

Linear Algebra: Lecture notes from Kolman and Hill 9th edition.

First we introduce the sets that are going to serve as the generalizations of the scalars.

Chapter 2:Determinants. Section 2.1: Determinants by cofactor expansion

EIGENVALUES AND EIGENVECTORS

Lecture 23: Trace and determinants! (1) (Final lecture)

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Transcription:

Remarks on the Cayley Representation of Orthogonal Matrices and on Perturbing the Diagonal of a Matrix to Make it Invertible Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA jean@cis.upenn.edu June 20, 2014 Abstract. This note contains two remarks. The first remark concerns the extension of the well-known Cayley representation of rotation matrices by skew-symmetric matrices to rotation matrices admitting 1 as an eigenvalue and then to all orthogonal matrices. We review a method due to Hermann Weyl and another method involving multiplication by a diagonal matrix whose entries are +1 or 1. The second remark has to do with ways of flipping the signs of the entries of a diagonal matrix, C, with nonzero diagonal entries, obtaining a new matrix, E, so that E+A is invertible, where A is any given matrix (invertible or not). 1

1 The Cayley Representation of Orthogonal Matrices Given any rotation matrix, R SO(n), if R does not admit 1 as an eigenvalue, then there is a unique skew-symmetric matrix, S, (S = S) so that R = (I S)(I + S) 1. This is a classical result of Cayley [3] (1846) and R is called the Cayley transform of S. Among other sources, a proof can be found in Hermann Weyl s beautiful book The Classical Groups [7], Chapter II, Section 10, Theorem 2.10.B (page 57). As we can see, this representation misses rotation matrices admitting the eigenvalue 1, and of course, as det((i S)(I + S) 1 ) = +1, it misses improper orthogonal matrices, i.e., those matrices R O(n) with det(r) = 1. Question 1. Is there a way to extend the Cayley representation to all rotation matrices (matrices in SO(n))? Question 2. Is there a way to extend the Cayley representation to all orthogonal matrices (matrices in O(n))? Answer: Yes in both cases! An answer to Question 1 is given in Weyl s book [7], Chapter II, Section 10, Lemma 2.10.D (page 60): Proposition 1.1 (Weyl) Every rotation matrix, R SO(n), can be expressed as a product R = (I S 1 )(I + S 1 ) 1 (I S 2 )(I + S 2 ) 1, where S 1 and S 2 are skew-symmetric matrices. Thus, if we allow two Cayley representation matrices, we can capture orthogonal matrices having an even number of 1 as eigenvalues. Actually, proposition 1.1 can be sharpened slightly as follows: Proposition 1.2 Every rotation matrix, R SO(n), can be expressed as where S is a skew-symmetric matrix. ) 2 R = ((I S)(I + S) 1 Proposition 1.2 can be easily proved using the following well-known normal form for orthogonal matrices: 2

Proposition 1.3 For every orthogonal matrix, R O(n), there is an orthogonal matrix P and a block diagonal matrix D such that R = P D P, where D is of the form D 1... D 2... D =......... D p such that each block D i is either 1, 1, or a two-dimensional matrix of the form ( ) cos θi sin θ D i = i sin θ i cos θ i where 0 < θ i < π. In particular, if R is a rotation matrix (R SO(n)), then it has an even number of eigenvalues 1. So, they can be grouped into two-dimensional rotation matrices of the form ( ) 1 0, 0 1 i.e., we allow θ i = π and we may assume that D does not contain one-dimensional blocks of the form 1. A proof of Proposition 1.3 can be found in Gantmacher [5], Chapter IX, Section 13 (page 285), or Berger [2], or Gallier [4], Chapter 11, Section 11.4 (Theorem 11.4.5). Now, for every two-dimensional rotation matrix ( ) cos θ sin θ T = sin θ cos θ with 0 < θ π, observe that ( ) T 1 2 cos(θ/2) sin(θ/2) = sin(θ/2) cos(θ/2) does not admit 1 as an eigenvalue (since 0 < θ/2 π/2) and T = ( T 1 2 ) 2. Thus, if we form the matrix R 1 2 by replacing each two-dimensional block D i in the above normal ( form ) by D 1 2 2 i, we obtain a rotation matrix that does not admit 1 as an eigenvalue, R = R 1 2 and the Cayley transform of R 1 2 is well defined. Therefore, we have proved Proposition 1.2. Next, why is the answer to Question 2 also yes? This is because 3

Proposition 1.4 For any orthogonal matrix, R O(n), there is some diagonal matrix, E, whose entries are +1 or 1, and some skew-symmetric matrix, S, so that R = E(I S)(I + S) 1. As such matrices E are orthogonal, all matrices E(I S)(I + S) 1 are orthogonal, so we have a Cayley-style representation of all orthogonal matrices. I am not sure when Proposition 1.4 was discovered and originally published. Since I could not locate this result in Weyl s book [7], I assume that it was not known before 1946, but I did stumble on it as an exercise in Richard Bellman s classic [1], first published in 1960, Chapter 6, Section 4, Exercise 11, page 91-92 (see also, Exercises, 7, 8, 9, and 10). Why does this work? Fact E: Because, for every n n matrix, A (invertible or not), there some diagonal matrix, E, whose entries are +1 or 1, so that I + EA is invertible! This is Exercise 10 in Bellman [1] (Chapter 6, Section 4, page 91). Using Fact E, it is easy to prove Proposition 1.4. Proof of Proposition 1.4. Let R O(n) be any orthogonal matrix. By Fact E, we can find a diagonal matrix, E (with diagonal entries ±1), so that I + ER is invertible. But then, as E is orthogonal, ER is an orthogonal matrix that does not admit the eigenvalue 1 and so, by the Cayley representation theorem, there is a skew-symmetric matrix, S, so that However, notice that E 2 = I, so we get as claimed. But Why does Fact E hold? ER = (I S)(I + S) 1. R = E(I S)(I + S) 1, As we just observed, E 2 = I, so by multiplying by E, I + EA is invertible iff E + A is. Thus, we are naturally led to the following problem: If A is any n n matrix, is there a way to perturb the diagonal entries of A, i.e., to add some diagonal matrix, C = diag(c 1,..., c n ), to A so that C + A becomes invertible? Indeed this can be done, and we will show in the next section that what matters is not the magnitude of the perturbation but the signs of the entries being added. 4

2 Perturbing the Diagonal of a Matrix to Make it Invertible In this section we prove the following result: Proposition 2.1 For every n n matrix (invertible or not), A, and every any diagonal matrix, C = diag(c 1,..., c n ), with c i 0 for i = 1,..., n, there an assignment of signs, ɛ i = ±1, so that if E = diag(ɛ 1 c 1,..., ɛ n c n ), then E + A is invertible. Proof. Let us evaluate the determinant of C+A. We see that = det(c+a) is a polynomial of degree n in the variables c 1,..., c n and that all the monomials of consist of products of distinct variables (i.e., every variable occurring in a monomial has degree 1). In particular, contains the monomial c 1 c n. In order to prove Proposition 2.1, it will suffice to prove Proposition 2.2 Given any polyomial, P (x 1,..., x n ), of degree n (in the indeterminates x 1,..., x n and over any integral domain of characteristic unequal to 2), if every monomial in P is a product of distinct variables, for every n-tuple (c 1,..., c n ) such that c i 0 for i = 1,..., n, then there is an assignment of signs, ɛ i = ±1, so that P (ɛ 1 c 1,..., ɛ n c n ) 0. Clearly, any assignment of signs given by Proposition 2.2 will make det(e + A) 0, proving Proposition 2.1. It remains to prove Proposition 2.2. Proof of Proposition 2.2. We proceed by induction on n (starting with n = 1). For n = 1, the polynomial P (x 1 ) is of the form P (x 1 ) = a + bx 1, with b 0 since deg(p ) = 1. Obviously, for any c 0, either a + bc 0 or a bc 0 (otherwise, 2bc = 0, contradicting b 0, c 0 and the ring being an integral domain of characteristic 2). Assume the induction hypothesis holds for any n 1 and let P (x 1,..., x n+1 ) be a polynomial of degree n + 1 satisfying the conditions of Proposition 2.2. Then, P must be of the form P (x 1,..., x n, x n+1 ) = Q(x 1,..., x n ) + S(x 1,..., x n )x n+1, where both Q(x 1,..., x n ) and S(x 1,..., x n ) are polynomials in x 1,..., x n and S(x 1,..., x n ) is of degree n and all monomials in S are products of distinct variables. By the induction hypothesis, we can find (ɛ 1,..., ɛ n ), with ɛ i = ±1, so that S(ɛ 1 c 1,..., ɛ n c n ) 0. But now, we are back to the case n = 1 with the polynomial Q(ɛ 1 c 1,..., ɛ n c n ) + S(ɛ 1 c 1,..., ɛ n c n )x n+1, 5

and we can find ɛ n+1 = ±1 so that P (ɛ 1 c 1,..., ɛ n c n, ɛ n+1 c n+1 ) = Q(ɛ 1 c 1,..., ɛ n c n ) + S(ɛ 1 c 1,..., ɛ n c n )ɛ n+1 c n+1 0, establishing the induction hypothesis. Note that in Proposition 2.1, the c i can be made arbitrarily small or large, as long as they are not zero. What matters is the signs that are assigned to the perturbation. Another nice proof of Fact E is given in a short note by William Kahan [6]. Due to its elegance, we feel compelled to sketch Kahan s proof. This proof uses two facts: (1) If A = (A 1,..., A n 1, U) and B = (A 1,..., A n 1, V ) are two n n matrices that differ in their last column, then det(a + B) = 2 n 1 (det(a) + det(b)). This is because determinants are multilinear (alternating) maps of their columns. Therefore, if det(a) = det(b) = 0, then det(a + B) = 0. Obviously, this fact also holds whenever A and B differ by just one column (not just the last one). (2) For every k = 0,... 2 n 1, if we write k in binary as k = k n k 1, then let E k be the diagonal matrix whose ith diagonal entry is 1 iff k i = 1, else +1 iff k i = 0. For example, E 0 = I and E 2 n 1 = I. Observe that E k and E k+1 differ by exactly one column. Then, it is easy to see that E 0 + E 1 + + E 2 n 1 = 0. The proof proceeds by contradiction. Assume that det(i + E k A) = 0, for k = 0,..., 2 n 1. The crux of the proof is that det(i + E 0 A + I + E 1 A + I + E 2 A + + I + E 2 n 1A) = 0. However, as E 0 + E 1 + + E 2 n 1 = 0, we see that and so, I + E 0 A + I + E 1 A + I + E 2 A + + I + E 2 n 1A = 2 n I, 0 = det(i + E 0 A + I + E 1 A + I + E 2 A + + I + E 2 n 1A) = det(2 n I) = 2 n 0, a contradiction! To prove that det(i + E 0 A + I + E 1 A + I + E 2 A + + I + E 2 n 1A) = 0, we observe using fact (2) that, det(i + E 2i A + I + E 2i+1 A) = det(i + E 2i A) + det(i + E 2i+1 A) = 0, 6

for i = 0,..., 2 n 1 1; similarly, det(i + E 4i A + I + E 4i+1 A + I + E 4i+2 A + I + E 4i+3 A) = 0, for i = 0,..., 2 n 2 1; by induction, we get det(i + E 0 A + I + E 1 A + I + E 2 A + + I + E 2 n 1A) = 0, which concludes the proof. Final Questions: (1) When was Fact E first stated and by whom (similarly for Proposition 1.4)? (2) Can Proposition 2.1 be generalized to non-diagonal matrices (in an interesting way)? References [1] Richard Bellman. Introduction to Matrix Analysis. Classics in Applied Mathematics, No. 12. SIAM Publications, first edition, 1995. [2] Marcel Berger. Géométrie 1. Nathan, 1990. English edition: Geometry 1, Universitext, Springer Verlag. [3] Arthur Cayley. About the algebraic structure of the orthogonal group and the other classical groups in a field of characteristic zero or a prime characteristic. J. Reine Angew. Math., 32, 1846. [4] Jean H. Gallier. Geometric Methods and Applications, For Computer Science and Engineering. TAM, Vol. 38. Springer, first edition, 2000. [5] F.R. Gantmacher. The Theory of Matrices, Vol. I. AMS Chelsea, first edition, 1977. [6] William Kahan. Is there a small skew Cayley transform with zero diagonal. Technical report, University of California, Berkeley, Mathematics Department, Berkeley, CA 94720-3840, 2004. pdf file available from http://www.cs.berkeley.edu/ wkahan. [7] Hermann Weyl. The Classical Groups. Their Invariants and Representations. Princeton Mathematical Series, No. 1. Princeton University Press, second edition, 1946. 7