QR-Decomposition for Arbitrary Matrices

Similar documents
The Cartan Dieudonné Theorem

4.2. ORTHOGONALITY 161

Basics of Hermitian Geometry

Important Matrix Factorizations

Schur s Triangularization Theorem. Math 422

ISOMETRIES OF R n KEITH CONRAD

Chapter 5. Basics of Euclidean Geometry

9.1 Eigenvectors and Eigenvalues of a Linear Map

Review of linear algebra

f(a)f(b) = ab for all a, b E. WhenE = F, an affine isometry f: E E is also called a rigid motion.

1 Planar rotations. Math Abstract Linear Algebra Fall 2011, section E1 Orthogonal matrices and rotations

Section 6.4. The Gram Schmidt Process

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Linear Algebra Review. Vectors

Lecture 4 Orthonormal vectors and QR factorization

Linear Analysis Lecture 16

Math 25a Practice Final #1 Solutions

Lecture 19: Isometries, Positive operators, Polar and singular value decompositions; Unitary matrices and classical groups; Previews (1)

INNER PRODUCT SPACE. Definition 1

Math 240 Calculus III

Math Linear Algebra II. 1. Inner Products and Norms

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Lecture Notes Introduction to Cluster Algebra

Fall, 2003 CIS 610. Advanced geometric methods. Homework 3. November 11, 2003; Due November 25, beginning of class

October 25, 2013 INNER PRODUCT SPACES

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Solution of Linear Equations

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Review problems for MA 54, Fall 2004.

Math 215B: Solutions 1

The Orthogonal Geometry of R n

Review of Linear Algebra

SOLUTIONS: ASSIGNMENT Use Gaussian elimination to find the determinant of the matrix. = det. = det = 1 ( 2) 3 6 = 36. v 4.

LINEAR ALGEBRA KNOWLEDGE SURVEY

Quantum Computing Lecture 2. Review of Linear Algebra

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Symmetric matrices and dot products

No books, notes, any calculator, or electronic devices are allowed on this exam. Show all of your steps in each answer to receive a full credit.

Jordan Normal Form and Singular Decomposition

The Spectral Theorem for normal linear maps

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

18.06 Professor Johnson Quiz 1 October 3, 2007

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Conceptual Questions for Review

Singular Value Decomposition (SVD) and Polar Form

Vector Spaces, Affine Spaces, and Metric Spaces

Final EXAM Preparation Sheet

Introduction to Linear Algebra. Tyrone L. Vincent

Assignment #9: Orthogonal Projections, Gram-Schmidt, and Least Squares. Name:

Spectral Theorems in Euclidean and Hermitian Spaces

235 Final exam review questions

ft-uiowa-math2550 Assignment NOTRequiredJustHWformatOfQuizReviewForExam3part2 due 12/31/2014 at 07:10pm CST

Some notes on Coxeter groups

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

7. Symmetric Matrices and Quadratic Forms

Math 113 Final Exam: Solutions

LinGloss. A glossary of linear algebra

Lecture notes: Applied linear algebra Part 1. Version 2

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

This can be accomplished by left matrix multiplication as follows: I

Linear Algebra II. 7 Inner product spaces. Notes 7 16th December Inner products and orthonormal bases

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

EECS 275 Matrix Computation

Introduction to Lie Groups

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

LINEAR ALGEBRA SUMMARY SHEET.

MATH 22A: LINEAR ALGEBRA Chapter 4

Rigid Geometric Transformations

Spring, 2012 CIS 515. Fundamentals of Linear Algebra and Optimization Jean Gallier

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

5.6. PSEUDOINVERSES 101. A H w.

Topic 2 Quiz 2. choice C implies B and B implies C. correct-choice C implies B, but B does not imply C

ALGEBRA 8: Linear algebra: characteristic polynomial

1. General Vector Spaces

Math Matrix Theory - Spring 2012

Matrix Factorization and Analysis

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

1 Last time: least-squares problems

6. Orthogonality and Least-Squares

22m:033 Notes: 3.1 Introduction to Determinants

Warm-up. True or false? Baby proof. 2. The system of normal equations for A x = y has solutions iff A x = y has solutions

MATH 581D FINAL EXAM Autumn December 12, 2016

Diagonalizing Matrices

MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS

Math 396. Quotient spaces

Vector Spaces ปร ภ ม เวกเตอร

2. Every linear system with the same number of equations as unknowns has a unique solution.

MATH 223A NOTES 2011 LIE ALGEBRAS 35

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Differential Topology Final Exam With Solutions

TBP MATH33A Review Sheet. November 24, 2018

Math113: Linear Algebra. Beifang Chen

Spectral Theorems in Euclidean and Hermitian Spaces

Rigid Geometric Transformations

Transcription:

Chapter 10 QR-Decomposition for Arbitrary Matrices 10.1 Orthogonal Reflections Orthogonal symmetries are a very important example of isometries. First let us review the definition of a (linear) projection. Given a vector space E, letf and G be subspaces of E that form a direct sum E = F G. Since every u 2 E can be written uniquely as u = v + w, wherev 2 F and w 2 G, wecandefinethe two projections p F : E! F and p G : E! G, suchthat p F (u) =v and p G (u) =w. 565

566 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES It is immediately verified that p G and p F are linear maps, and that p 2 F = p F, p 2 G = p G, p F p G = p G p F =0,and p F + p G =id. Definition 10.1. Given a vector space E, foranytwo subspaces F and G that form a direct sum E = F G, the symmetry with respect to F and parallel to G, or reflection about F is the linear map s: E! E, defined such that s(u) =2p F (u) u, for every u 2 E. Because p F + p G =id,notethatwealsohave s(u) =p F (u) p G (u) and s(u) =u 2p G (u), s 2 =id,s is the identity on F,ands = id on G.

10.1. ORTHOGONAL REFLECTIONS 567 We now assume that E is a Euclidean space of finite dimension. Definition 10.2. Let E be a Euclidean space of finite dimension n. ForanytwosubspacesF and G, iff and G form a direct sum E = F G and F and G are orthogonal, i.e. F = G?,theorthogonal symmetry with respect to F and parallel to G, or orthogonal reflection about F is the linear map s: E! E, definedsuchthat for every u 2 E. s(u) =2p F (u) u, When F is a hyperplane, we call s an hyperplane symmetry with respect to F or reflection about F,andwhen G is a plane, we call s a flip about F. It is easy to show that s is an isometry.

568 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES Using Proposition 9.8, it is possible to find an orthonormal basis (e 1,...,e n )ofe consisting of an orthonormal basis of F and an orthonormal basis of G. Assume that F has dimension p, sothatg has dimension n p. With respect to the orthonormal basis (e 1,...,e n ), the symmetry s has a matrix of the form Ip 0 0 I n p

10.1. ORTHOGONAL REFLECTIONS 569 Thus, det(s) =( 1) n p,andsis a rotation i n p is even. In particular, when F is a hyperplane H, wehave p = n 1, and n p =1,sothats is an improper orthogonal transformation. When F = {0}, wehaves = id, which is called the symmetry with respect to the origin. The symmetry with respect to the origin is a rotation i n is even, and an improper orthogonal transformation i n is odd. When n is odd, we observe that every improper orthogonal transformation is the composition of a rotation with the symmetry with respect to the origin.

570 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES When G is a plane, p = n 2, and det(s) =( 1) 2 =1, so that a flip about F is a rotation. In particular, when n =3,F is a line, and a flip about the line F is indeed a rotation of measure. When F = H is a hyperplane, we can give an explicit formula for s(u)in terms of any nonnull vector w orthogonal to H. We get s(u) =u 2 (u w) kwk 2 w. Such reflections are represented by matrices called Householder matrices, andtheyplayanimportantroleinnumerical matrix analysis. Householder matrices are symmetric and orthogonal.

10.1. ORTHOGONAL REFLECTIONS 571 Over an orthonormal basis (e 1,...,e n ), a hyperplane reflection about a hyperplane H orthogonal to a nonnull vector w is represented by the matrix H = I n 2 WW> kw k 2 = I n 2 WW> W > W, where W is the column vector of the coordinates of w. Since p G (u) = (u w) kwk 2 w, the matrix representing p G is WW > W > W, and since p H + p G =id,thematrixrepresentingp H is I n WW > W > W.

572 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES The following fact is the key to the proof that every isometry can be decomposed as a product of reflections. Proposition 10.1. Let E be any nontrivial Euclidean space. For any two vectors u, v 2 E, if kuk = kvk, then there is an hyperplane H such that the reflection s about H maps u to v, and if u 6= v, then this reflection is unique. We now show that Hyperplane reflections can be used to obtain another proof of the QR-decomposition.

10.2. QR-DECOMPOSITION USING HOUSEHOLDER MATRICES 573 10.2 QR-Decomposition Using Householder Matrices First, we state the result geometrically. When translated in terms of Householder matrices, we obtain the fact advertised earlier that every matrix (not necessarily invertible) has a QR-decomposition. Proposition 10.2. Let E be a nontrivial Euclidean space of dimension n. Given any orthonormal basis (e 1,...,e n ), for any n-tuple of vectors (v 1,...,v n ), there is a sequence of n isometries h 1,...,h n, such that h i is a hyperplane reflection or the identity, and if (r 1,...,r n ) are the vectors given by r j = h n h 2 h 1 (v j ), then every r j is a linear combination of the vectors (e 1,...,e j ), (1 apple j apple n). Equivalently, the matrix R whose columns are the components of the r j over the basis (e 1,...,e n ) is an upper triangular matrix. Furthermore, the h i can be chosen so that the diagonal entries of R are nonnegative.

574 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES Remarks. (1)Sinceeveryh i is a hyperplane reflection or the identity, = h n h 2 h 1 is an isometry. (2) If we allow negative diagonal entries in R, thelast isometry h n may be omitted. (3) Instead of picking r k,k = ku 00 k k,whichmeansthat w k = r k,k e k u 00 k, where 1 apple k apple n, itmightbepreferabletopick r k,k = ku 00 k k if this makes kw kk 2 larger, in which case w k = r k,k e k + u 00 k. Indeed, since the definition of h k involves division by kw k k 2,itisdesirabletoavoiddivisionbyverysmallnumbers. Proposition 10.2 immediately yields the QR-decomposition in terms of Householder transformations.

10.2. QR-DECOMPOSITION USING HOUSEHOLDER MATRICES 575 Theorem 10.3. For every real n n-matrix A, there is a sequence H 1,...,H n of matrices, where each H i is either a Householder matrix or the identity, and an upper triangular matrix R, such that R = H n H 2 H 1 A. As a corollary, there is a pair of matrices Q, R, where Q is orthogonal and R is upper triangular, such that A = QR (a QR-decomposition of A). Furthermore, R can be chosen so that its diagonal entries are nonnegative. Remarks. (1)Letting A k+1 = H k H 2 H 1 A, with A 1 = A, 1apple k apple n, theproofofproposition10.2 can be interpreted in terms of the computation of the sequence of matrices A 1,...,A n+1 = R.

576 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES The matrix A k+1 has the shape A k+1 = 0 u k+1 1 1 0...... 0 0 u k+1 k 0 0 0 u k+1 k+1 0 0 0 u k+1 k+2 B........ @ 0 0 0 u k+1 n 1 C A 0 0 0 u k+1 n where the (k +1)thcolumnofthematrixisthevector u k+1 = h k h 2 h 1 (v k+1 ), and thus and u 0 k+1 =(u k+1 1,...,u k+1 k ), u 00 k+1 =(u k+1 k+1,uk+1 k+2,...,uk+1 n ). If the last n k 1entriesincolumnk +1areallzero, there is nothing to do and we let H k+1 = I.

10.2. QR-DECOMPOSITION USING HOUSEHOLDER MATRICES 577 Otherwise, we kill these n k 1entriesbymultiplying A k+1 on the left by the Householder matrix H k+1 sending (0,...,0,u k+1 k+1,...,uk+1 n )to(0,...,0,r k+1,k+1, 0,...,0), where r k+1,k+1 = (u k+1 k+1,...,uk+1 n ). (2) If we allow negative diagonal entries in R, thematrix H n may be omitted (H n = I). (3) If A is invertible and the diagonal entries of R are positive, it can be shown that Q and R are unique.

578 CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES (4) The method allows the computation of the determinant of A. Wehave det(a) =( 1) m r 1,1 r n,n, where m is the number of Householder matrices (not the identity) among the H i. (5) The condition number of the matrix A is preserved. This is very good for numerical stability.