This property turns out to be a general property of eigenvectors of a symmetric A that correspond to distinct eigenvalues as we shall see later.

Similar documents
Chap 3. Linear Algebra

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

1 Last time: least-squares problems

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form:

Econ Slides from Lecture 7

Recall the convention that, for us, all vectors are column vectors.

Chapter 3 Transformations

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Chapter 6: Orthogonality

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

. The following is a 3 3 orthogonal matrix: 2/3 1/3 2/3 2/3 2/3 1/3 1/3 2/3 2/3

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

Preliminary Linear Algebra 1. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 100

Review of Linear Algebra

Applied Linear Algebra in Geoscience Using MATLAB

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

Linear Models Review

Applied Linear Algebra

Conceptual Questions for Review

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Tutorials in Optimization. Richard Socher

Cheat Sheet for MATH461

MAT Linear Algebra Collection of sample exams

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Symmetric matrices and dot products

Transpose & Dot Product

Linear Algebra Review. Vectors

1. First and Second Order Conditions for A Local Min or Max.

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

I. Multiple Choice Questions (Answer any eight)

complex dot product x, y =

MTH 2032 SemesterII

Properties of Matrices and Operations on Matrices

Linear Algebra 2 Spectral Notes

TBP MATH33A Review Sheet. November 24, 2018

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

235 Final exam review questions

Transpose & Dot Product

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

7. Symmetric Matrices and Quadratic Forms

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

EXAM. Exam 1. Math 5316, Fall December 2, 2012

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

Review problems for MA 54, Fall 2004.

Solutions to Review Problems for Chapter 6 ( ), 7.1

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Lecture 7: Positive Semidefinite Matrices

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Introduction to Matrix Algebra

Lecture 15, 16: Diagonalization

MATRICES ARE SIMILAR TO TRIANGULAR MATRICES

Conjugate Gradient (CG) Method

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Math 3191 Applied Linear Algebra

Notes on Eigenvalues, Singular Values and QR

Repeated Eigenvalues and Symmetric Matrices

Numerical Linear Algebra Homework Assignment - Week 2

INNER PRODUCT SPACE. Definition 1

The Singular Value Decomposition

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Stat 159/259: Linear Algebra Notes

Homework 11 Solutions. Math 110, Fall 2013.

Total 170. Name. Final Examination M340L-CS

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

8. Diagonalization.

ELE/MCE 503 Linear Algebra Facts Fall 2018

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of

Linear Algebra Solutions 1

Warm-up. True or false? Baby proof. 2. The system of normal equations for A x = y has solutions iff A x = y has solutions

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Math Linear Algebra

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Fundamentals of Engineering Analysis (650163)

STAT200C: Review of Linear Algebra

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Appendix A: Matrices

(v, w) = arccos( < v, w >

Section 6.2, 6.3 Orthogonal Sets, Orthogonal Projections

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Linear Algebra: Matrix Eigenvalue Problems

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

E2 212: Matrix Theory (Fall 2010) Solutions to Test - 1

Lec 2: Mathematical Economics

Section 6.4. The Gram Schmidt Process

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 8, 2012

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

5 Compact linear operators

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1 Inner Product and Orthogonality

6 Inner Product Spaces

Transcription:

34 To obtain an eigenvector x 2 0 2 for l 2 = 0, define: B 2 A - l 2 I 2 = È 1, 1, 1 Î 1-0 È 1, 0, 0 Î 1 = È 1, 1, 1 Î 1. To transform B 2 into an upper triangular matrix, subtract the first row of B 2 from the second row of B 2 and we obtain U 2 : U 2 È 1, 0, 0 1 Î. The first zero element of U 2 2 is u 22 = 0. Hence set 2 x2 = 1 and solve U 2 x 2 È = 1, 0, 1 È Î 0 Î Í 1 = È 0 Î 0 x 1 2 for x 1 2 = -1. Hence x 2 = [-1, 1] T is an eigenvector of A that corresponds to the eigenvalue l 2 = 0. Note that the two eigenvectors x 1 and x 2 that were constructed in Examples 2 and 3 above had the property. (139) x 1T x 2 = 0. This property turns out to be a general property of eigenvectors of a symmetric A that correspond to distinct eigenvalues as we shall see later. Problems: 17. Calculate the 2 eigenvalues l 1 and l 2 of A È 2 1 1 Î 2 eigenvectors x 1 and x 2 that correspond to l 1 and l 2. and calculate È 18. Calculate the 3 eigenvalues and eigenvectors of A 0 2 0 2 0 Í 1 Î 0 1 2. Hint: Taking into account the block diagonal structure of A, you may find the results of Problem 17 useful. (140) Definition: Two N dimensional vectors x and y are orthogonal or perpendicular iff x T y = 0. To see why x T y = 0 implies x and y are perpendicular, let x and y be two N dimensional vectors of unit length; i.e., 34

35 (141) x T N 2 x = S i=1xi = 1 and y T N 2 y = S i=1 yi = 1. Now look at the vector z x - y which has the same length as the line segment s which joins x to y. If N = 2, we have the following picture. x 2, y 2 y = (y 1,y 2 ) segment s joining x and y x = (x 1, x 2 ) segment z = x - y x 1, y 1 If x and y are perpendicular and of unit length, then by Pythagoras' Theorem in geometry, the length of the segment s (which is equal to the length of z) is 2 ; i.e., we have (142) z T z = x T x + y T y = 1 + 1 = 2. Now substitute z = x - y into (142) and simplify: 2 = (x - y) T (x - y) = x T x - x T y - y T x + y T y = 2 - x T y - y T x using (141) (143) = 2-2x T y since x T y = y T x. Equation (143) implies that x T y = 0. Thus if x and y are of unit length and are perpendicular, then x T y = 0. This argument can be extended to the case where x 0 N and y 0 N. (If x and y are perpendicular, then so are x/(x T x) 1/2 a and y/(y T y) 1/2 b. Now a and b are perpendicular and of unit length and so we can apply the above argument to conclude that a T b = 0. But this implies x T y = 0 as well). (144) Theorem: Suppose that the N eigenvalues l 1, l 2,..., l N of the N by N symmetric matrix A are all distinct. Let x 1, x 2,..., x N be eigenvectors corresponding to these distinct eigenvalues. Then x it x j = 0 for all i j; i.e., the eigenvectors x 1, x 2,..., x N are mutually orthogonal. 35

36 Proof: Let i j. By the definition of x i 0 N and l i being an eigenvalue and eigenvector of A, we have: (145) Ax i = l i x i and by the definition of l j and x j 0 N being an eigenvalue and eigenvector of A, we have: (146) Ax j = l j x j. Premultiply both sides of (145) by x jt and obtain: (147) x jt Ax i = l i x j T x i. Now take transposes of both sides of (147) and get: (148) l i x it x j = x it A T x j = x it Ax j where the second equality in (148) follows from A = A T. Premultiply both sides of (146) by x it and obtain: (149) l j x it x j = x it Ax j. Since the right hand sides of (148) and (149) are equal, so are the left hand sides, so we obtain: (150) l i x it x j = l j x it x j or (151) (l i - l j )x it x j = 0. Since l i l j by assumption, (151) implies that x it x j = 0; i.e., x i is orthogonal to x j. Q.E.D. 10. The Diagonalization of a Symmetric Matrix by an Orthonormal Transformation Suppose that A is an N by N symmetric matrix with distinct eigenvalues l 1, l 2,..., l N with corresponding nonzero eigenvectors x 1, x 2,.., x N. (We will deal with the case where the eigenvalues are not necessarily distinct later in this section). We normalize the eigenvectors x i 0 N so that they are of unit length; i.e., for i = 1, 2,..., N, define the normalized eigenvector u i by (152) u i = x i /(x it x i ) 1/2 ; i = 1, 2,..., N. 36

37 It can be seen that since Ax i = l i x i for i = 1,..., N, we also have: (153) Au i = l i u i ; i = 1, 2,..., N, and the u i also satisfy: (154) u it u i = [x i /(x it xi) 1/2 ] T [x i /(x it x i ) 1/2 ] = x it x i /(x it x i 1/2 + 1/2 ) = 1 for i = 1,..., N. From Theorem (144), we have x it x j = 0 if i j. This implies that the u i have the same orthogonality properties; i.e., we have: (155) u it u j = 0 if i j. Define the N by N matrix U of the normalized eigenvectors of A by: (156) U [u 1, u 2,..., u N ]. Using (154) and (155), it can be seen that (157) U T U = I N. But (157) tells us that U T is a left inverse (and hence is an inverse) for U; i.e., (158) U -1 = U T. Taking determinants of both sides of (157) yields: (158) 1 = I N = U T U = U T U = U 2, or (159) U = +1 or -1. A square matrix U that satisfies (157) is called an orthonormal matrix, and (159) shows us that its determinant equals +1 or -1. Return to the N eigenvalue, normalized eigenvector equations (153). Using definition (156), it can be seen that the N equations (153) can be rewritten as the following matrix equation: AU = [l 1 u 1, l 2 u 2,..., l N u N ] È l 1 0... 0 0 l 2... 0 = [u 1, u 2,..., u N Í ] Í :. :. :. Î Í 0 0... l N 37

38 (160) = UL where L is a diagonal matrix with the eigenvalues of A on the main diagonal of L. Now premultiply both sides of (160) by U T and get: (161) U T AU = U T UL = L using (157) We can now use the eigenvector-eigenvalue method for diagonalizing A, the matrix equation (161) above, in exactly the same way that we used the Gauss- Lagrange diagonalization method (59) above in order to determine the definiteness properties of A. For example, since U 0, it can be seen that the conditions for positive definiteness of A, (162) x T Ax > 0 for all x 0 N are equivalent to 0 < x T Ax, x 0 N = y T U T AUy, letting x =Uy so y 0 N = y T Ly, using (161) (163) N 2 = S i=1li y i. Thus A is positive (negative) definite iff all of the eigenvalues of A, l 1, l 2,..., l N, are positive (negative). Of course A is positive (negative) semidefinite iff all of the eigenvalues of A are nonnegative (nonpositive). However, we established the above results under the assumption that all of the eigenvalues of A were distinct. In the next section, we shall relax this assumption that the eigenvalues of A are distinct, but we will still get the same results as in the above paragraph. 11. The Diagonalization of a Symmetric Matrix in the General Case In order to derive the matrix equation (161) in the previous section without the assumption that the eigenvalues of the symmetric N by N matrix are distinct, we need two preliminary results. (164) The Gram-Schmidt Orthogonalization Procedure: Let the N dimensional vectors x 1, x 2,..., x K be linearly independent (so that K N). Then for k = 1, 2,...,K; each vector x K can be expressed as a linear combination of k orthonormal vectors y 1, y 2,..., y K ; i.e., for k = 1, 2,...,K: (165) x k k = S j=1 a kj y j where 38

39 (166) y it y j Ï 1 if i = j and = Ì Ó 0 if i j. The a kj in equations (165) are scalars. Proof: For k = 1, take y 1 x 1 /(x 1T x 1 ) 1/2. Note that x 1T x 1 > 0 since if x 1 = 0 N, then the x k would not be linearly independent. Thus we have x 1 = a 11 y 1 where a 11 (x 1T x 1 ) 1/2. For k = 2, define the vector z 2 as follows: (167) z 2 x 2 - (x 2T y 1 )y 1. Suppose z 2 = 0 N. Then x 2 = (x 2T y 1 )y 1 = x 1 (x 2T y 1 )/(x 1T x 1 ) 1/2 which would imply that x 2 is a multiple of x 1 and hence x 1 and x 2 would be linearly dependent. Thus our supposition is false and z 2 0 N. We show that z 2 is orthogonal to y 1 : (168) y 1T z 2 = y 1T [x 2 - (x 2T y 1 )y 1 ] = y 1T x 2 - (x 2T y 1 )y 1T y 1 = y 1T x 2 - (x 2T y 1 ) using y 1T y 1 = 1 = 0 since (x 2T y 1 ) T = y 1T x 2. Now define y 2 as the following normalization of z 2 : (169) y 2 z 2 /(z 2T z 2 ) 1/2 and so we have using (168) and (169)): (170) y 1T y 2 = 0, y 2T y 2 = 1 and y 1T y 1 = 1. If we substitute (169) into (167) and rearrange terms, we have: (171) x 2 = (x 2T y 1 )y 1 = (z 2T z 2 ) 1/2 y 2 a 21 y 1 + a 11 y 2 which is (165) for k = 2. For a general k, having defined y 1, y 2,..., y k-1, define z k as follows: (172) z k x k k -1 -S j=1 (x kt y j )y j. 39

40 If z k = 0 N, then we can show that x 1, x 2,..., x k are linearly dependent which contradicts our assumption. Thus z k 0 N. It is also straightforward to show that for l = 1, 2,..., k-1: (173) y lt z k = y lt [x k k -1 - S j=1 (x kt y j )y j ] = y lt x k k -1 - S j=1 (x kt y j )y lt y j = y lt x k - (x kt y l )y lt y l since y lt y j = 0 if l j = y lt x k - (x kt y k ) since y lt y l = 1 = 0. Thus z k is orthogonal to y l, y 2,..., y k-1. Now define y k as the following normalization of z k : (174) y k z k /(z kt z k ) 1/2. Now substitute (174) into (172) and we obtain: (175) x k k -1 = S j=1 (x kt y j )y j + (z kt z k ) 1/2 y k k -1 S j=1 akj y j + a kk y k. Q.E.D. The Gram-Schmidt orthogonalization procedure is very useful in econometrics. We need another preliminary result about orthonormal N by N matrices U (recall this means U T U = I N ). (176) Lemma: Let U 1 and U 2 be two N by N orthonormal matrices. Then the product matrix U U 1 U 2 is also orthonormal. Proof: U T U = (U 1 U 2 ) T (U 1 U 2 ) = U 2T U 1T U 1 U 2 = U 2T I N U 2 = I N. Q.E.D. Now we are ready to derive the diagonal decomposition of the N by N matrix A, U T AU = L where U T U = I N, assuming only that A is symmetric; i.e., we want to drop our hypothesis that we made in the previous section that the eigenvalues of A were distinct. Let A be N by N and symmetric and let l 1 be any eigenvalue of A with corresponding normalized eigenvector y 1 ; i.e., we have (177) Ay 1 = l 1 y 1 with y 1T y 1 = 1. 40