Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

Similar documents
Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

(here, we allow β j = 0). Generally, vectors are represented as column vectors, not row vectors. β 1. β n. crd V (x) = R n

The following definition is fundamental.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Vector Spaces and Linear Transformations

Math 108b: Notes on the Spectral Theorem

1. General Vector Spaces

Linear Algebra. Session 12

235 Final exam review questions

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Lecture 7: Positive Semidefinite Matrices

NATIONAL UNIVERSITY OF SINGAPORE DEPARTMENT OF MATHEMATICS SEMESTER 2 EXAMINATION, AY 2010/2011. Linear Algebra II. May 2011 Time allowed :

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

Review problems for MA 54, Fall 2004.

Chap 3. Linear Algebra

Further Mathematical Methods (Linear Algebra)

Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, f(x + h) (f(x) + ah) = a = 0 = 0. f(x + h) (f(x) + T x (h))

UNIT 6: The singular value decomposition.

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Lecture 8 : Eigenvalues and Eigenvectors

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

The Singular Value Decomposition

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Definition 3 A Hamel basis (often just called a basis) of a vector space X is a linearly independent set of vectors in X that spans X.

Quadratic forms. Defn

Math Linear Algebra II. 1. Inner Products and Norms

Eigenvalues and Eigenvectors A =

Chapter 4 Euclid Space

A Review of Linear Algebra

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

REPRESENTATION THEORY WEEK 7

The Spectral Theorem for normal linear maps

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Definitions for Quizzes

Linear algebra and applications to graphs Part 1

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

Elementary linear algebra

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

Practice Exam. 2x 1 + 4x 2 + 2x 3 = 4 x 1 + 2x 2 + 3x 3 = 1 2x 1 + 3x 2 + 4x 3 = 5

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Linear Algebra Massoud Malek

Linear Algebra Final Exam Study Guide Solutions Fall 2012

Quantum Computing Lecture 2. Review of Linear Algebra

Linear Algebra: Matrix Eigenvalue Problems

MATH 220 FINAL EXAMINATION December 13, Name ID # Section #

October 25, 2013 INNER PRODUCT SPACES

Algebra II. Paulius Drungilas and Jonas Jankauskas

Mathematical foundations - linear algebra

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

LINEAR ALGEBRA REVIEW

Linear Algebra 2 Spectral Notes

Chapter 6 Inner product spaces

Final Exam. Linear Algebra Summer 2011 Math S2010X (3) Corrin Clarkson. August 10th, Solutions

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

EIGENVALUES AND EIGENVECTORS 3

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

Linear Algebra Highlights

Linear Algebra Review. Vectors

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

MTH 2032 SemesterII

Lecture 9. Econ August 20

Linear Algebra- Final Exam Review

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Linear Algebra Lecture Notes-II

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

I. Multiple Choice Questions (Answer any eight)

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

LINEAR ALGEBRA SUMMARY SHEET.

MATH 369 Linear Algebra

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

MA 265 FINAL EXAM Fall 2012

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

Diagonalization of Matrix

Symmetric and self-adjoint matrices

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Optimization Theory. A Concise Introduction. Jiongmin Yong

NORMS ON SPACE OF MATRICES

MAT Linear Algebra Collection of sample exams

TBP MATH33A Review Sheet. November 24, 2018

INNER PRODUCT SPACE. Definition 1

Week Quadratic forms. Principal axes theorem. Text reference: this material corresponds to parts of sections 5.5, 8.2,


Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

Math Solutions to homework 5

Linear algebra 2. Yoav Zemel. March 1, 2012

Math Linear Algebra Final Exam Review Sheet

Review of some mathematical tools

Mathematical foundations - linear algebra

Check that your exam contains 30 multiple-choice questions, numbered sequentially.

Transcription:

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Diagonalization of Symmetric Real Matrices (from Handout Definition 1 Let δ ij = { 1 if i = j 0 if i j A basis V = {v 1,..., v n } of R n is orthonormal if v i v j = δ ij. In other words, a basis is orthonormal if each basis element has unit length ( v i 2 = v i v i = 1 for each i, and distinct basis elements are perpendicular (v i v j = 0 for i j. Remark: Suppose that x = n α j v j where {v 1,..., v n } is an orthonormal basis of R n. Then x v k = α j v j v k = = = α k α j (v j v k α j δ jk so x = (x v j v j Example: The standard basis of R n is orthonormal. Recall that for a real n m matrix A, A denotes the transpose of A: the (i, j th entry of A is the (j, i th entry of A. So the i th row of A is the i th column of A. Definition 2 A real n n matrix A is unitary if A = A 1. Theorem 3 A real n n matrix A is unitary if and only if the columns of A are orthonormal. Proof: Let v j denote the j th column of A. A = A 1 A A = I v i v j = δ ij {v 1,..., v n } is orthonormal 1

If A is unitary, let V be the set of columns of A and W be the standard basis of R n. Since A is unitary, it is invertible, so V is a basis of R n. A = A 1 = Mtx V,W (id Since V is orthonormal, the transformation between bases W and V preserves all geometry, including lengths and angles. Theorem 4 Let T L(R n,r n and W be the standard basis of R n. Suppose that Mtx W (T is symmetric. Then the eigenvectors of T are all real, and there is an orthonormal basis V = {v 1,..., v n } of R n consisting of eigenvectors of T, so that Mtx W (T is diagonalizable: Mtx W (T = Mtx W,V (id Mtx V (T Mtx V,W (id where Mtx V T is diagonal and the change of basis matrices Mtx V,W (id and Mtx W,V (id are unitary. Proof: (Sketch The proof of the theorem requires a lengthy digression into the linear algebra of complex vector spaces. Here is a very brief outline. 1. Let M = Mtx W (T. 2. The inner product in C n is defined as follows: x y = x j y j where c denotes the complex conjugate of any c C; note that this implies that x y = y x. The usual inner product in R n is the restriction of this inner product on C n to R n. 3. Given any complex matrix A, define A to be the matrix whose (i, j th entry is a ji ; in other words, A is formed by taking the complex conjugate of each element of the transpose of A. It is easy to verify that given x, y C n and a complex n n matrix A, Ax y = x A y. Since M is real and symmetric, M = M. 4. If M is real and symmetric, and λ C is an eigenvalue of M, with eigenvector x C n, then λ x 2 = λ(x x = (λx x = (Mx x = x (M x 2

= x (Mx = x (λx = (λx x = λ(x x = λ x 2 = λ x 2 which proves that λ = λ, hence λ R. 5. If M is real (not necessarily symmetric and λ R is an eigenvalue, then det(m λi = 0 v R n s.t. (M λiv = 0, so there is at least one real eigenvector. Symmetry implies that, if λ has multiplicity m, there are m independent real eigenvectors corresponding to λ (but unfortunately we don t have time to show this. Thus, there is a basis of eigenvectors, hence M is diagonalizable over R. 6. If M is real and symmetric, eigenvectors corresponding to distinct eigenvalues are orthogonal: Suppose that Mx = λx and My = ρy with ρ λ. Then λ(x y = (λx y = (Mx y = (Mx y = ( x M y = ( x M y = x (My = x (ρy = x (ρy = ρ(x y so (λ ρ(x y = 0; since λ ρ 0, we must have x y = 0. 7. Using the Gram-Schmidt method, we can get an orthonormal basis of eigenvectors: Let X λ = {x R n : Mx = λx}, the set of all eigenvectors corresponding to λ. Notice that if Mx = λx and My = λy, then M(αx + βy = αmx + βmy = αλx + βλy = λ(αx + βy so X λ is a vector subspace. Thus, given any basis of X λ, we wish to find an orthonormal basis of X λ ; all elements of this orthonormal basis will be eigenvectors corresponding to λ. Suppose X λ is m-dimensional and we are given independent vectors x 1,..., x m X λ. The Gram-Schmidt method finds an orthonormal basis {v 1,..., v m } for X λ. Let v 1 = x 1 x 1. Note that v 1 = 1. 3

Suppose we have found an orthonormal set {v 1,..., v k } such that span {v 1,..., v k } = span {x 1,..., x k }, with k < m. Let k y k+1 = x k+1 (x k+1 v j v j, v k+1 = y k+1 y k+1 span {v 1,..., v k+1 } = span {v 1,..., v k, v k+1 } = span {v 1,..., v k, y k+1 } = span {v 1,..., v k, x k+1 } = span {x 1,..., x k, x k+1 } For i = 1,..., k, k y k+1 v i = x k+1 (x k+1 v j v j v i K = x k+1 v i (x k+1 v j (v j v i K = x k+1 v i (x k+1 v j δ ij = x k+1 v i x k+1 v i = 0 v k+1 v i = y k+1 v i y k+1 0 = y k+1 = 0 v k+1 = y k+1 y k+1 = 1 Application to Quadratic Forms Consider a quadratic form f(x 1,..., x n = α ii x 2 i + β ij x i x j (1 i=1 i<j Let α ij = { βij 2 if i < j β ji 2 if i > j 4

Let so α 11 α 1n A =..... α n1 α nn f(x = x Ax Example: Let Let so A is symmetric and f(x = αx 2 1 + βx 1 x 2 + γx 2 2 A = ( α β 2 β 2 γ ( ( α β x1 (x 1, x 2 2 β γ x 2 2 ( αx1 + β = (x 1, x 2 x 2 2 β x 2 1 + γx 2 = αx 2 1 + βx 1 x 2 + γx 2 2 = f(x Returning to the general quadratic form in Equation (1, A is symmetric, so let V = {v 1,..., v n } be an orthonormal basis of eigenvectors of A with corresponding eigenvalues λ 1,..., λ n. A = U DU λ 1 0 0 0 λ 2 0 D =...... 0 0 λ n U = Mtx V,W (id is unitary The columns of U (the rows of U are the coordinates of v 1,..., v n, expressed in terms of the standard basis W. Given x R n, recall x = γ i v i where γ i = x v i i=1 f(x = f ( γi v i = ( γi v i A ( γi v i 5

The equation for the level sets of f is = ( γi v i U DU ( γi v i = ( U γ i v i D ( U γi v i = ( γi Uv i D ( γi Uv i = (γ 1,..., γ n D = λ i γ 2 i λ i γi 2 = C i=1 γ 1. γ n If λ i 0 for all i, the level set is an ellipsoid, with principal axes in the directions v 1,..., v n. The length of the principal axis along v i is C/λ i if C 0 (if λ i = 0, the level set is a degenerate ellipsoid with principal axis of infinite length in that direction. The level set is empty if C < 0. See Figure 1. If λ i 0 for all i, the level set is an ellipsoid, with principal axes in the directions v 1,..., v n. The length of the principal axis along v i is C/λ i if C 0 (if λ i = 0, the level set is a degenerate ellipsoid with principal axis of infinite length in that direction. The level set is empty if C > 0. If λ i > 0 for some i and λ j < 0 for some j, the level set is a hyperboloid. For example, suppose n = 2, λ 1 > 0, λ 2 < 0. The equation is This is a hyperbola with asymptotes C = λ 1 γ1 2 + λ 2 γ2 2 ( ( = λ 1 γ 1 + λ 2 γ 2 λ 1 γ 1 λ 2 γ 2 0 = λ 1 γ 1 + λ 2 γ 2 λ2 γ 1 = γ 2 λ 1 ( 0 = λ 1 γ 1 λ 2 γ 2 γ 1 = λ2 λ 1 γ 2 See Figure 2. This proves the following corollary of Theorem 4. 6

Corollary 5 Consider the quadratic form (1. 1. f has a global minimum at 0 if and only if λ i 0 for all i; the level sets of f are ellipsoids with principal axes aligned with the orthonormal eigenvectors v 1,..., v n. 2. f has a global maximum at 0 if and only if λ i 0 for all i; the level sets of f are ellipsoids with principal axes aligned with the orthonormal eigenvectors v 1,..., v n. 3. If λ i < 0 for some i and λ j > 0 for some j, then f has a saddle point at 0; the level sets of f are hyperboloids with principal axes aligned with the orthonormal eigenvectors v 1,..., v n. Section 3.4. Linear Maps between Normed Spaces Definition 6 Suppose X, Y are normed vector spaces and T L(X, Y. We say T is bounded if β R s.t. T(x Y β x X x X Note this implies that T is Lipschitz with constant β. Theorem 7 (Thms. 4.1, 4.3 Let X, Y be normed vector spaces and T L(X, Y. Then T is continuous at some point x 0 X T is continuous at every x X T is uniformly continuous on X T is Lipschitz T is bounded Proof: Suppose T is continuous at x 0. Fix ε > 0. Then there exists δ > 0 such that z x 0 < δ T(z T(x 0 < ε Now suppose x is any element of X. If y x < δ, let z = y x+x 0, so z x 0 = y x < δ. T(y T(x = T(y x = T(y x + x 0 x 0 = T(z T(x 0 < ε which proves that T is continuous at every x, and uniformly continuous. 7

We claim that T is bounded if and only if T is continuous at 0. Suppose T is not bounded. Then {x n } s.t. T(x n > n x n n Note that x n 0. Let ε = 1. Fix δ > 0 and choose n such that 1 < δ. Let n x x n n = n x n x n = x n n x n T(x n T(0 = 1 n < δ = T(x n 1 = n x n T(x n > n x n n x n = 1 = ε Since this is true for every δ, T is not continuous at 0. Therefore, T continuous at 0 implies T is bounded. Now, suppose T is bounded, so find M such that T(x M x for every x X. Given ε > 0, let δ = ε/m. Then x 0 < δ x < δ T(x T(0 = T(x < Mδ T(x T(0 < ε so T is continuous at 0. Thus, we have shown that continuity at some point x 0 implies uniform continuity, which implies continuity at every point, which implies T is continuous at 0, which implies that T is bounded, which implies that T is continuous at 0, which implies that T is continuous at some x 0, so all of the statements except possibly the Lipschitz statement are equivalent. Suppose T is bounded, with constant M. Then T(x T(y = T(x y M x y so T is Lipschitz with constant M; conversely, if T is Lipschitz with constant M, then T is bounded with constant M. So all the statements are equivalent. Every linear map on a finite-dimensional normed vector space is bounded (and thus continuous, uniformly continuous, and Lipschitz continuous. 8

Theorem 8 (Thm. 4.5 Let X, Y be normed vector spaces with dim X <. Every T L(X, Y is bounded. Proof: See de la Fuente. Definition 9 A topological isomorphism between normed vector spaces X and Y is a linear transformation T L(X, Y that is invertible (one-to-one, onto, continuous, and has a continuous inverse. Two normed vector spaces X and Y are topologically isomorphic if there is a topological isomorphism T : X Y. Suppose X and Y are normed vector spaces. We define B(X, Y = {T L(X, Y : T is bounded} { } T(x Y T B(X,Y = sup, x X, x 0 x X = sup{ T(x Y : x X = 1} Theorem 10 (Thm. 4.8 Let X, Y be normed vector spaces. Then ( B(X, Y, B(X,Y is a normed vector space. Proof: See de la Fuente. Theorem 11 (Thm. 4.9 Let T L(R n,r m (= B(R n,r m with matrix A = (a ij with respect to the standard bases. Let M = max{ a ij : 1 i m, 1 j n} Then M T M mn Proof: See de la Fuente. Theorem 12 (Thm. 4.10 Let R L(R m,r n and S L(R n,r p. Then S R S R 9

Proof: See de la Fuente. Define Ω(R n = {T L(R n,r n : T is invertible} Theorem 13 (Thm. 4.11 Suppose T L(R n,r n and E is the standard basis of R n. Then T is invertible kert = {0} det(mtx E (T 0 det(mtx V,V (T 0 for every basis V det(mtx V,W (T 0 for every pair of bases V, W Theorem 14 (Thm. 4.12 If S, T Ω(R n, then S T Ω(R n and (S T 1 = T 1 S 1 Theorem 15 (Thm. 4.14 Let S, T L(R n,r n. If T is invertible and T S < 1 T 1 then S is invertible. In particular, Ω(R n is open in L(R n,r n = B(R n,r n. Proof: See de la Fuente. Theorem 16 (4.15 The function ( 1 : Ω(R n Ω(R n that assigns T 1 to each T Ω(R n is continuous. Proof: See de la Fuente. 10

Figure 1: If λ 1, λ 2 > 0 and C > 0, the level set is an ellipsoid, with principal axes in the directions v 1, v 2. The length of the principal axis along v i is C/λ i. 11

Figure 2: If λ 1 > 0 and λ 2 < 0, the level set is a hyperbola with asymptotes γ 1 = λ2 λ 1 γ 2. 12