MATH 583A REVIEW SESSION #1

Similar documents
1. General Vector Spaces

Calculating determinants for larger matrices

235 Final exam review questions

Econ Slides from Lecture 7

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

Eigenvalues and Eigenvectors

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Linear Algebra: Matrix Eigenvalue Problems

Review of some mathematical tools

Linear Algebra Review

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Linear algebra and applications to graphs Part 1

4. Linear transformations as a vector space 17

Foundations of Matrix Analysis

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Review problems for MA 54, Fall 2004.

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Math Matrix Algebra

. = V c = V [x]v (5.1) c 1. c k

A Brief Outline of Math 355

Linear Algebra - Part II

Symmetric and anti symmetric matrices

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

ANSWERS. E k E 2 E 1 A = B

Linear Algebra Practice Problems

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

LINEAR ALGEBRA REVIEW

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Definitions for Quizzes

Solutions to Final Exam

Then since v is an eigenvector of T, we have (T λi)v = 0. Then

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Math Linear Algebra Final Exam Review Sheet

2. Every linear system with the same number of equations as unknowns has a unique solution.

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

NORMS ON SPACE OF MATRICES

Review of Linear Algebra

Eigenvectors and Hermitian Operators

Linear Algebra Exercises

MATHEMATICS 217 NOTES

Lecture Summaries for Linear Algebra M51A

Dimension. Eigenvalue and eigenvector

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Knowledge Discovery and Data Mining 1 (VO) ( )

Eigenvalues and Eigenvectors

Eigenvalues, Eigenvectors, and Diagonalization

Study Guide for Linear Algebra Exam 2

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

MTH 102: Linear Algebra Department of Mathematics and Statistics Indian Institute of Technology - Kanpur. Problem Set

Mathematical Methods wk 2: Linear Operators

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

Math 110 Linear Algebra Midterm 2 Review October 28, 2017

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

October 25, 2013 INNER PRODUCT SPACES

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

MATH 221, Spring Homework 10 Solutions

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Study Notes on Matrices & Determinants for GATE 2017

Linear Algebra Highlights

Linear Algebra. Min Yan

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Background Mathematics (2/2) 1. David Barber

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Online Exercises for Linear Algebra XM511

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Properties of Matrices and Operations on Matrices

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial.

Elementary linear algebra

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Definition (T -invariant subspace) Example. Example

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Review of linear algebra

Analysis Preliminary Exam Workshop: Hilbert Spaces

MATH 235. Final ANSWERS May 5, 2015

6 Inner Product Spaces

Transcription:

MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space), subspace, linear independence, basis, dimension. 1.1. Coordinatization. Choosing a basis {v 1,..., v n } for a vector space V enables us to express any vector x V in terms of the elements of the basis: n x = x i v i. i=1 Thus, we can identify the (abstract) vector x with a (concrete) column matrix [x 1,..., x n ] T, having coefficients in this expansion (coordinates) for elements. The set of all column matrices, denoted F n (F can be either R or C), forms a vector space which is isomorphic to the abstract vector space V, but we always need to keep in mind the basis with respect to which the coordinatization was carried out. (Coordinates always imply a system of reference!) There are many possible bases for any vector space, and the matrix representation will be different moving from one basis to another. However, it is easily verified that the transition from one basis to another translates into a linear transformation of the corresponding matrices. Let {u 1,..., u n } and {v 1,..., v n } be two bases for the vector space V, and let s ij be the coordinates of v j in the basis of u s, n v j = s ij u i i=1 The matrix S therefore has coordinates of v j s for columns. Let x V be any vector; we denote its coordinates in the u- and v-basis by x(u) = [ξ 1,..., ξ n ] T and x(v) = [η 1,..., η n ] T, respectively. It is easy to verify that n ξ i = s ij η j, j=1 i.e. S is the transition matrix from coordinates in the v-basis to coordinates in the u-basis: 1 x(u) = Sx(v) Date: September 1, 2006. 1 This is intuitively clear when you think of the multiplication Sx as a linear combination of columns of S with coefficients x i. Then x i are coordinates in the basis of columns of S, and the resulting product will be the coordinates in whatever basis the columns of S are expressed. 1

2 BOJAN DURICKOVIC The inverse transformation is obtained by applying the inverse matrix: x(v) = S 1 x(u) 2. Linear Operators A linear operator A : V W (V, W vector spaces) is a map that preserves sums and scalar multiples, i.e. A(x + y) = Ax + Ay A(αx) = αax x, y V, α F. The range (also called column space with matrices) of the operator is its range as a function, ran A = {Ax x V } W The kernel (or null space) of the operator is the subset of its domain that is mapped to the null vector, ker A = {x V Ax = 0} V The linearity of the operator ensures that both the range and the kernel are always subspaces of W and V, respectively. The dimensions of these subspaces are related in the following way. Proposition. dim(ran A) + dim(ker A) = dim V Note that the dimension of the image space W has no bearing in the relation. 2 The following nice property of linear operators is another direct consequence of their linearity: Proposition. A linear operator is uniquely determined by its action on a basis. In other words, given a basis {v 1,..., v n } of V, if we specify the images of the elements of the basis Av i = w i W, i = 1,..., n, the operator is completely determined, since this definition is extended by linearity to all vectors in V : Ax = A x i v i = x i Av i = x i w i A linear operator A : V V (from a vector space to itself!) is non-singular if it is invertible, i.e. if A 1 : V V such that AA 1 = A 1 A = I. We can tell whether an operator is non-singular by simply looking at its null space: Proposition. An operator A : V V is non-singular if and only if its kernel is trivial, i.e. ker A = {0}. 2 Think of it as if reducing the dimension of W will tend to reduce the dimension of ran A, while increasing the dimension of ker A.

MATH 583A REVIEW SESSION #1 3 2.1. Matrix Representation. Analogously to the identification of vectors and column matrices for a chosen basis, we can coordinatize a linear operator. Since a linear operator is uniquely determined by its action on the basis, its coordinate representation can be taken as the collection of n column vectors which are the coordinate representations of the vectors w i = Av i W, i = 1,..., n. Combining these n m-dimensional column vectors into a rectangular matrix proves to be a useful representation, since the action of the operator on a vector translates into matrix multiplication. Again, it should be stressed that although the space of linear operators V W is entirely isomorphic to the space of rectangular matrices of corresponding dimensions, the actual correspondence between an operator and a matrix always requires the knowledge of the bases both in V and W. Dealing with abstract operators has the advantage that we don t need to specify bases, while matrices are needed for any practical computation. The matrix representation of an operator will be different from one pair of bases to another, generally requiring two transformation matrices S and T, for the bases in V and W respectively. (Transformation matrices are always square, but here S and T may have different dimensions if dim V dim W.) Most frequently, we are dealing with operators from the vector space V to itself, and in that case there is only one basis involved: the transformation of coordinates is then performed by a similarity transformation A = S 1 AS The way to think of this transformation is the following: A and A are matrix representations of an operator in the basis u = {u 1,..., u n } and v = {v 1,..., v n }, respectively, and S is the transformation matrix from the basis v to the basis u, i.e. its columns are the coordinates of v j in the basis u. Reading S 1 AS from right to left, we have: (1) multiplication by S: transformation from basis v to basis u; (2) multiplication by A: performing the action of the operator in the basis u; (3) multiplication by S 1 : transforming back from basis u to basis v. Although the concrete matrix representation depends on the basis (or basis pair), the properties of the matrix that remain unchanged in a change of basis (i.e. that are invariant with respect to similarity transformations in case of square matrices) are worth identifying: these are the properties of the abstract operator itself. 2.2. Rank of an operator. The (column) rank of a matrix is the number of linearly independent columns. The rank does not change in similarity transformations (prove this!), thus it is a property of the abstract operator. Proposition. The row rank of a matrix is equal to its column rank. Proof. Exercise. A matrix is said to be (or to have) full rank if it has a maximal rank. An m n matrix, where m > n, is full rank if all its (n) columns are linearly independent. For square matrices, full rank is actually a characterization of non-singularity: Proposition. A square matrix is full rank if and only if it is non-singular. Full rank and non-singularity are also operator properties independent of the choice of basis. Other properties that don t change in similarity transformations

4 BOJAN DURICKOVIC include the determinant and trace. Therefore, although the determinant and trace are computed from a matrix representation, it makes perfect sense to speak of a determinant and trace of an abstract operator. 2.3. Characteristic polynomial. The characteristic polynomial of a square matrix A is det(a λi) The characteristic polynomial of an n n matrix is a polynomial of order n: c A (λ) = c n λ n + c n 1 λ n 1 + + c 0 Proposition. The characteristic polynomial of a matrix A has the following coefficients: c n = ( 1) n, c n 1 = ( 1) n 1 tr A, c 0 = det A The above proposition can be useful, in particular for quickly computing the characteristic polynomial of a 2 2 matrix: c A (λ) = λ 2 λ tr A + det A Using the Binet-Cauchy theorem (det(ab) = det A det B), it is easy to show the following: Proposition. Similar matrices share the same characteristic polynomial. Thus, we can speak of a characteristic polynomial of an operator, which is independent of the operator s particular matrix representation. Theorem (Hamilton-Cayley). Every square matrix cancels its own characteristic polynomial, i.e. c A (A) = 0 This is an important theorem that we will be using for computing the function of a matrix. 3. Spectral Theory 3.1. Eigenvalue problem. Given an operator A : V V, 3 the eigenvalue problem consists in finding scalars λ (eigenvalues) and non-zero vectors v (eigenvectors) such that Av = λv Such a pair (λ, v) is collectively termed an eigenpair. The set of all eigenvalues of A is called the spectrum of A (notation: σ(a)). Proposition. Eigenvectors corresponding to different eigenvalues are linearly independent. The existence of eigenvalues is guaranteed by the following theorem. 4 3 Note that the eigenvalue problem makes sense only for operators from the space V to itself, i.e. for square matrices! 4 Note, however, that, except for low dimensions, this is not how eigenvalues are actually computed in practice, due to the prohibitively high cost of computing a determinant. Usually, they are found using some matrix iteration scheme, until they start appearing on the diagonal.

MATH 583A REVIEW SESSION #1 5 Theorem. A scalar λ is an eigenvalue of the operator A if and only if λ is a root of the characteristic polynomial, Proof. c A (λ) = 0 λ eigenvalue (A λi)v = 0, v 0 det(a λi) = 0 c A (λ) = 0 Thus, a linear operator acting on an n-dimensional vector space has exactly n (not necessarily distinct) eigenvalues. The algebraic multiplicity of the eigenvalue is its multiplicity as the root of the characteristic polynomial. By definition, the existence of eigenvalues implies the existence of corresponding eigenvectors. Every eigenvector spans a one-dimensional invariant subspace. However, there may be several eigenvectors corresponding to one eigenvalue, spanning an invariant subspace of dimension greater than one. The dimension of this space is called the geometric multiplicity of the eigenvalue. Proposition. The geometric multiplicity of an eigenvalue is less than or equal to its algebraic multiplicity. Therefore, although there is at least one eigenvector for each eigenvalue, some eigenvalues with multiplicity larger than one may run short of eigenvectors to match their multiplicity. The eigenvalues for which the geometric multiplicity is less than the algebraic one are called defective. A defective operator/matrix is one having one or more defective eigenvalues. 3.2. Diagonalization. An operator is diagonalizable if there is a basis in which the operator is represented by a diagonal matrix. (A matrix is diagonalizable if it can be reduced to diagonal form by similarity transformations.) Theorem. An operator A : V V is diagonalizable if and only if there is a basis of V composed of eigenvectors of A. Proof. (sketch, one direction only) Let {v 1,..., v n } be a basis of eigenvectors of A. Combining the corresponding column vectors into a matrix S = [v 1,..., v n ] we have a full rank matrix, which is therefore invertible, and we got ourselves a similarity transformation S 1 AS. We claim that the result is a diagonal matrix. The matrix AS has for columns Av j = λ j v j, thus can be rewritten as AS = SΛ (operations on columns = multiplication from the right; columns are just multiplied by scalars = Λ is diagonal). Therefore S 1 AS = Λ = diag(λ 1,..., λ n ). A useful way to think of this is by looking at a transformation to the eigenvector basis: Ax = SΛS 1 x The action of A is therefore obtained by transforming to the eigenvector basis with S 1, in which the action consists simply in multiplying each coordinate with the

6 BOJAN DURICKOVIC corresponding eigenvalue (hence multiplication with a diagonal matrix), and then transforming back to the original basis with S. 5 An immediate consequence is that if all eigenvalues are distinct, the operator is diagonalizable, since there are n (linearly independent) eigenvectors. Note that defective operators are not diagonalizable they are short of eigenvectors. 3.3. Normal operators. A normal operator N : V V is an operator that commutes with its adjoint 6, NN = N N Proposition. Let N be a normal operator. If (λ, v) is an eigenpair of N, then (λ, v) is an eigenpair of N. Proof. Exercise. The most useful property of normal operators on complex vector spaces is given by the Spectral Theorem: 7 Theorem (Spectral Theorem for Normal Operators). An operator N : V V is normal if and only if there is an orthonormal basis of V composed of eigenvectors of N. Therefore, normal operators on complex spaces are diagonalizable. The situation is somewhat more complicated in the real case, since in that case the characteristic polynomial may have less than n real roots. This is not a concern with self-adjoint operators... 3.4. Self-adjoint operators. An operator L is self-adjoint (or Hermitian) if L = L. Self-adjoint operators are a special case of normal operators, hence they are diagonalizable. They also have the additional nice property that all their eigenvalues are real. This theorem is frequently used: Theorem (Spectral Theorem). Let L : V V be a self-adjoint operator. Then: (1) The eigenvalues of L are real. (2) There is an orthonormal basis of V composed of eigenvectors of L. 3.5. Computing the function of a matrix. Taylor series enable us to give meaning to a function of a square matrix, since we know how to compute powers of a matrix. If the matrix is in diagonal form, this task is very simple: Λ k = diag(λ k 1, λ k 2,..., λ k n), so that f(λ) = diag(f(λ 1 ), f(λ 2 ),..., f(λ n )). 5 If you are having problems remembering whether A is S 1 ΛS or SΛS 1, where the columns of S are eigenvectors, remember that you need to have SΛ in there, since you need to multiply eigenvectors with corresponding eigenvalues, and columns are scaled by multiplying from the right with a diagonal matrix. 6 The adjoint A (sometimes also denoted A ) of a matrix A is its complex conjugate transpose, A = A T. 7 It is called Spectral Theorem since it allows us to decompose the entire space V into mutually orthogonal invariant subspaces of N. The words spectral, invariant subspace and the prefix eigen often denote the same context. It s all about the eigenvalue problem.

MATH 583A REVIEW SESSION #1 7 But how to compute the function of a non-diagonal matrix? The obvious approach is diagonalizing (assuming the matrix is diagonalizable), computing the function of a diagonal form, and then transforming back to the original basis. However, there is a way of computing this without performing the change of basis. To do this, we don t need the actual eigenvectors of the matrix, but only its eigenvalues. Consider the exponential function to be specific. Its Taylor series converges on the whole set of real numbers: e x = k=0 x k k!, x R. The exponential of a matrix is thus defined as e A A k = k!. k=0 We can compute this infinite series in a finite number of operations by using the Hamilton-Cayley theorem: since the matrix A cancels its characteristic polynomial, its nth power can be expressed in terms of lower powers: 8 n 1 A n = c j A j The series thus reduces to a finite sum: n 1 e A = ξ k A k = ξ n 1 A n 1 + + ξ 1 A + ξ 0 I, k=0 j=0 where ξ k are some coefficients to be determined. These are found by applying the equation to all n eigenvectors, yielding a linear system for ξ 0,..., ξ n 1 involving the n eigenvalues: e λ1 e λ2 e λn. = ξ n 1 λ n 1 1 + + ξ 1 λ 1 + ξ 0 = ξ n 1 λ n 1 2 + + ξ 1 λ 2 + ξ 0 = ξ n 1 λ n 1 n + + ξ 1 λ n + ξ 0 You should recognize the Vandermonde matrix. Its determinant will be nonzero if all eigenvalues are different; for now our method is limited to this case. Then we have a unique solution for ξ 0,..., ξ n 1. We have thus found e A using nothing but the eigenvalues of A, and all that needed to be computed are the first n 1 powers of A and the solution of one n n system. Exercise. Compute the exponential of the matrix [ ] 2 1 A = 1 2 (a) by diagonalizing and undiagonalizing; (b) by the above method. 8 The coefficients cj here coincinde possibly up to a sign with the coefficients of the characteristic polynomial: remember that the leading coefficient of c A is ( 1) n.