Matrices and Linear transformations

Similar documents
1 Last time: multiplying vectors matrices

Eigenvalues, Eigenvectors, and Diagonalization

Eigenvalues and Eigenvectors

Matrices related to linear transformations

k is a product of elementary matrices.

A Note on the Eigenvalues and Eigenvectors of Leslie matrices. Ralph Howard Department of Mathematics University of South Carolina

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Usually, when we first formulate a problem in mathematics, we use the most familiar

Solutions to Final Exam

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Announcements Monday, October 29

Chapter 4 & 5: Vector Spaces & Linear Transformations

Topics in linear algebra

Linear Algebra (MATH ) Spring 2011 Final Exam Practice Problem Solutions

From Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play?

Vector Spaces and Linear Transformations

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Linear algebra and differential equations (Math 54): Lecture 13

22 Approximations - the method of least squares (1)

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

1 Last time: inverses

Study Guide for Linear Algebra Exam 2

21 Symmetric and skew-symmetric matrices

MAT224 Practice Exercises - Week 7.5 Fall Comments

( 9x + 3y. y 3y = (λ 9)x 3x + y = λy 9x + 3y = 3λy 9x + (λ 9)x = λ(λ 9)x. (λ 2 10λ)x = 0

MAT 1302B Mathematical Methods II

MATH 310, REVIEW SHEET 2

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

EXAM 2 REVIEW DAVID SEAL

8 Extremal Values & Higher Derivatives

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

Eigenspaces and Diagonalizable Transformations

Generalized Eigenvectors and Jordan Form

Practice Final Exam Solutions

ENGINEERING MATH 1 Fall 2009 VECTOR SPACES

3.2 Gaussian Elimination (and triangular matrices)

Linear Algebra, Summer 2011, pt. 2

REVIEW OF DIFFERENTIAL CALCULUS

MATH 115A: SAMPLE FINAL SOLUTIONS

Solutions to Math 51 First Exam October 13, 2015

Chapter 5. Eigenvalues and Eigenvectors

Diagonalization of Matrices

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

12. Perturbed Matrices

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Math 396. Quotient spaces

CSL361 Problem set 4: Basic linear algebra

Numerical Linear Algebra Homework Assignment - Week 2

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Numerical Optimization

0.1 Rational Canonical Forms

Linear Algebra- Final Exam Review

Eigenvalues and Eigenvectors 7.1 Eigenvalues and Eigenvecto

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

Chapter 3. Vector spaces

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Matrices and Linear Algebra

Properties of Matrices and Operations on Matrices

Contents. 1 Vectors, Lines and Planes 1. 2 Gaussian Elimination Matrices Vector Spaces and Subspaces 124

Math Final December 2006 C. Robinson

MATH 205 HOMEWORK #3 OFFICIAL SOLUTION. Problem 1: Find all eigenvalues and eigenvectors of the following linear transformations. (a) F = R, V = R 3,

Algebra II. Paulius Drungilas and Jonas Jankauskas

SECTION 3.3. PROBLEM 22. The null space of a matrix A is: N(A) = {X : AX = 0}. Here are the calculations of AX for X = a,b,c,d, and e. =

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Differential Equations

1 Inner Product and Orthogonality

MATH 221, Spring Homework 10 Solutions

MATH 431: FIRST MIDTERM. Thursday, October 3, 2013.

Notes on Linear Algebra I. # 1

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Eigenvalues and eigenvectors

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

Chapters 5 & 6: Theory Review: Solutions Math 308 F Spring 2015

MATH 1553-C MIDTERM EXAMINATION 3

= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E.

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

Linear Algebra March 16, 2019

Chapter 1: Linear Equations

MATH 583A REVIEW SESSION #1

Math 344 Lecture # Linear Systems

Name Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise.

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

MAT Linear Algebra Collection of sample exams

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

MATH 310, REVIEW SHEET

UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013

Chapter 1: Linear Equations

MAT1302F Mathematical Methods II Lecture 19

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

LECTURES 14/15: LINEAR INDEPENDENCE AND BASES

Lecture 22: Section 4.7

Diagonalization. Hung-yi Lee

Spectral Theorem for Self-adjoint Linear Operators

Recall : Eigenvalues and Eigenvectors

Announcements Monday, November 20

Notes on multivariable calculus

Transcription:

Matrices and Linear transformations We have been thinking of matrices in connection with solutions to linear systems of equations like Ax = b. It is time to broaden our horizons a bit and start thinking of matrices as functions. In particular, if A is m n, we can use A to define a function f A from R n to R m which sends v R n to Av R m. That is, f A (v) = Av. If then Let A 2 3 = 2 3 4 5 6 x v = y z R 3,. x 2 3 x + 2y + 3z f A (v) = Av = y = 4 5 6 4x + 5y + 6z z sends the vector v R 3 to Av R 2. Definition: A function f : R n R m is said to be linear if f(v + v 2 ) = f(v ) + f(v 2 ), and f(cv) = cf(v) for all v, v 2 R n and for all scalars c. A linear function f is also known as a linear transformation.

Examples: Define f : R 3 R by f x y z = 3x 2y + z. Then f is linear because for any v = x y z, and v 2 = x 2 y 2 z 2, we have f(v + v 2 ) = f x + x 2 y + y 2 = 3(x + x 2 ) 2(y + y 2 ) + (z + z 2 ). z + z 2 And the right hand side can be rewritten as (3x 2y +z )+(3x 2 2y 2 +z 2 ), which is the same as f(v )+f(v 2. So the first property holds. So does the second, since f(cv) = 3cx 2cy + cz = c(3x 2y + z) = cf(v). Notice that the function f is actually f A for the right A: if A 3 = (3, 2, ), then f(v) = Av. If A m n is a matrix, then f A : R n R m is a linear transformation because f A (v + v 2 ) = A(v +v 2 ) = Av +Av 2 (this is a fundamental property of matrix multiplication) = f A (v ) + f A (v 2 ). And A(cv) = cav f A (cv) = cf A (v). Although we don t give the proof, it can be shown that any linear transformation can be written as f A for a suitable matrix A. 2

There are many other examples of linear transformations; some of the most interesting do not go from R n to R m :. If f and g are differentiable functions, then d df (f + g) = dx dx + dg dx, and The function D(f) = df/dx is linear. d df (cf) = c dx dx. 2. If f is continuous, then we can define If(x) = x 0 f(s) ds, and I is linear, by well-known properties of the integral. 3. The Laplace operator,, defined before, is linear. 4. Let y be twice continuously differentiable and define L(y) = y 2y 3y. Then L is linear, as you can (and should!) verify. Linear transformations on functions, like the above, are generally known as linear operators. They re a bit more complicated than matrix multiplication operators, but they have the same essential property of linearity. Exercises:. Give an example of a function from R 2 to itself which is not linear. 2. Identify all the linear transformations from R to R. 3

3. If f : R n R m is linear then Ker(f) := {v R n such that f(v) = 0} is a subspace of R n, called the kernel of f. 4. If f : R n R m is linear, then Range(f) = {y R m such that y = f(v) for some v} is a subspace of R m called the range of f. Remark: Everything we ve been doing regarding the solution of linear systems of equations can be recast in the framework of linear transformations. In particular, if f A is multiplication by some matrix A, then the range of f A is just the set of all y such that the linear system Av = y has a solution. And the kernel of f A is the set of all solutions to the homogeneous equation Av = 0. We now want to study square matrices, regarding an n n matrix A as a linear transformation from R n to itself. We ll just write Av for f A (v) to simplify the notation, and to keep things really simple, we ll just talk about 2 2 matrices -- all the problems that exist in higher dimensions are present in R 2. There are several questions that present themselves: Can we visualize the linear transformation x Ax? One thing we can t do is draw a graph! Why not? Connected with the first question is: can we choose a better coordinate system in which to view the problem? 4

The answer is not an unequivocal "yes" to either of these, but we can generally do some useful things. To pick up at the end of the last lecture, note that when we write f A (v) = y = Av, we are actually using the coordinate vector of v in the standard basis. Suppose we change the basis to some other basis {e, e 2 } using the invertible matrix E. Then we can rewrite the equation in the new coordinates and basis: We have v = Ev e, and y = Ey e, so y = Av Ey e = AEv e, and y e = E AEv e That is, the matrix equation y = Av is given in the new basis by the equation y e = E AEv e or y e = A e v e, where A e = E AE. We can now restate the second question: Can we find a nonsingular matrix E so that E AE is particularly "nice"? The primary example of such a "nice" matrix is a diagonal matrix. Definition: The matrix A is diagonal if the only nonzero entries lie on the main diagonal. That is, a ij = 0 if i j. is diagonal. A = 4 0 0 3 This is nice because we can (partially) visualize the linear transformation corresponding to multiplication by A: a vector v lying along the first coordinate 5

axis mapped to 4v, a multiple of itself. A vector w lying along the second coordinate axis is also mapped to a multiple of itself: Aw = 3w. It s length is tripled, and its direction is reversed. Now it turns out that we can find vectors like v and w, which are mapped to multiples of themselves, without first finding the matrix E. This is the subject of the next few sections. Eigenvalues and eigenvectors Definitions: If a vector v 0 satisfies the equation Av = λv, for some real number λ, then λ is said to be an eigenvalue of the matrix A, and v is said to be an eigenvector of A corresponding to λ. then If A = 2 3 3 2 Av =, and v = 5 5 = 5v., So λ = 5 is an eigenvalue of A, and v an eigenvector corresponding to this eigenvalue. Remark: Note that the definition of eigenvector requires that v 0. The reason for this is that if v = 0 were allowed, then any number λ would be an eigenvalue since the statement A0 = λ0 holds for any λ. On the other hand, we can have λ = 0. See the exercise below. 6

Exercises: Show that is also an eigenvector of the matrix A above. What s the eigenvalue? Eigenvectors are not unique. In particular, if v is an eigenvector for A, then so is cv, for any real number c 0. Suppose λ is an eigenvalue of A. Define E λ = {v R n such that Av = λv} Show that E λ is a subspace of R n. (N.b: the definition of E λ does not require v to be an eigenvector of A, so v = 0 is allowed; otherwise, it wouldn t be a subspace.) E 0 = Ker(f A ) is just the null space of the matrix A. The matrix A = 0 0 = cos(π/2) sin(π/2) sin(π/2) cos(π/2) represents a counterclockwise rotation through the angle π/2. Apart from 0, there is no vector which is mapped by A to a multiple of itself. So not every matrix has eigenvectors. 7