Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples

Size: px
Start display at page:

Download "Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples"

Transcription

1 Math b TA: Padraic Bartlett Recitation 9: Probability Matrices and Real Symmetric Matrices Week 9 Caltech 20 Random Question Show that = ϕ, the golden ratio, which is = Homework comments Section average: 72/80 = 90%. This was about 2 3% above the class average; nice work! As the above comment suggests, people did pretty well! There isn t much to say, other than please include explanations of all the things you do. Today, our talk is pretty much split into two topics: probability matrices and real symmetric matrices. For each subject, we ll include examples of both the calculations you ll be expected to do on the HW, as well as proofs that you might be asked to understand or replicate. 3 Probability Matrices: Definitions and Examples Definition. A n n matrix P is called a probability matrix if and only if the following two properties are satisfied: P 0; in other words, p ij 0 for every entry p ij of P. The column sums of P are all ; in other words, n i= p ij =, for every j. Why do we call these kinds of matrices probabilitiy matrices? Well, consider the following way of interpreting such a matrix P : Suppose that you have an object that is capable of being in n different states. For example, if the object you re modeling is a Caltech undergrad, you could roughly model them as having the following three distinct states: {sleeping, doing sets, eating}

2 Suppose furthermore that at the end of every hour ( or more generally, at the end of every step, where your step size is some arbitrary unit), the undergrad you re studying has certain probabilities of switching from the state they re in into a different state. I.e. you could assume that if your undergrad is working on a set for a hour, there s a decent chance (30%) that they pass out because it s difficult, a better chance (0%) that they keep working on the set, and a small chance (0%) that they get hungry and stop for a sandwich. If you have this information for all of your states, you can create a probability matrix with this data! To do this, let p ij contain the probability of switching from state j to state i. Then, we have a matrix all of whose columns sum to (this is because your undergrad is always in *one* of these three states, and thus when they leave any state j the sum of their chances of landing in the other states must be.) In other words: a probability matrix! As it turns out, this method of describing a finite-state system isn t just notationally useful: we in fact have a pair of remarkably useful theorems for probability matrices! 4 Probability Matrices: Two Useful Theorems Theorem If we have a probability matrix P representing some finite system with n states {,... n}, then the probability of starting in state j and ending in state i in precisely m steps is the (i, j)-th entry in P m. Proof. Last recitation, we proved that for a n n matrix P, the (i, j)-th entry of P m can be written as the sum n c,...c m = p i,c p c,c 2,... p cm 2,c m p cm,j But what *is* one of these terms p i,c p c,c 2,... p cm 2,c m p cm,j? Well, if we read it from right to left, it s just the product (probability of going from j to c m ) (probability of going from c m to c m 2 )... (probability of going from c to i). If we just multiply everything out, we can see that this is just the probability that we take the specific path j c m... i from j to i! And our sum has one term for every such path from i to j: therefore, because our sum is just the sum of all of these probabilities, it represents the chances of us taking any of these paths from i to j. Therefore, the (i, j)-th entry represents the probability of us going from j to i in m steps along any of these paths which is what we claimed. An example of this theorem follows below: 2

3 Example. Suppose we have a Caltech student capable of entering three possible states {sleeping, sets 2, eating 3 }, and suppose it switches between these three states every hour according to the following probability matrix: P = (I.e. if our student is asleep at noon, it has a 40% chance of staying asleep at, a 50% chance of waking up and starting a set, and a 0% chance of deciding it s hungry for a kebab.) If your student is asleep at 2am, what are the chances that it will be working on a set at 4am? Solution. So: by the theorem above, we just need to look at the (2, ) cell in P 2 to figure this out. Calculating gives us that P 2 = and thus that our poor student has a 2% chance of working on their sets at 4am. This theorem allows us to say what happens at specific points and times in the future. However: what if we re not interested in specific points and times in the future, but rather in the long-term behavior of the system as a whole? The following two definitions and pair of theorems tell us what to do in that situation: Definition. A vector v R n is called a probability vector if and only if n i= v i =. Definition. For a probability matrix P, v is called a stable vector if and only if v is a probability vector that is also an eigenvector of P with corresponding eigenvalue. Theorem 2 Every probability matrix has at least one stable vector. Theorem 3 If P is a probability matrix such that P m > 0 for some m i.e. p ij > 0 for every entry in P m, for some m then P has exactly one stable vector, x. Furthermore, this stable vector is an attractor for P : in other words, if v is any probability vector, we have that lim n P n v = x. Basically, this means that if we run the system represented by P for long enough, the odds of us being in any one of our n states are given by the entries of x. The proof of this theorem is in the lecture notes and is kinda tricky, so we ve omitted it here because of time constraints. Let me know if you have questions about it, though! Instead, we offer an example, to illustrate what s going on here: Example. If we take our Caltech student from earlier, in the long run, what are they more likely to be doing sets or sleeping? 3

4 Solution. So, as we noted before, our probability matrix is P = In our earlier example, we showed that P 2 > 0; therefore, we know that there is exactly one stable vector, and this stable vector tells us precisely what our student is likely to be doing in the long run. So: let s find it! Recall that a stable vector is simply an eigenvector for that happens to also be a probability vector. So, to find our stable vector x, we simply need to find E : E = nullspace (A I)..3 0 = nullspace = all (x, y, z) such that (A I)..3 0 = nullspace = 3y x, 0 = 5x 3y + 7z, and 0 = x 7z. x y z = (0, 0, 0); i.e. Solving the above equations for x, y and z tells us that E is made up out of vectors of the form c (7, 4, ). Thus, if we want a vector of this form to have its entries sum up to, we simply let c = /22 and get that our unique stable vector is ( 7 22, 4 22, ). 22 Consequently, we can say that in the long run, our student is twice as likely to be studying (4/22) as it is to be sleeping (7/22). This is most of what we know about probability matrices! We now change gears here somewhat abruptly, to begin our discussion of real-valued symmetric matrices: 5 Real Symmetric Matrices: A Theorem and its Use First, as a reminder: Definition. A n n matrix A is called symmetric iff a ij = a ji, for every i and j; equivalently, A is called symmetric iff A = A T. 4

5 In lecture, we discussed the following remarkable theorem about real-valued symmetric matrices: Theorem 4 If A is a real-valued symmetric matrix, then A has only real-valued eigenvalues, A is diagonalizable, and furthermore A is diagonalizable by an orthogonal matrix : in other words, there is an orthogonal matrix E made of eigenvectors and diagonal matrix D made out of eigenvalues such that A = EDE T. As the proof was discussed in class and is kinda ponderous, we omit it here in favor of discussing two things: () how to use the above theorem, and (2) an example of the kinds of proofs you might be asked to do on HW/the final involving symmetric matrices. First: how *do* we use this theorem? Well, as it turns out, you can pretty much follow the exact same process we used for diagonalizing matrices in general. Suppose that we have a real symmetric matrix A; then, if we want to diagonalize it, we need to simply do the following:. First, find all of A s eigenvalues λ... λ k. 2. Once you ve done that, find each eigenspace E λi. 3. For each eigenspace E λi, find an orthonormal basis for E λi. This is the only difference between this process and normal diagonalization the bit about making sure that your basis is orthogonal. (To do this, simply use Gram-Schmidt on a normal basis for E λi, and then normalize all of your vectors by dividing by their length if you ve forgotten how to do Gram-Schmidt, consult week 4 s notes, or contact me!) 4. Take all of the vectors you got from these orthogonal bases, and use all of them as the columns in a matrix E; then, take their corresponding eigenvalues, and put them in the diagonal entries in some diagonal matrix D. Then, A = EDE T! and you re done. Again, for emphasis: the *only* difference between this and normal diagonalization is the part about finding orthogonal bases for the E λi s. We include an example of this algorithm here: Example. Diagonalize the matrix A = with orthogonal matrices: i.e. find an orthogonal matrix E and diagonal matrix D such that A = EDE T. A is called an orthogonal matrix iff A = A T : i.e. iff A 2 = I. 5

6 Solution. We follow our blueprint above. First, we find A s eigenvalues, by calculating its characteristic polynomial: λ 5 0 det(λi A) = det λ λ ( ) ( λ = (λ 5) det ( ) det 0 λ 0 λ = (λ 5) 2 (λ ) (λ ) = ((λ 5) 2 )(λ ) = (λ 2 0λ + 25 )(λ ) = (λ )(λ 4)(λ ) ) Therefore, and 4 are our eigenvalues. We now find their eigenspaces: E 4 = nullspace(a 4I) 0 = nullspace 0, which is spanned by the vector (,, 0). Normalizing gives us the vector ( / 2, / 2, 0). E = nullspace(a I) 0 = nullspace 0, which consists of all vectors (x, y, z) such that x = y. One such vector is (,, ); by using Gram-Schmidt or just being clever, another such vector that s orthogonal to (,, ) is (,, 2). Normalizing gives us (/ 3, / 3, / 3) and (/, /, 2/ ) as our vectors. Now that we ve found all of our vectors, we use them as the columns of the matrix / 2 / 3 / E = / 2 / 3 / / 3 2/. Then, if we let D = 0 0 be the diagonal 0 0 matrix corresponding to the appropriate eigenvalues of our matrix, by our theorem, we know that as claimed. A = EDE T

7 Real Symmetric Matrices: A Sample Proof Finally, we have an example that illustrates the kind of proofs we may ask you to do in your HW, or on the final, involving symmetric matrices: Proposition 5 A matrix A is symmetric if and only if x, Ay = Ax, y for every pair of vectors x, y R n. Proof. As this is an if and only if proof, we must prove both directions of our claim. We start by assuming that A is symmetric. Then, if we look at Ax, y for any pair of vectors x, y, we have (by definition!) that Ax, y = (Ax) T y = x T A T y = x T A y = x T (Ay) = x, Ay, which is exactly what we claimed. Conversely, assume that x, Ay = Ax, y for every pair of vectors x, y. It s not entirely clear how to proceed from here: so, let s try just exploring and seeing what this property gives us. In particular, let s see what this property tells us about A when we let x, y be the standard basis vectors of R n : i.e. x = e i = the vector that has a in its i-th spot and zeroes elsewhere, and y = e j = the vector that has a in its j-th spot and zeroes elsewhere. (In general, if you have a vector property and don t understand it, try to find out what it means when you plug in things like the standard basis vectors, the all- s vector, and any simple examples you can think of. This will often work!) So: with x, y defined as above, we have that A x = a... a n a n... a nn e i = the i-th column of A Ax, y = (the i-th column of A) e j = a ji, and a... a n A y = e j a n... a nn = the k-th column of A x, Ay = e i (the i-th column of A) = a ij But if these two inner products are equal, we ve just shown that a ij = a ji, for every i and j! In other words, A must be symmetric, as claimed. 7

Recitation 8: Graphs and Adjacency Matrices

Recitation 8: Graphs and Adjacency Matrices Math 1b TA: Padraic Bartlett Recitation 8: Graphs and Adjacency Matrices Week 8 Caltech 2011 1 Random Question Suppose you take a large triangle XY Z, and divide it up with straight line segments into

More information

Lecture 2: Eigenvalues and their Uses

Lecture 2: Eigenvalues and their Uses Spectral Graph Theory Instructor: Padraic Bartlett Lecture 2: Eigenvalues and their Uses Week 3 Mathcamp 2011 As you probably noticed on yesterday s HW, we, um, don t really have any good tools for finding

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

Recitation 5: Elementary Matrices

Recitation 5: Elementary Matrices Math 1b TA: Padraic Bartlett Recitation 5: Elementary Matrices Week 5 Caltech 2011 1 Random Question Consider the following two-player game, called Angels and Devils: Our game is played on a n n chessboard,

More information

Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices

Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices Math 108B Professor: Padraic Bartlett Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices Week 6 UCSB 2014 1 Lies Fun fact: I have deceived 1 you somewhat with these last few lectures! Let me

More information

MAT 1302B Mathematical Methods II

MAT 1302B Mathematical Methods II MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information

Practice problems for Exam 3 A =

Practice problems for Exam 3 A = Practice problems for Exam 3. Let A = 2 (a) Determine whether A is diagonalizable. If so, find a matrix S such that S AS is diagonal. If not, explain why not. (b) What are the eigenvalues of A? Is A diagonalizable?

More information

Lecture 15, 16: Diagonalization

Lecture 15, 16: Diagonalization Lecture 15, 16: Diagonalization Motivation: Eigenvalues and Eigenvectors are easy to compute for diagonal matrices. Hence, we would like (if possible) to convert matrix A into a diagonal matrix. Suppose

More information

MAT1302F Mathematical Methods II Lecture 19

MAT1302F Mathematical Methods II Lecture 19 MAT302F Mathematical Methods II Lecture 9 Aaron Christie 2 April 205 Eigenvectors, Eigenvalues, and Diagonalization Now that the basic theory of eigenvalues and eigenvectors is in place most importantly

More information

Lecture 10: Powers of Matrices, Difference Equations

Lecture 10: Powers of Matrices, Difference Equations Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each

More information

Lecture 4: Applications of Orthogonality: QR Decompositions

Lecture 4: Applications of Orthogonality: QR Decompositions Math 08B Professor: Padraic Bartlett Lecture 4: Applications of Orthogonality: QR Decompositions Week 4 UCSB 204 In our last class, we described the following method for creating orthonormal bases, known

More information

A PRIMER ON SESQUILINEAR FORMS

A PRIMER ON SESQUILINEAR FORMS A PRIMER ON SESQUILINEAR FORMS BRIAN OSSERMAN This is an alternative presentation of most of the material from 8., 8.2, 8.3, 8.4, 8.5 and 8.8 of Artin s book. Any terminology (such as sesquilinear form

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar

More information

k is a product of elementary matrices.

k is a product of elementary matrices. Mathematics, Spring Lecture (Wilson) Final Eam May, ANSWERS Problem (5 points) (a) There are three kinds of elementary row operations and associated elementary matrices. Describe what each kind of operation

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

Lecture 4: Constructing the Integers, Rationals and Reals

Lecture 4: Constructing the Integers, Rationals and Reals Math/CS 20: Intro. to Math Professor: Padraic Bartlett Lecture 4: Constructing the Integers, Rationals and Reals Week 5 UCSB 204 The Integers Normally, using the natural numbers, you can easily define

More information

PROBLEM SET. Problems on Eigenvalues and Diagonalization. Math 3351, Fall Oct. 20, 2010 ANSWERS

PROBLEM SET. Problems on Eigenvalues and Diagonalization. Math 3351, Fall Oct. 20, 2010 ANSWERS PROBLEM SET Problems on Eigenvalues and Diagonalization Math 335, Fall 2 Oct. 2, 2 ANSWERS i Problem. In each part, find the characteristic polynomial of the matrix and the eigenvalues of the matrix by

More information

Matrices related to linear transformations

Matrices related to linear transformations Math 4326 Fall 207 Matrices related to linear transformations We have encountered several ways in which matrices relate to linear transformations. In this note, I summarize the important facts and formulas

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

MATH 310, REVIEW SHEET

MATH 310, REVIEW SHEET MATH 310, REVIEW SHEET These notes are a summary of the key topics in the book (and follow the book pretty closely). You should be familiar with everything on here, but it s not comprehensive, so please

More information

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Math 291-2: Lecture Notes Northwestern University, Winter 2016 Math 291-2: Lecture Notes Northwestern University, Winter 2016 Written by Santiago Cañez These are lecture notes for Math 291-2, the second quarter of MENU: Intensive Linear Algebra and Multivariable Calculus,

More information

MATH 310, REVIEW SHEET 2

MATH 310, REVIEW SHEET 2 MATH 310, REVIEW SHEET 2 These notes are a very short summary of the key topics in the book (and follow the book pretty closely). You should be familiar with everything on here, but it s not comprehensive,

More information

Eigenvalues and Eigenvectors A =

Eigenvalues and Eigenvectors A = Eigenvalues and Eigenvectors Definition 0 Let A R n n be an n n real matrix A number λ R is a real eigenvalue of A if there exists a nonzero vector v R n such that A v = λ v The vector v is called an eigenvector

More information

Linear Algebra 2 Spectral Notes

Linear Algebra 2 Spectral Notes Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex

More information

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix

More information

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2 Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch

More information

Problems for M 11/2: A =

Problems for M 11/2: A = Math 30 Lesieutre Problem set # November 0 Problems for M /: 4 Let B be the basis given by b b Find the B-matrix for the transformation T : R R given by x Ax where 3 4 A (This just means the matrix for

More information

and let s calculate the image of some vectors under the transformation T.

and let s calculate the image of some vectors under the transformation T. Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =

More information

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes Written by Santiago Cañez These are notes which provide a basic summary of each lecture for Math 290-2, the second

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Dot Products, Transposes, and Orthogonal Projections

Dot Products, Transposes, and Orthogonal Projections Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +

More information

MA 265 FINAL EXAM Fall 2012

MA 265 FINAL EXAM Fall 2012 MA 265 FINAL EXAM Fall 22 NAME: INSTRUCTOR S NAME:. There are a total of 25 problems. You should show work on the exam sheet, and pencil in the correct answer on the scantron. 2. No books, notes, or calculators

More information

PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper)

PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper) PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper) In class, we saw this graph, with each node representing people who are following each other on Twitter: Our

More information

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left

More information

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0 LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Lecture 8: A Crash Course in Linear Algebra

Lecture 8: A Crash Course in Linear Algebra Math/CS 120: Intro. to Math Professor: Padraic Bartlett Lecture 8: A Crash Course in Linear Algebra Week 9 UCSB 2014 Qué sed de saber cuánto! Pablo Neruda, Oda a los Números 1 Linear Algebra In the past

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

Math 31 Lesson Plan. Day 5: Intro to Groups. Elizabeth Gillaspy. September 28, 2011

Math 31 Lesson Plan. Day 5: Intro to Groups. Elizabeth Gillaspy. September 28, 2011 Math 31 Lesson Plan Day 5: Intro to Groups Elizabeth Gillaspy September 28, 2011 Supplies needed: Sign in sheet Goals for students: Students will: Improve the clarity of their proof-writing. Gain confidence

More information

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions. The exam will cover Sections 6.-6.2 and 7.-7.4: True/False 30% Definitions 0% Computational 60% Skip Minors and Laplace Expansion in Section 6.2 and p. 304 (trajectories and phase portraits) in Section

More information

Linear Algebra Highlights

Linear Algebra Highlights Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to

More information

Math 20F Practice Final Solutions. Jor-el Briones

Math 20F Practice Final Solutions. Jor-el Briones Math 2F Practice Final Solutions Jor-el Briones Jor-el Briones / Math 2F Practice Problems for Final Page 2 of 6 NOTE: For the solutions to these problems, I skip all the row reduction calculations. Please

More information

Problems for M 10/26:

Problems for M 10/26: Math, Lesieutre Problem set # November 4, 25 Problems for M /26: 5 Is λ 2 an eigenvalue of 2? 8 Why or why not? 2 A 2I The determinant is, which means that A 2I has 6 a nullspace, and so there is an eigenvector

More information

Math 115A: Homework 5

Math 115A: Homework 5 Math 115A: Homework 5 1 Suppose U, V, and W are finite-dimensional vector spaces over a field F, and that are linear a) Prove ker ST ) ker T ) b) Prove nullst ) nullt ) c) Prove imst ) im S T : U V, S

More information

12x + 18y = 30? ax + by = m

12x + 18y = 30? ax + by = m Math 2201, Further Linear Algebra: a practical summary. February, 2009 There are just a few themes that were covered in the course. I. Algebra of integers and polynomials. II. Structure theory of one endomorphism.

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

Lecture 6: Finite Fields

Lecture 6: Finite Fields CCS Discrete Math I Professor: Padraic Bartlett Lecture 6: Finite Fields Week 6 UCSB 2014 It ain t what they call you, it s what you answer to. W. C. Fields 1 Fields In the next two weeks, we re going

More information

Math 205, Summer I, Week 4b:

Math 205, Summer I, Week 4b: Math 205, Summer I, 2016 Week 4b: Chapter 5, Sections 6, 7 and 8 (5.5 is NOT on the syllabus) 5.6 Eigenvalues and Eigenvectors 5.7 Eigenspaces, nondefective matrices 5.8 Diagonalization [*** See next slide

More information

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:

More information

Eigenvectors and Hermitian Operators

Eigenvectors and Hermitian Operators 7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding

More information

Math 3191 Applied Linear Algebra

Math 3191 Applied Linear Algebra Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction MAT4 : Introduction to Applied Linear Algebra Mike Newman fall 7 9. Projections introduction One reason to consider projections is to understand approximate solutions to linear systems. A common example

More information

Math 110 Linear Algebra Midterm 2 Review October 28, 2017

Math 110 Linear Algebra Midterm 2 Review October 28, 2017 Math 11 Linear Algebra Midterm Review October 8, 17 Material Material covered on the midterm includes: All lectures from Thursday, Sept. 1st to Tuesday, Oct. 4th Homeworks 9 to 17 Quizzes 5 to 9 Sections

More information

Math 205, Summer I, Week 4b: Continued. Chapter 5, Section 8

Math 205, Summer I, Week 4b: Continued. Chapter 5, Section 8 Math 205, Summer I, 2016 Week 4b: Continued Chapter 5, Section 8 2 5.8 Diagonalization [reprint, week04: Eigenvalues and Eigenvectors] + diagonaliization 1. 5.8 Eigenspaces, Diagonalization A vector v

More information

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4 Linear Systems Math Spring 8 c 8 Ron Buckmire Fowler 9 MWF 9: am - :5 am http://faculty.oxy.edu/ron/math//8/ Class 7 TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5. Summary

More information

Solutions to Final Exam

Solutions to Final Exam Solutions to Final Exam. Let A be a 3 5 matrix. Let b be a nonzero 5-vector. Assume that the nullity of A is. (a) What is the rank of A? 3 (b) Are the rows of A linearly independent? (c) Are the columns

More information

Differential Equations

Differential Equations This document was written and copyrighted by Paul Dawkins. Use of this document and its online version is governed by the Terms and Conditions of Use located at. The online version of this document is

More information

18.06 Quiz 2 April 7, 2010 Professor Strang

18.06 Quiz 2 April 7, 2010 Professor Strang 18.06 Quiz 2 April 7, 2010 Professor Strang Your PRINTED name is: 1. Your recitation number or instructor is 2. 3. 1. (33 points) (a) Find the matrix P that projects every vector b in R 3 onto the line

More information

Problem 1: Solving a linear equation

Problem 1: Solving a linear equation Math 38 Practice Final Exam ANSWERS Page Problem : Solving a linear equation Given matrix A = 2 2 3 7 4 and vector y = 5 8 9. (a) Solve Ax = y (if the equation is consistent) and write the general solution

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

22A-2 SUMMER 2014 LECTURE Agenda

22A-2 SUMMER 2014 LECTURE Agenda 22A-2 SUMMER 204 LECTURE 2 NATHANIEL GALLUP The Dot Product Continued Matrices Group Work Vectors and Linear Equations Agenda 2 Dot Product Continued Angles between vectors Given two 2-dimensional vectors

More information

Lecture 5: Latin Squares and Magic

Lecture 5: Latin Squares and Magic Latin Squares Instructor: Padraic Bartlett Lecture 5: Latin Squares and Magic Week Mathcamp 0 Today s application is to magic! Not the friendship kind though ; instead we re going to talk about magic squares

More information

Math 215 HW #9 Solutions

Math 215 HW #9 Solutions Math 5 HW #9 Solutions. Problem 4.4.. If A is a 5 by 5 matrix with all a ij, then det A. Volumes or the big formula or pivots should give some upper bound on the determinant. Answer: Let v i be the ith

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that

More information

Eigenvalues and Eigenvectors. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Eigenvalues and Eigenvectors. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB Eigenvalues and Eigenvectors Consider the equation A x = λ x, where A is an nxn matrix. We call x (must be non-zero) an eigenvector of A if this equation can be solved for some value of λ. We call λ an

More information

Math 314H Solutions to Homework # 3

Math 314H Solutions to Homework # 3 Math 34H Solutions to Homework # 3 Complete the exercises from the second maple assignment which can be downloaded from my linear algebra course web page Attach printouts of your work on this problem to

More information

Topics in linear algebra

Topics in linear algebra Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over

More information

MTH50 Spring 07 HW Assignment 7 {From [FIS0]}: Sec 44 #4a h 6; Sec 5 #ad ac 4ae 4 7 The due date for this assignment is 04/05/7 Sec 44 #4a h Evaluate the erminant of the following matrices by any legitimate

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

Lecture 19: The Determinant

Lecture 19: The Determinant Math 108a Professor: Padraic Bartlett Lecture 19: The Determinant Week 10 UCSB 2013 In our last class, we talked about how to calculate volume in n-dimensions Specifically, we defined a parallelotope:

More information

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. Math 5327 Fall 2018 Homework 7 1. For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. 3 1 0 (a) A = 1 2 0 1 1 0 x 3 1 0 Solution: 1 x 2 0

More information

The minimal polynomial

The minimal polynomial The minimal polynomial Michael H Mertens October 22, 2015 Introduction In these short notes we explain some of the important features of the minimal polynomial of a square matrix A and recall some basic

More information

DIFFERENTIAL EQUATIONS REVIEW. Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it.

DIFFERENTIAL EQUATIONS REVIEW. Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it. DIFFERENTIAL EQUATIONS REVIEW PEYAM RYAN TABRIZIAN Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it. Welcome to the special Friday after-school special of That

More information

Columbus State Community College Mathematics Department Public Syllabus

Columbus State Community College Mathematics Department Public Syllabus Columbus State Community College Mathematics Department Public Syllabus Course and Number: MATH 2568 Elementary Linear Algebra Credits: 4 Class Hours Per Week: 4 Prerequisites: MATH 2153 with a C or higher

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

0 2 0, it is diagonal, hence diagonalizable)

0 2 0, it is diagonal, hence diagonalizable) MATH 54 TRUE/FALSE QUESTIONS FOR MIDTERM 2 SOLUTIONS PEYAM RYAN TABRIZIAN 1. (a) TRUE If A is diagonalizable, then A 3 is diagonalizable. (A = P DP 1, so A 3 = P D 3 P = P D P 1, where P = P and D = D

More information

Eigenvalues & Eigenvectors

Eigenvalues & Eigenvectors Eigenvalues & Eigenvectors Page 1 Eigenvalues are a very important concept in linear algebra, and one that comes up in other mathematics courses as well. The word eigen is German for inherent or characteristic,

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Final Exam Practice Problems Answers Math 24 Winter 2012

Final Exam Practice Problems Answers Math 24 Winter 2012 Final Exam Practice Problems Answers Math 4 Winter 0 () The Jordan product of two n n matrices is defined as A B = (AB + BA), where the products inside the parentheses are standard matrix product. Is the

More information

Chapter 4 & 5: Vector Spaces & Linear Transformations

Chapter 4 & 5: Vector Spaces & Linear Transformations Chapter 4 & 5: Vector Spaces & Linear Transformations Philip Gressman University of Pennsylvania Philip Gressman Math 240 002 2014C: Chapters 4 & 5 1 / 40 Objective The purpose of Chapter 4 is to think

More information

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler. Math 453 Exam 3 Review The syllabus for Exam 3 is Chapter 6 (pages -2), Chapter 7 through page 37, and Chapter 8 through page 82 in Axler.. You should be sure to know precise definition of the terms we

More information

= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E.

= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E. 3.3 Diagonalization Let A = 4. Then and are eigenvectors of A, with corresponding eigenvalues 2 and 6 respectively (check). This means 4 = 2, 4 = 6. 2 2 2 2 Thus 4 = 2 2 6 2 = 2 6 4 2 We have 4 = 2 0 0

More information

Designing Information Devices and Systems I Fall 2018 Homework 5

Designing Information Devices and Systems I Fall 2018 Homework 5 Last Updated: 08-09-9 0:6 EECS 6A Designing Information Devices and Systems I Fall 08 Homework 5 This homework is due September 8, 08, at 3:59. Self-grades are due October, 08, at 3:59. Submission Format

More information

Solution Set 7, Fall '12

Solution Set 7, Fall '12 Solution Set 7, 18.06 Fall '12 1. Do Problem 26 from 5.1. (It might take a while but when you see it, it's easy) Solution. Let n 3, and let A be an n n matrix whose i, j entry is i + j. To show that det

More information

FINAL EXAM Ma (Eakin) Fall 2015 December 16, 2015

FINAL EXAM Ma (Eakin) Fall 2015 December 16, 2015 FINAL EXAM Ma-00 Eakin Fall 05 December 6, 05 Please make sure that your name and GUID are on every page. This exam is designed to be done with pencil-and-paper calculations. You may use your calculator

More information

Appendix A. Vector addition: - The sum of two vectors is another vector that also lie in the space:

Appendix A. Vector addition: - The sum of two vectors is another vector that also lie in the space: Tor Kjellsson Stockholm University Appendix A A.1 Q. Consider the ordinary vectors in 3 dimensions (a x î+a y ĵ+a zˆk), with complex components. Do the following subsets constitute a vector space? If so,

More information

Math Lecture 26 : The Properties of Determinants

Math Lecture 26 : The Properties of Determinants Math 2270 - Lecture 26 : The Properties of Determinants Dylan Zwick Fall 202 The lecture covers section 5. from the textbook. The determinant of a square matrix is a number that tells you quite a bit about

More information

Math 18.6, Spring 213 Problem Set #6 April 5, 213 Problem 1 ( 5.2, 4). Identify all the nonzero terms in the big formula for the determinants of the following matrices: 1 1 1 2 A = 1 1 1 1 1 1, B = 3 4

More information

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if UNDERSTANDING THE DIAGONALIZATION PROBLEM Roy Skjelnes Abstract These notes are additional material to the course B107, given fall 200 The style may appear a bit coarse and consequently the student is

More information

5.3.5 The eigenvalues are 3, 2, 3 (i.e., the diagonal entries of D) with corresponding eigenvalues. Null(A 3I) = Null( ), 0 0

5.3.5 The eigenvalues are 3, 2, 3 (i.e., the diagonal entries of D) with corresponding eigenvalues. Null(A 3I) = Null( ), 0 0 535 The eigenvalues are 3,, 3 (ie, the diagonal entries of D) with corresponding eigenvalues,, 538 The matrix is upper triangular so the eigenvalues are simply the diagonal entries, namely 3, 3 The corresponding

More information

From Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play?

From Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play? Overview Last week introduced the important Diagonalisation Theorem: An n n matrix A is diagonalisable if and only if there is a basis for R n consisting of eigenvectors of A. This week we ll continue

More information

MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS

MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS 1. (7 pts)[apostol IV.8., 13, 14] (.) Let A be an n n matrix with characteristic polynomial f(λ). Prove (by induction) that the coefficient of λ n 1 in f(λ) is

More information

Homework 3 Solutions Math 309, Fall 2015

Homework 3 Solutions Math 309, Fall 2015 Homework 3 Solutions Math 39, Fall 25 782 One easily checks that the only eigenvalue of the coefficient matrix is λ To find the associated eigenvector, we have 4 2 v v 8 4 (up to scalar multiplication)

More information

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar

More information

Topic 15 Notes Jeremy Orloff

Topic 15 Notes Jeremy Orloff Topic 5 Notes Jeremy Orloff 5 Transpose, Inverse, Determinant 5. Goals. Know the definition and be able to compute the inverse of any square matrix using row operations. 2. Know the properties of inverses.

More information