Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples
|
|
- Marsha Wiggins
- 5 years ago
- Views:
Transcription
1 Math b TA: Padraic Bartlett Recitation 9: Probability Matrices and Real Symmetric Matrices Week 9 Caltech 20 Random Question Show that = ϕ, the golden ratio, which is = Homework comments Section average: 72/80 = 90%. This was about 2 3% above the class average; nice work! As the above comment suggests, people did pretty well! There isn t much to say, other than please include explanations of all the things you do. Today, our talk is pretty much split into two topics: probability matrices and real symmetric matrices. For each subject, we ll include examples of both the calculations you ll be expected to do on the HW, as well as proofs that you might be asked to understand or replicate. 3 Probability Matrices: Definitions and Examples Definition. A n n matrix P is called a probability matrix if and only if the following two properties are satisfied: P 0; in other words, p ij 0 for every entry p ij of P. The column sums of P are all ; in other words, n i= p ij =, for every j. Why do we call these kinds of matrices probabilitiy matrices? Well, consider the following way of interpreting such a matrix P : Suppose that you have an object that is capable of being in n different states. For example, if the object you re modeling is a Caltech undergrad, you could roughly model them as having the following three distinct states: {sleeping, doing sets, eating}
2 Suppose furthermore that at the end of every hour ( or more generally, at the end of every step, where your step size is some arbitrary unit), the undergrad you re studying has certain probabilities of switching from the state they re in into a different state. I.e. you could assume that if your undergrad is working on a set for a hour, there s a decent chance (30%) that they pass out because it s difficult, a better chance (0%) that they keep working on the set, and a small chance (0%) that they get hungry and stop for a sandwich. If you have this information for all of your states, you can create a probability matrix with this data! To do this, let p ij contain the probability of switching from state j to state i. Then, we have a matrix all of whose columns sum to (this is because your undergrad is always in *one* of these three states, and thus when they leave any state j the sum of their chances of landing in the other states must be.) In other words: a probability matrix! As it turns out, this method of describing a finite-state system isn t just notationally useful: we in fact have a pair of remarkably useful theorems for probability matrices! 4 Probability Matrices: Two Useful Theorems Theorem If we have a probability matrix P representing some finite system with n states {,... n}, then the probability of starting in state j and ending in state i in precisely m steps is the (i, j)-th entry in P m. Proof. Last recitation, we proved that for a n n matrix P, the (i, j)-th entry of P m can be written as the sum n c,...c m = p i,c p c,c 2,... p cm 2,c m p cm,j But what *is* one of these terms p i,c p c,c 2,... p cm 2,c m p cm,j? Well, if we read it from right to left, it s just the product (probability of going from j to c m ) (probability of going from c m to c m 2 )... (probability of going from c to i). If we just multiply everything out, we can see that this is just the probability that we take the specific path j c m... i from j to i! And our sum has one term for every such path from i to j: therefore, because our sum is just the sum of all of these probabilities, it represents the chances of us taking any of these paths from i to j. Therefore, the (i, j)-th entry represents the probability of us going from j to i in m steps along any of these paths which is what we claimed. An example of this theorem follows below: 2
3 Example. Suppose we have a Caltech student capable of entering three possible states {sleeping, sets 2, eating 3 }, and suppose it switches between these three states every hour according to the following probability matrix: P = (I.e. if our student is asleep at noon, it has a 40% chance of staying asleep at, a 50% chance of waking up and starting a set, and a 0% chance of deciding it s hungry for a kebab.) If your student is asleep at 2am, what are the chances that it will be working on a set at 4am? Solution. So: by the theorem above, we just need to look at the (2, ) cell in P 2 to figure this out. Calculating gives us that P 2 = and thus that our poor student has a 2% chance of working on their sets at 4am. This theorem allows us to say what happens at specific points and times in the future. However: what if we re not interested in specific points and times in the future, but rather in the long-term behavior of the system as a whole? The following two definitions and pair of theorems tell us what to do in that situation: Definition. A vector v R n is called a probability vector if and only if n i= v i =. Definition. For a probability matrix P, v is called a stable vector if and only if v is a probability vector that is also an eigenvector of P with corresponding eigenvalue. Theorem 2 Every probability matrix has at least one stable vector. Theorem 3 If P is a probability matrix such that P m > 0 for some m i.e. p ij > 0 for every entry in P m, for some m then P has exactly one stable vector, x. Furthermore, this stable vector is an attractor for P : in other words, if v is any probability vector, we have that lim n P n v = x. Basically, this means that if we run the system represented by P for long enough, the odds of us being in any one of our n states are given by the entries of x. The proof of this theorem is in the lecture notes and is kinda tricky, so we ve omitted it here because of time constraints. Let me know if you have questions about it, though! Instead, we offer an example, to illustrate what s going on here: Example. If we take our Caltech student from earlier, in the long run, what are they more likely to be doing sets or sleeping? 3
4 Solution. So, as we noted before, our probability matrix is P = In our earlier example, we showed that P 2 > 0; therefore, we know that there is exactly one stable vector, and this stable vector tells us precisely what our student is likely to be doing in the long run. So: let s find it! Recall that a stable vector is simply an eigenvector for that happens to also be a probability vector. So, to find our stable vector x, we simply need to find E : E = nullspace (A I)..3 0 = nullspace = all (x, y, z) such that (A I)..3 0 = nullspace = 3y x, 0 = 5x 3y + 7z, and 0 = x 7z. x y z = (0, 0, 0); i.e. Solving the above equations for x, y and z tells us that E is made up out of vectors of the form c (7, 4, ). Thus, if we want a vector of this form to have its entries sum up to, we simply let c = /22 and get that our unique stable vector is ( 7 22, 4 22, ). 22 Consequently, we can say that in the long run, our student is twice as likely to be studying (4/22) as it is to be sleeping (7/22). This is most of what we know about probability matrices! We now change gears here somewhat abruptly, to begin our discussion of real-valued symmetric matrices: 5 Real Symmetric Matrices: A Theorem and its Use First, as a reminder: Definition. A n n matrix A is called symmetric iff a ij = a ji, for every i and j; equivalently, A is called symmetric iff A = A T. 4
5 In lecture, we discussed the following remarkable theorem about real-valued symmetric matrices: Theorem 4 If A is a real-valued symmetric matrix, then A has only real-valued eigenvalues, A is diagonalizable, and furthermore A is diagonalizable by an orthogonal matrix : in other words, there is an orthogonal matrix E made of eigenvectors and diagonal matrix D made out of eigenvalues such that A = EDE T. As the proof was discussed in class and is kinda ponderous, we omit it here in favor of discussing two things: () how to use the above theorem, and (2) an example of the kinds of proofs you might be asked to do on HW/the final involving symmetric matrices. First: how *do* we use this theorem? Well, as it turns out, you can pretty much follow the exact same process we used for diagonalizing matrices in general. Suppose that we have a real symmetric matrix A; then, if we want to diagonalize it, we need to simply do the following:. First, find all of A s eigenvalues λ... λ k. 2. Once you ve done that, find each eigenspace E λi. 3. For each eigenspace E λi, find an orthonormal basis for E λi. This is the only difference between this process and normal diagonalization the bit about making sure that your basis is orthogonal. (To do this, simply use Gram-Schmidt on a normal basis for E λi, and then normalize all of your vectors by dividing by their length if you ve forgotten how to do Gram-Schmidt, consult week 4 s notes, or contact me!) 4. Take all of the vectors you got from these orthogonal bases, and use all of them as the columns in a matrix E; then, take their corresponding eigenvalues, and put them in the diagonal entries in some diagonal matrix D. Then, A = EDE T! and you re done. Again, for emphasis: the *only* difference between this and normal diagonalization is the part about finding orthogonal bases for the E λi s. We include an example of this algorithm here: Example. Diagonalize the matrix A = with orthogonal matrices: i.e. find an orthogonal matrix E and diagonal matrix D such that A = EDE T. A is called an orthogonal matrix iff A = A T : i.e. iff A 2 = I. 5
6 Solution. We follow our blueprint above. First, we find A s eigenvalues, by calculating its characteristic polynomial: λ 5 0 det(λi A) = det λ λ ( ) ( λ = (λ 5) det ( ) det 0 λ 0 λ = (λ 5) 2 (λ ) (λ ) = ((λ 5) 2 )(λ ) = (λ 2 0λ + 25 )(λ ) = (λ )(λ 4)(λ ) ) Therefore, and 4 are our eigenvalues. We now find their eigenspaces: E 4 = nullspace(a 4I) 0 = nullspace 0, which is spanned by the vector (,, 0). Normalizing gives us the vector ( / 2, / 2, 0). E = nullspace(a I) 0 = nullspace 0, which consists of all vectors (x, y, z) such that x = y. One such vector is (,, ); by using Gram-Schmidt or just being clever, another such vector that s orthogonal to (,, ) is (,, 2). Normalizing gives us (/ 3, / 3, / 3) and (/, /, 2/ ) as our vectors. Now that we ve found all of our vectors, we use them as the columns of the matrix / 2 / 3 / E = / 2 / 3 / / 3 2/. Then, if we let D = 0 0 be the diagonal 0 0 matrix corresponding to the appropriate eigenvalues of our matrix, by our theorem, we know that as claimed. A = EDE T
7 Real Symmetric Matrices: A Sample Proof Finally, we have an example that illustrates the kind of proofs we may ask you to do in your HW, or on the final, involving symmetric matrices: Proposition 5 A matrix A is symmetric if and only if x, Ay = Ax, y for every pair of vectors x, y R n. Proof. As this is an if and only if proof, we must prove both directions of our claim. We start by assuming that A is symmetric. Then, if we look at Ax, y for any pair of vectors x, y, we have (by definition!) that Ax, y = (Ax) T y = x T A T y = x T A y = x T (Ay) = x, Ay, which is exactly what we claimed. Conversely, assume that x, Ay = Ax, y for every pair of vectors x, y. It s not entirely clear how to proceed from here: so, let s try just exploring and seeing what this property gives us. In particular, let s see what this property tells us about A when we let x, y be the standard basis vectors of R n : i.e. x = e i = the vector that has a in its i-th spot and zeroes elsewhere, and y = e j = the vector that has a in its j-th spot and zeroes elsewhere. (In general, if you have a vector property and don t understand it, try to find out what it means when you plug in things like the standard basis vectors, the all- s vector, and any simple examples you can think of. This will often work!) So: with x, y defined as above, we have that A x = a... a n a n... a nn e i = the i-th column of A Ax, y = (the i-th column of A) e j = a ji, and a... a n A y = e j a n... a nn = the k-th column of A x, Ay = e i (the i-th column of A) = a ij But if these two inner products are equal, we ve just shown that a ij = a ji, for every i and j! In other words, A must be symmetric, as claimed. 7
Recitation 8: Graphs and Adjacency Matrices
Math 1b TA: Padraic Bartlett Recitation 8: Graphs and Adjacency Matrices Week 8 Caltech 2011 1 Random Question Suppose you take a large triangle XY Z, and divide it up with straight line segments into
More informationLecture 2: Eigenvalues and their Uses
Spectral Graph Theory Instructor: Padraic Bartlett Lecture 2: Eigenvalues and their Uses Week 3 Mathcamp 2011 As you probably noticed on yesterday s HW, we, um, don t really have any good tools for finding
More information[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]
Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the
More informationRecitation 5: Elementary Matrices
Math 1b TA: Padraic Bartlett Recitation 5: Elementary Matrices Week 5 Caltech 2011 1 Random Question Consider the following two-player game, called Angels and Devils: Our game is played on a n n chessboard,
More informationLecture 6: Lies, Inner Product Spaces, and Symmetric Matrices
Math 108B Professor: Padraic Bartlett Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices Week 6 UCSB 2014 1 Lies Fun fact: I have deceived 1 you somewhat with these last few lectures! Let me
More informationMAT 1302B Mathematical Methods II
MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture
More informationMATH 304 Linear Algebra Lecture 34: Review for Test 2.
MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1
More informationPractice problems for Exam 3 A =
Practice problems for Exam 3. Let A = 2 (a) Determine whether A is diagonalizable. If so, find a matrix S such that S AS is diagonal. If not, explain why not. (b) What are the eigenvalues of A? Is A diagonalizable?
More informationLecture 15, 16: Diagonalization
Lecture 15, 16: Diagonalization Motivation: Eigenvalues and Eigenvectors are easy to compute for diagonal matrices. Hence, we would like (if possible) to convert matrix A into a diagonal matrix. Suppose
More informationMAT1302F Mathematical Methods II Lecture 19
MAT302F Mathematical Methods II Lecture 9 Aaron Christie 2 April 205 Eigenvectors, Eigenvalues, and Diagonalization Now that the basic theory of eigenvalues and eigenvectors is in place most importantly
More informationLecture 10: Powers of Matrices, Difference Equations
Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each
More informationLecture 4: Applications of Orthogonality: QR Decompositions
Math 08B Professor: Padraic Bartlett Lecture 4: Applications of Orthogonality: QR Decompositions Week 4 UCSB 204 In our last class, we described the following method for creating orthonormal bases, known
More informationA PRIMER ON SESQUILINEAR FORMS
A PRIMER ON SESQUILINEAR FORMS BRIAN OSSERMAN This is an alternative presentation of most of the material from 8., 8.2, 8.3, 8.4, 8.5 and 8.8 of Artin s book. Any terminology (such as sesquilinear form
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationREVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION
REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar
More informationk is a product of elementary matrices.
Mathematics, Spring Lecture (Wilson) Final Eam May, ANSWERS Problem (5 points) (a) There are three kinds of elementary row operations and associated elementary matrices. Describe what each kind of operation
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More informationLecture 4: Constructing the Integers, Rationals and Reals
Math/CS 20: Intro. to Math Professor: Padraic Bartlett Lecture 4: Constructing the Integers, Rationals and Reals Week 5 UCSB 204 The Integers Normally, using the natural numbers, you can easily define
More informationPROBLEM SET. Problems on Eigenvalues and Diagonalization. Math 3351, Fall Oct. 20, 2010 ANSWERS
PROBLEM SET Problems on Eigenvalues and Diagonalization Math 335, Fall 2 Oct. 2, 2 ANSWERS i Problem. In each part, find the characteristic polynomial of the matrix and the eigenvalues of the matrix by
More informationMatrices related to linear transformations
Math 4326 Fall 207 Matrices related to linear transformations We have encountered several ways in which matrices relate to linear transformations. In this note, I summarize the important facts and formulas
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationMATH 310, REVIEW SHEET
MATH 310, REVIEW SHEET These notes are a summary of the key topics in the book (and follow the book pretty closely). You should be familiar with everything on here, but it s not comprehensive, so please
More informationMath 291-2: Lecture Notes Northwestern University, Winter 2016
Math 291-2: Lecture Notes Northwestern University, Winter 2016 Written by Santiago Cañez These are lecture notes for Math 291-2, the second quarter of MENU: Intensive Linear Algebra and Multivariable Calculus,
More informationMATH 310, REVIEW SHEET 2
MATH 310, REVIEW SHEET 2 These notes are a very short summary of the key topics in the book (and follow the book pretty closely). You should be familiar with everything on here, but it s not comprehensive,
More informationEigenvalues and Eigenvectors A =
Eigenvalues and Eigenvectors Definition 0 Let A R n n be an n n real matrix A number λ R is a real eigenvalue of A if there exists a nonzero vector v R n such that A v = λ v The vector v is called an eigenvector
More informationLinear Algebra 2 Spectral Notes
Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex
More information22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices
m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More informationProblems for M 11/2: A =
Math 30 Lesieutre Problem set # November 0 Problems for M /: 4 Let B be the basis given by b b Find the B-matrix for the transformation T : R R given by x Ax where 3 4 A (This just means the matrix for
More informationand let s calculate the image of some vectors under the transformation T.
Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =
More informationMath 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes
Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes Written by Santiago Cañez These are notes which provide a basic summary of each lecture for Math 290-2, the second
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More informationMA 265 FINAL EXAM Fall 2012
MA 265 FINAL EXAM Fall 22 NAME: INSTRUCTOR S NAME:. There are a total of 25 problems. You should show work on the exam sheet, and pencil in the correct answer on the scantron. 2. No books, notes, or calculators
More informationPageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper)
PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper) In class, we saw this graph, with each node representing people who are following each other on Twitter: Our
More informationMATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL
MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left
More informationc 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0
LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationLecture 8: A Crash Course in Linear Algebra
Math/CS 120: Intro. to Math Professor: Padraic Bartlett Lecture 8: A Crash Course in Linear Algebra Week 9 UCSB 2014 Qué sed de saber cuánto! Pablo Neruda, Oda a los Números 1 Linear Algebra In the past
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationMath 31 Lesson Plan. Day 5: Intro to Groups. Elizabeth Gillaspy. September 28, 2011
Math 31 Lesson Plan Day 5: Intro to Groups Elizabeth Gillaspy September 28, 2011 Supplies needed: Sign in sheet Goals for students: Students will: Improve the clarity of their proof-writing. Gain confidence
More informationExamples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.
The exam will cover Sections 6.-6.2 and 7.-7.4: True/False 30% Definitions 0% Computational 60% Skip Minors and Laplace Expansion in Section 6.2 and p. 304 (trajectories and phase portraits) in Section
More informationLinear Algebra Highlights
Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to
More informationMath 20F Practice Final Solutions. Jor-el Briones
Math 2F Practice Final Solutions Jor-el Briones Jor-el Briones / Math 2F Practice Problems for Final Page 2 of 6 NOTE: For the solutions to these problems, I skip all the row reduction calculations. Please
More informationProblems for M 10/26:
Math, Lesieutre Problem set # November 4, 25 Problems for M /26: 5 Is λ 2 an eigenvalue of 2? 8 Why or why not? 2 A 2I The determinant is, which means that A 2I has 6 a nullspace, and so there is an eigenvector
More informationMath 115A: Homework 5
Math 115A: Homework 5 1 Suppose U, V, and W are finite-dimensional vector spaces over a field F, and that are linear a) Prove ker ST ) ker T ) b) Prove nullst ) nullt ) c) Prove imst ) im S T : U V, S
More information12x + 18y = 30? ax + by = m
Math 2201, Further Linear Algebra: a practical summary. February, 2009 There are just a few themes that were covered in the course. I. Algebra of integers and polynomials. II. Structure theory of one endomorphism.
More informationMATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION
MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether
More informationLecture 6: Finite Fields
CCS Discrete Math I Professor: Padraic Bartlett Lecture 6: Finite Fields Week 6 UCSB 2014 It ain t what they call you, it s what you answer to. W. C. Fields 1 Fields In the next two weeks, we re going
More informationMath 205, Summer I, Week 4b:
Math 205, Summer I, 2016 Week 4b: Chapter 5, Sections 6, 7 and 8 (5.5 is NOT on the syllabus) 5.6 Eigenvalues and Eigenvectors 5.7 Eigenspaces, nondefective matrices 5.8 Diagonalization [*** See next slide
More informationMATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.
MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:
More informationEigenvectors and Hermitian Operators
7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding
More informationMath 3191 Applied Linear Algebra
Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful
More informationLINEAR ALGEBRA KNOWLEDGE SURVEY
LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.
More informationMAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction
MAT4 : Introduction to Applied Linear Algebra Mike Newman fall 7 9. Projections introduction One reason to consider projections is to understand approximate solutions to linear systems. A common example
More informationMath 110 Linear Algebra Midterm 2 Review October 28, 2017
Math 11 Linear Algebra Midterm Review October 8, 17 Material Material covered on the midterm includes: All lectures from Thursday, Sept. 1st to Tuesday, Oct. 4th Homeworks 9 to 17 Quizzes 5 to 9 Sections
More informationMath 205, Summer I, Week 4b: Continued. Chapter 5, Section 8
Math 205, Summer I, 2016 Week 4b: Continued Chapter 5, Section 8 2 5.8 Diagonalization [reprint, week04: Eigenvalues and Eigenvectors] + diagonaliization 1. 5.8 Eigenspaces, Diagonalization A vector v
More informationLinear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4
Linear Systems Math Spring 8 c 8 Ron Buckmire Fowler 9 MWF 9: am - :5 am http://faculty.oxy.edu/ron/math//8/ Class 7 TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5. Summary
More informationSolutions to Final Exam
Solutions to Final Exam. Let A be a 3 5 matrix. Let b be a nonzero 5-vector. Assume that the nullity of A is. (a) What is the rank of A? 3 (b) Are the rows of A linearly independent? (c) Are the columns
More informationDifferential Equations
This document was written and copyrighted by Paul Dawkins. Use of this document and its online version is governed by the Terms and Conditions of Use located at. The online version of this document is
More information18.06 Quiz 2 April 7, 2010 Professor Strang
18.06 Quiz 2 April 7, 2010 Professor Strang Your PRINTED name is: 1. Your recitation number or instructor is 2. 3. 1. (33 points) (a) Find the matrix P that projects every vector b in R 3 onto the line
More informationProblem 1: Solving a linear equation
Math 38 Practice Final Exam ANSWERS Page Problem : Solving a linear equation Given matrix A = 2 2 3 7 4 and vector y = 5 8 9. (a) Solve Ax = y (if the equation is consistent) and write the general solution
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More information22A-2 SUMMER 2014 LECTURE Agenda
22A-2 SUMMER 204 LECTURE 2 NATHANIEL GALLUP The Dot Product Continued Matrices Group Work Vectors and Linear Equations Agenda 2 Dot Product Continued Angles between vectors Given two 2-dimensional vectors
More informationLecture 5: Latin Squares and Magic
Latin Squares Instructor: Padraic Bartlett Lecture 5: Latin Squares and Magic Week Mathcamp 0 Today s application is to magic! Not the friendship kind though ; instead we re going to talk about magic squares
More informationMath 215 HW #9 Solutions
Math 5 HW #9 Solutions. Problem 4.4.. If A is a 5 by 5 matrix with all a ij, then det A. Volumes or the big formula or pivots should give some upper bound on the determinant. Answer: Let v i be the ith
More informationA Brief Outline of Math 355
A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting
More informationDIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix
DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that
More informationEigenvalues and Eigenvectors. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB
Eigenvalues and Eigenvectors Consider the equation A x = λ x, where A is an nxn matrix. We call x (must be non-zero) an eigenvector of A if this equation can be solved for some value of λ. We call λ an
More informationMath 314H Solutions to Homework # 3
Math 34H Solutions to Homework # 3 Complete the exercises from the second maple assignment which can be downloaded from my linear algebra course web page Attach printouts of your work on this problem to
More informationTopics in linear algebra
Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over
More informationMTH50 Spring 07 HW Assignment 7 {From [FIS0]}: Sec 44 #4a h 6; Sec 5 #ad ac 4ae 4 7 The due date for this assignment is 04/05/7 Sec 44 #4a h Evaluate the erminant of the following matrices by any legitimate
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationLecture 19: The Determinant
Math 108a Professor: Padraic Bartlett Lecture 19: The Determinant Week 10 UCSB 2013 In our last class, we talked about how to calculate volume in n-dimensions Specifically, we defined a parallelotope:
More informationHomework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.
Math 5327 Fall 2018 Homework 7 1. For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. 3 1 0 (a) A = 1 2 0 1 1 0 x 3 1 0 Solution: 1 x 2 0
More informationThe minimal polynomial
The minimal polynomial Michael H Mertens October 22, 2015 Introduction In these short notes we explain some of the important features of the minimal polynomial of a square matrix A and recall some basic
More informationDIFFERENTIAL EQUATIONS REVIEW. Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it.
DIFFERENTIAL EQUATIONS REVIEW PEYAM RYAN TABRIZIAN Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it. Welcome to the special Friday after-school special of That
More informationColumbus State Community College Mathematics Department Public Syllabus
Columbus State Community College Mathematics Department Public Syllabus Course and Number: MATH 2568 Elementary Linear Algebra Credits: 4 Class Hours Per Week: 4 Prerequisites: MATH 2153 with a C or higher
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More information0 2 0, it is diagonal, hence diagonalizable)
MATH 54 TRUE/FALSE QUESTIONS FOR MIDTERM 2 SOLUTIONS PEYAM RYAN TABRIZIAN 1. (a) TRUE If A is diagonalizable, then A 3 is diagonalizable. (A = P DP 1, so A 3 = P D 3 P = P D P 1, where P = P and D = D
More informationEigenvalues & Eigenvectors
Eigenvalues & Eigenvectors Page 1 Eigenvalues are a very important concept in linear algebra, and one that comes up in other mathematics courses as well. The word eigen is German for inherent or characteristic,
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationFinal Exam Practice Problems Answers Math 24 Winter 2012
Final Exam Practice Problems Answers Math 4 Winter 0 () The Jordan product of two n n matrices is defined as A B = (AB + BA), where the products inside the parentheses are standard matrix product. Is the
More informationChapter 4 & 5: Vector Spaces & Linear Transformations
Chapter 4 & 5: Vector Spaces & Linear Transformations Philip Gressman University of Pennsylvania Philip Gressman Math 240 002 2014C: Chapters 4 & 5 1 / 40 Objective The purpose of Chapter 4 is to think
More informationMath 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.
Math 453 Exam 3 Review The syllabus for Exam 3 is Chapter 6 (pages -2), Chapter 7 through page 37, and Chapter 8 through page 82 in Axler.. You should be sure to know precise definition of the terms we
More information= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E.
3.3 Diagonalization Let A = 4. Then and are eigenvectors of A, with corresponding eigenvalues 2 and 6 respectively (check). This means 4 = 2, 4 = 6. 2 2 2 2 Thus 4 = 2 2 6 2 = 2 6 4 2 We have 4 = 2 0 0
More informationDesigning Information Devices and Systems I Fall 2018 Homework 5
Last Updated: 08-09-9 0:6 EECS 6A Designing Information Devices and Systems I Fall 08 Homework 5 This homework is due September 8, 08, at 3:59. Self-grades are due October, 08, at 3:59. Submission Format
More informationSolution Set 7, Fall '12
Solution Set 7, 18.06 Fall '12 1. Do Problem 26 from 5.1. (It might take a while but when you see it, it's easy) Solution. Let n 3, and let A be an n n matrix whose i, j entry is i + j. To show that det
More informationFINAL EXAM Ma (Eakin) Fall 2015 December 16, 2015
FINAL EXAM Ma-00 Eakin Fall 05 December 6, 05 Please make sure that your name and GUID are on every page. This exam is designed to be done with pencil-and-paper calculations. You may use your calculator
More informationAppendix A. Vector addition: - The sum of two vectors is another vector that also lie in the space:
Tor Kjellsson Stockholm University Appendix A A.1 Q. Consider the ordinary vectors in 3 dimensions (a x î+a y ĵ+a zˆk), with complex components. Do the following subsets constitute a vector space? If so,
More informationMath Lecture 26 : The Properties of Determinants
Math 2270 - Lecture 26 : The Properties of Determinants Dylan Zwick Fall 202 The lecture covers section 5. from the textbook. The determinant of a square matrix is a number that tells you quite a bit about
More informationMath 18.6, Spring 213 Problem Set #6 April 5, 213 Problem 1 ( 5.2, 4). Identify all the nonzero terms in the big formula for the determinants of the following matrices: 1 1 1 2 A = 1 1 1 1 1 1, B = 3 4
More informationUNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if
UNDERSTANDING THE DIAGONALIZATION PROBLEM Roy Skjelnes Abstract These notes are additional material to the course B107, given fall 200 The style may appear a bit coarse and consequently the student is
More information5.3.5 The eigenvalues are 3, 2, 3 (i.e., the diagonal entries of D) with corresponding eigenvalues. Null(A 3I) = Null( ), 0 0
535 The eigenvalues are 3,, 3 (ie, the diagonal entries of D) with corresponding eigenvalues,, 538 The matrix is upper triangular so the eigenvalues are simply the diagonal entries, namely 3, 3 The corresponding
More informationFrom Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play?
Overview Last week introduced the important Diagonalisation Theorem: An n n matrix A is diagonalisable if and only if there is a basis for R n consisting of eigenvectors of A. This week we ll continue
More informationMA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS
MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS 1. (7 pts)[apostol IV.8., 13, 14] (.) Let A be an n n matrix with characteristic polynomial f(λ). Prove (by induction) that the coefficient of λ n 1 in f(λ) is
More informationHomework 3 Solutions Math 309, Fall 2015
Homework 3 Solutions Math 39, Fall 25 782 One easily checks that the only eigenvalue of the coefficient matrix is λ To find the associated eigenvector, we have 4 2 v v 8 4 (up to scalar multiplication)
More informationMATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by
MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar
More informationTopic 15 Notes Jeremy Orloff
Topic 5 Notes Jeremy Orloff 5 Transpose, Inverse, Determinant 5. Goals. Know the definition and be able to compute the inverse of any square matrix using row operations. 2. Know the properties of inverses.
More information