Solving a system by back-substitution, checking consistency of a system (no rows of the form

Similar documents
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

2. Every linear system with the same number of equations as unknowns has a unique solution.

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Study Guide for Linear Algebra Exam 2

LINEAR ALGEBRA SUMMARY SHEET.

Dimension. Eigenvalue and eigenvector

235 Final exam review questions

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Announcements Monday, October 29

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

ft-uiowa-math2550 Assignment OptionalFinalExamReviewMultChoiceMEDIUMlengthForm due 12/31/2014 at 10:36pm CST

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Math Linear Algebra Final Exam Review Sheet

SUMMARY OF MATH 1600

1 9/5 Matrices, vectors, and their applications

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

What is on this week. 1 Vector spaces (continued) 1.1 Null space and Column Space of a matrix

Solutions to Final Exam

W2 ) = dim(w 1 )+ dim(w 2 ) for any two finite dimensional subspaces W 1, W 2 of V.

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Chapter 5. Eigenvalues and Eigenvectors

Review Notes for Linear Algebra True or False Last Updated: February 22, 2010

Third Midterm Exam Name: Practice Problems November 11, Find a basis for the subspace spanned by the following vectors.

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Warm-up. True or false? Baby proof. 2. The system of normal equations for A x = y has solutions iff A x = y has solutions

Math 1553, Introduction to Linear Algebra

MTH 464: Computational Linear Algebra

Linear Algebra Final Exam Study Guide Solutions Fall 2012

Math113: Linear Algebra. Beifang Chen

Quizzes for Math 304

Lecture Summaries for Linear Algebra M51A

Eigenvalues, Eigenvectors, and Diagonalization

March 27 Math 3260 sec. 56 Spring 2018

Chapter 5 Eigenvalues and Eigenvectors

Linear Algebra Highlights

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Calculating determinants for larger matrices

1. Let A = (a) 2 (b) 3 (c) 0 (d) 4 (e) 1

Chapter 4 & 5: Vector Spaces & Linear Transformations

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

MATH 221, Spring Homework 10 Solutions

spring, math 204 (mitchell) list of theorems 1 Linear Systems Linear Transformations Matrix Algebra

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

Math 3108: Linear Algebra

Definitions for Quizzes

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

Linear Algebra Summary. Based on Linear Algebra and its applications by David C. Lay

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Math 54 HW 4 solutions

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Problem 1: Solving a linear equation

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Topic 2 Quiz 2. choice C implies B and B implies C. correct-choice C implies B, but B does not imply C

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Introduction to Linear Algebra, Second Edition, Serge Lange

LINEAR ALGEBRA REVIEW

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Chapter 3. Directions: For questions 1-11 mark each statement True or False. Justify each answer.

MA 265 FINAL EXAM Fall 2012

1. General Vector Spaces

PRACTICE PROBLEMS FOR THE FINAL

(a) only (ii) and (iv) (b) only (ii) and (iii) (c) only (i) and (ii) (d) only (iv) (e) only (i) and (iii)

Practice Final Exam. Solutions.

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Math 2114 Common Final Exam May 13, 2015 Form A

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Conceptual Questions for Review

Math Final December 2006 C. Robinson

Math 21b. Review for Final Exam

ft-uiowa-math2550 Assignment NOTRequiredJustHWformatOfQuizReviewForExam3part2 due 12/31/2014 at 07:10pm CST

Summer Session Practice Final Exam

MAT188H1S LINEAR ALGEBRA: Course Information as of February 2, Calendar Description:

1. Select the unique answer (choice) for each problem. Write only the answer.

MATH 310, REVIEW SHEET 2

MAT Linear Algebra Collection of sample exams

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

ANSWERS (5 points) Let A be a 2 2 matrix such that A =. Compute A. 2

Recall : Eigenvalues and Eigenvectors

WI1403-LR Linear Algebra. Delft University of Technology

ANSWERS. E k E 2 E 1 A = B

MATH 1553 SAMPLE FINAL EXAM, SPRING 2018

Math 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix

2 Eigenvectors and Eigenvalues in abstract spaces.

Linear Algebra March 16, 2019

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

MATH 1553 PRACTICE FINAL EXAMINATION

Transcription:

MATH 520 LEARNING OBJECTIVES SPRING 2017 BROWN UNIVERSITY SAMUEL S. WATSON Week 1 (23 Jan through 27 Jan) Definition of a system of linear equations, definition of a solution of a linear system, elementary row operations (switch rows, scale a row, and add to one row a multiple of another) Augmented matrix and matrix of coefficients of a system of equations, row echelon form (all entries below a leading entry are zero, and rows are sorted by the column of their leading entry), row reduction algorithm (move a row with leftmost leading entry to the top, and use that entry to zero-out the entries below it; repeat with remaining rows) Solving a system by back-substitution, checking consistency of a system (no rows of the form [0 0 ]) Week 2 (30 Jan through 3 Feb) Reduced row echelon form (row echelon plus leading entries all 1 plus all zeros above each leading entry), Solution of a system in terms of free variables Vector addition and scalar multiplication Definition of a linear combination, definition of the span of a collection of vectors Matrix form Ax = b of a system of linear equations, properties of the matrix-vector product The columns of an m n matrix span R m if and only if the matrix has a pivot entry in every row Week 3 (6 Feb through 10 Feb) Basic facts about homogeneous and nonhomogeneous systems: every homogeneous system Ax = b has at least the trivial solution x = 0, and nontrivial solutions as well if A has some non-pivot column A nonhomogeneous system can be inconsistent, but if it s consistent then its solution takes the form (particular solution) + (solution of homogeneous system) Applications: balancing economies, balancing chemical reactions, and traffic networks Linear independence: a list of vectors is linearly dependent if 0 is a nontrivial linear combination of those vectors; alternatively, if one of the vectors is in the span of the others. Geometrically, this means that the vectors are collinear (for two vectors) or coplanar (for three vectors) Week 4 (13 Feb through 17 Feb) Linear transformations map equally spaced lines to equally spaces lines (or points) and the origin to the origin. Basic examples in 2D: rotating, scaling, shearing, projecting, and reflecting. Multiplying a vector x R n on the left by a matrix A R m n represents a linear transformation, and every linear transformation from R n to R m can be represented that way The matrix representing a given linear transformation from R 2 to R m is obtained by letting its columns

be the vectors to which e 1 = (1, 0) and e 2 = (0, 1) are mapped (and similarly for any R n, e.g., for R 3 we use e 1 = (1, 0, 0), e 2 = (0, 1, 0), and e 3 = (0, 0, 1)) A matrix transformation from R n to R m is surjective (i.e., every vector in R m has a vector in R n mapping to it) iff the columns of the matrix span R m, and it s injective (i.e., no two vectors in R n map to the same output) iff all the columns of the matrix are pivot columns A linear transformation from R n to R m can only be bijective if m = n Matrix multiplication is defined for an m n matrix times an n p matrix; the result is an m p matrix which represents the composition of the matrices corresponding linear transformations: the (i, j)th entry of the product is the ith row of the first matrix dotted with the jth column of the second matrix (where dot means multiply in pairs and sum). Matrix multiplication is not commutative This table summarizes some helpful relationships; A is an m n matrix A has a pivot in every row A has a pivot in every column Ax = b has at least one solution, for all b R m Ax = 0 has only the trivial solution x = 0 T(x) = Ax is surjective (onto) T(x) = Ax is injective (one-to-one) Week 5 (20 Feb through 24 Feb) The inverse of a bijective linear transformation from R n to R n is again a linear transformation To find the inverse of A, we find the vectors that A maps to e 1, e 2,..., e n This is the same as row-reducing the augmented matrix [A I], where I is the n n identity matrix, and reading off the resulting matrix right of the bar. If the matrix left of the bar isn t the identity, then A wasn t invertible (AB) 1 = B 1 A 1 [ a b 2 2 matrix inversion formula: c d ] 1 [ = ad bc 1 d b c a ] (not invertible if ad bc = 0) You can solve equations whose variables represent matrices like you would if they represent numbers, except: (a) never use division: use matrix inverses instead, and (b) watch out for non-commutativity. So like C + BA 1 X = 0 implies X = AB 1 C Week 6 (27 Feb through 3 Mar) Definition of a linear subspace of R n (contains 0, closed under scalar multiplication and vector addition) The span of any list of vectors is a linear subspace The column space of a matrix is the span of its columns The null space of a matrix is the solution set of its homogeneous equation A basis of a linear subspace is a linearly indepent spanning list To find a basis for Nul A, solve the homogeneous system and write the solution in span form To find a basis for Col A, select all the pivot columns from A (not from the rref of A!)

Given a basis of a subspace, every vector in that subspace can be written uniquely as a linear combination of those basis vectors Every basis of a given linear subspace has the same number of vectors in it, and this number is called the dimension of the subspace The rank of a matrix is the dimension of the column space, and its nullity is the dimension of its null space The rank-nullity theorem tells us that rank plus nullity equals number of columns Week 7 (6 Mar through 10 Mar) A linear transformation T from R n to R n scales all n-dimensional volumes by the same factor: the (absolute value of the) determinant of T The sign of the determinant tells us whether the linear transformation reverses orientations, meaning that a small counterclockwise loop drawn on the outside of a closed surface maps to a counterclockwise loop (+) or a clockwise loop (-) [ ] a b The determinant of a 2 2 matrix is ad bc c d The determinant of a linear transformation can be computed from its matrix using cofactor expansions: (1) choose a row or column, (2) for each entry in that row or column, multiply (i) a checkerboardalternating sign, (ii) the matrix entry, and (iii) the determinant of the matrix obtained by removing that entry s row and column from the matrix. Finally, (3) sum all the terms from step 2 A rook arrangement is a selection of n positions in a n n matrix with the property that no two selected positions are in the same row or column The sign of a rook arrangement is the number of column switches it takes to get there from the maindiagonal rook arrangement The determinant of a matrix can also be computed as a sum over all rook arrangements on the matrix of the product of the entries in the rook arrangement s positions times the sign of the rook arrangement Row switching multiplies the determinant of a matrix by 1, row scaling scales the determinant, and row-add operations don t alter the determinant The determinant of an upper triangular matrix (all zeros below the main diagonal) is the product of the entries on the main diagonal (why? the main-diagonal rook arrangement is the only nonzero term in the rook expansion of the determinant) The computationally efficient way to compute determinants is to row reduce to echelon form, multiply down the diagonal, and account for any changes you made to the determinant whilst row reducing det AB = det A det B if A and B are both n n, because scaling volumes by a factor of k and then by a factor of l is equivalent to scaling volumes by a factor of kl det A 1 = (det A) 1, because if A scales volumes by a factor of k, then the inverse of A must scale volumes by a factor of 1/k Cramer s rule says that to find the ith component of the solution of Ax = b, we can replace the ith column of A with b, find the determinant of that matrix, and divide the result by det A Week 8 (13 Mar through 17 Mar) A vector space is a set V together with an operation + which tells us how to add elements of V and an

operation which tells us how to multiply elements of V by real numbers. These operations are required to satisfy a list of 10 basic properties, like commutativity and associativity of addition, etc. Working with abstract vector spaces is useful because it allows us to develop concepts we learned for R n in a way that permits their application to a variety of mathematical objects whose elements are not n-tuples of real numbers P n is the space of polynomials of degree at most n, with usual polynomial addition and scalar multiplication C([0, 1]) is the space of continuous functions from [0, 1] to R, with usual function addition and scalar multiplication A subspace of a vector space V is a subset of V which contains the zero vector and is closed under the vector operations of addition and scalar multiplication The span of a list of vectors in V is always a subspace of V A linear transformation T from V to W is a function which satisfies T(u + v) = T(u) + T(v) and T(cu) = ct(u) for all real numbers c and all u, v T The kernel of a linear transformation from V to W is the set of vectors that map to the zero vector in W The range of a linear transformation from V to W is the set of all vectors in W that get mapped to Linear independence of a list of vectors in an abstract vector space has the same definition as in R n The linear dependence lemma: a list of vectors is linearly dependent if and only if there is some vector in the list which is in the span of the vectors appearing before it in the list. Removing this vector does not change the span of the list Any linearly independent list in a vector space V is no longer than any list of vectors that spans V. Abbreviated: linearly independent lists are no longer than spanning lists Week 9 (20 Mar through 24 Mar) Each vector in a vector space V can be uniquely represented as a linear combination of the vectors in a given basis of V The weights in that linear combination are called the coordinates of the vector with respect to the given basis The coordinates of x with respect to B are denoted [x] B To convert the standard coordinates of v R n to coordinates with respect to the basis given by the columns of a matrix B, we left-multiply the standard coordinates by B 1 To convert from coordinates with respect to the columns of B to standard coordinates, we left-multiply the coordinates by B The coordinate mapping from V to R n which sends each vector to its coordinates with respect to some given (fixed) basis is called the coordinate mapping The coordinate mapping is linear and bijective, i.e., a vector space isomorphism Any linearly independent list of vectors in V can be extended to obtain a basis in V Any spanning list of vectors in V can be pruned to obtain a basis of V The rank of a linear transformation T : V W is the dimension of the range of T

The rank (or column rank) of an m n matrix is dimension of the matrix s column space, and the row rank is the dimension of the matrix s row space (the span of the matrix s rows) The row rank of any matrix equals its column rank The rank of an m n matrix A is equal to the smallest value of r such that A can be written as a product of an m r matrix and an r n matrix Week 11 (3 April through 7 April) If we have a vector represented via its coordinates with respect to some basis and we want its coordinates with respect to a different basis, then we left-multiply by the matix whose columns contain the coordinates of the old basis vectors with respect to the new ones The change of basis matrix from one basis B to another basis C is the inverse of the change of basis matrix from C back to B An eigenvector v of an n n matrix A is a vector with the property that Av = λv for some λ R (n other words, A maps v to a vector which is either zero or parallel to v) The eigenspace associated with a given eigenvalue λ is the set of all solutions to the homogeneous sytsem (A λi)v = 0, where I is the n n identity matrix If a matrix has n linearly independent eigenvectors, then A s action on R n can be understood simply. Each coordinate gets multiplied by the corresponding eigenvalue: A(c 1 v 1 + c n v n ) = c 1 λ 1 v 1 + c n λ n v n. This can be used to find a formula for the nth Fibonacci number in terms of the nth power of the golden ratio φ = 1+ 5 2. Week 12 (10 April through 14 April) Eigenvectors corresponding to distinct eigenvalues are linearly independent We can find eigenvalues by finding roots of the characteristic polynomial det(a λi) An n n matrix has at most n eigenvalues A matrix A is diagonalizable if A = PDP 1 for some invertible P and diagonal D Diagonalizability is equivalent to having n linearly independent eigenvectors If some eigenvalues of A are repeated roots of the characteristic polynomial, there might not be a basis of R n consisting of eigenvectors of A If two matrices A and B have the property that A = PBP 1 for some invertible matrix P, then we say A and B are similar Week 13 (17 April through 21 April) Complex numbers are expressions of the form a + bi where a and b are real numbers (so real numbers count as complex too) The conjugate of z = a + bi is z = a bi. The absolute value of z = a + bi is z = a 2 + b 2, and its argument is the angle between the positive x-axis and the pont (a, b)

These expressions plotted in the plane (at (a, b)), and they support all the standard arithmetic (multiplication, addition, multiplicative inverses, and additive inverses), with the rule that we reduce i 2 to 1 The characteristic polynomial of a matrix may have no real roots (such as the matrix of a 90-degree rotation in R 2 ), but every degree-n polynomial does have a complex root So every n n matrix A has a complex eigenvalue λ and a complex eigenvector (a vector v with n complex entries such that Av = λv) If λ is a complex eigenvalue with complex eigenvector v, then λ is a complex eigenvalue with eigenvector v. [ ] a b A matrix of the form represents a rotation with angle arg(a + bi) combined with a scaling by a b a factor of a 2 + b 2 Every 2 2 matrix with complex, non-real eigenvalues is similar to a scaling-rotation matrix (a matrix of the form described in the preceding item) The inner product of two vectors u and v in R n is defined to be the scalar u T v, also denoted u v and called the dot product u u = u 2 u v = u v cos θ, where θ is the angle between u and v (proof: law of cosines) u v = 0 if and only if u and v are orthogonal The dot product is linear: u (cv + w) = cu v + u w) A vector is called a unit vector if it has length 1 If V is a subspace of R n, then the orthogonal complement V is the set of vectors orthogonal to every vector in V V is a subspace If A is an m n matrix, then the orthogonal complement of the row space of A is the null space of A, and the orthogonal complement of the column space of A is the null space of A T We can put this together into the following schematic diagram:

Row(A) A Col(A) v u Nul(A) 0 0 u + v Nul(A T ) R n R m A list of vectors in R n is orthogonal if any pair of distinct vectors from the list is orthogonal Every orthogonal list is linearly independent The coordinates of a vector v with respect to an orthogonal basis {v 1,..., v n } are easy to compute using the inner product: if c j = v v j v j v j for all j from 1 to n, then v = c 1 v 1 + + c n v n. Week 14 (24 April through 28 April) The orthogonal projection of y R n onto the line spanned by a vector u R n is calculated by proj u y = y u u u u The orthogonal projection of y R n onto the space W spanned by a list of orthogonal vectors u 1,..., u p is calculated by proj W y = y u 1 u u 1 u 1 + + y u p u p 1 u p u p An orthonormal list is an orthogonal list all of whose vectors have length 1 A matrix whose columns are orthonormal is called an orthogonal matrix (yes, this is confusing) The orthogonal projection of y onto W is the point in W which is nearest to y If U is an orthogonal n p matrix, then UU T is the matrix of the transformation which projects each vector in R n onto the p-dimensional subspace of R n spanned by the columns of U To come up with a list of vectors b 1, b 2,..., b p which as the same span as some specified linearly inde-

pendent list v 1, v 2,..., v p, we successively subtract off orthogonal projections: b 1 = v 1 etc. b 2 = v 2 v 2 b 1 b 1 b 1 b 1 b 3 = v 3 v 3 b 1 b 1 b 1 b 1 v 3 b 2 b 2 b 2 b 2 The least squares problem in statistics asks for the slope and y-intercept of the line which best fits a collection of data points which are ordered pairs {(x i, y i )} i=1 n, in the sense that the sum of squared errors is minimized 1 x 1 1 x 2 We let A =.. and b = 1 x n (y 1 α βx 1 ) 2 + (y 2 α βx 2 ) 2 + + (y n α βx n ) 2 y 1 y 2. y n and x = [ α β ], so we re looking to minimize Ax b 2 To minimize Ax b 2, we find the orthogonal projection b of b onto the column space of A and choose a value of x satisfying Ax = b b is characterized by A T (b b) = 0 The solution of this equation is x = (A T A) 1 A T b if A T A is invertible