Linear Equations in Linear Algebra

Similar documents
Matrix Algebra 2.1 MATRIX OPERATIONS Pearson Education, Inc.

Linear Equations in Linear Algebra

Matrix operations Linear Algebra with Computer Science Application

ICS 6N Computational Linear Algebra Matrix Algebra

Matrix Arithmetic. a 11 a. A + B = + a m1 a mn. + b. a 11 + b 11 a 1n + b 1n = a m1. b m1 b mn. and scalar multiplication for matrices via.

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

Chapter 2 Notes, Linear Algebra 5e Lay

1/25/2018 LINEAR INDEPENDENCE LINEAR INDEPENDENCE LINEAR INDEPENDENCE LINEAR INDEPENDENCE

Review Let A, B, and C be matrices of the same size, and let r and s be scalars. Then

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

Math 4377/6308 Advanced Linear Algebra

Linear Algebra Tutorial for Math3315/CSE3365 Daniel R. Reynolds

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Linear Algebra and Matrix Inversion

Linear Independence x

Systems of Linear Equations and Matrices

Systems of Linear Equations and Matrices

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )

Matrix Algebra. Matrix Algebra. Chapter 8 - S&B

Linear Algebra Summary. Based on Linear Algebra and its applications by David C. Lay

Math 54 Homework 3 Solutions 9/

Elementary maths for GMT

Chapter 1 Vector Spaces

Fall Inverse of a matrix. Institute: UC San Diego. Authors: Alexander Knop

Kevin James. MTHSC 3110 Section 2.1 Matrix Operations

7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

2.1 Matrices. 3 5 Solve for the variables in the following matrix equation.

Review : Powers of a matrix

MATRICES AND MATRIX OPERATIONS

Section 9.2: Matrices.. a m1 a m2 a mn

n n matrices The system of m linear equations in n variables x 1, x 2,..., x n can be written as a matrix equation by Ax = b, or in full

Math 2331 Linear Algebra

Section 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.

MATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:

CS 246 Review of Linear Algebra 01/17/19

Linear Equations in Linear Algebra

Linear algebra comments. Sophie Marques

Linear Algebra. Linear Equations and Matrices. Copyright 2005, W.R. Winfrey

Matrices Gaussian elimination Determinants. Graphics 2009/2010, period 1. Lecture 4: matrices

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

Graduate Mathematical Economics Lecture 1

Undergraduate Mathematical Economics Lecture 1

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

Lecture Summaries for Linear Algebra M51A

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Math 54 HW 4 solutions

Math 3191 Applied Linear Algebra

Finite Mathematics Chapter 2. where a, b, c, d, h, and k are real numbers and neither a and b nor c and d are both zero.

MAT 2037 LINEAR ALGEBRA I web:

a11 a A = : a 21 a 22

MTH 464: Computational Linear Algebra

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics

Matrix Arithmetic. j=1

Chapter 7. Linear Algebra: Matrices, Vectors,

Matrix & Linear Algebra

CLASS 12 ALGEBRA OF MATRICES

Introduction to Matrix Algebra

Section 12.4 Algebra of Matrices

MATH Mathematics for Agriculture II

Matrix Algebra Determinant, Inverse matrix. Matrices. A. Fabretti. Mathematics 2 A.Y. 2015/2016. A. Fabretti Matrices

APPENDIX: MATHEMATICAL INDUCTION AND OTHER FORMS OF PROOF

Linear Algebra V = T = ( 4 3 ).

Mathematics 13: Lecture 10

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

Math Linear Algebra Final Exam Review Sheet

Announcements Wednesday, October 10

Announcements Monday, October 02

1300 Linear Algebra and Vector Geometry

Chapter 1: Systems of Linear Equations

Matrix Basic Concepts

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Chapter 2: Matrix Algebra

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Math 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

MATH2210 Notebook 2 Spring 2018

ECON 186 Class Notes: Linear Algebra

UNIT 1 DETERMINANTS 1.0 INTRODUCTION 1.1 OBJECTIVES. Structure

Final Exam Practice Problems Answers Math 24 Winter 2012

Linear Algebra: Lecture notes from Kolman and Hill 9th edition.

Chapter 3. Vector spaces

LINEAR SYSTEMS AND MATRICES

Mon Feb Matrix algebra and matrix inverses. Announcements: Warm-up Exercise:

Phys 201. Matrices and Determinants

Study Guide for Linear Algebra Exam 2

BASIC NOTIONS. x + y = 1 3, 3x 5y + z = A + 3B,C + 2D, DC are not defined. A + C =

[ Here 21 is the dot product of (3, 1, 2, 5) with (2, 3, 1, 2), and 31 is the dot product of

Review of Linear Algebra

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

ELEMENTARY LINEAR ALGEBRA

Math 60. Rumbos Spring Solutions to Assignment #17

Appendix A: Matrices

Math 220 F11 Lecture Notes

Midterm 1 Review. Written by Victoria Kala SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015

POLI270 - Linear Algebra

Matrices and Matrix Algebra.

Math113: Linear Algebra. Beifang Chen

Transcription:

1 Linear Equations in Linear Algebra 1.7 LINEAR INDEPENDENCE

LINEAR INDEPENDENCE Definition: An indexed set of vectors {v 1,, v p } in n is said to be linearly independent if the vector equation x x x 1 1 2 2 p p v + v +... + v = 0 has only the trivial solution. The set {v 1,, v p } is said to be linearly dependent if there exist weights c 1,, c p, not all zero, such that c c c 1 1 2 2 p p v + v +... + v = 0 ----(1) Slide 1.7-2

LINEAR INDEPENDENCE Equation (1) is called a linear dependence relation among v 1,, v p when the weights are not all zero. An indexed set is linearly dependent if and only if it is not linearly independent. 1 v = 2 1 3 4 v = 5 2 6 2 v = 1 3 0 Example 1: Let,, and. Slide 1.7-3

a. Determine if the set {v 1, v 2, v 3 } is linearly independent. b. If possible, find a linear dependence relation among v 1, v 2, and v 3. Solution: We must determine if there is a nontrivial solution of the following equation. 1 4 2 0 x 2 + x 5 + x 1 = 0 1 2 3 3 6 0 0 Slide 1.7-4

LINEAR INDEPENDENCE Row operations on the associated augmented matrix show that! # # # " 1 4 2 0 2 5 1 0 3 6 0 0 $! & # & # & # % " 1 4 2 0 0 3 3 0 0 0 0 0 x 1 and x 2 are basic variables, and x 3 is free. Each nonzero value of x 3 determines a nontrivial solution of (1). Hence, v 1, v 2, v 3 are linearly dependent.. $ & & & % Slide 1.7-5

LINEAR INDEPENDENCE b. To find a linear dependence relation among v 1, v 2, and v 3, row reduce the augmented matrix and write the new system: 1 0 2 0 0 1 1 0 0 0 0 0 2x = 0 1 3 2 3 Thus, x = 2x, x = x 1 3 2 3, and x 3 is free. Choose any nonzero value for x 3 say, x = 5. 3 Then x = 10 and x = 5. 1 2 x x + x = 0 0= 0 Slide 1.7-6

LINEAR INDEPENDENCE Substitute these values into equation (1) and obtain the equation below. 10v 5v + 5v = 0 1 2 3 This is one (out of infinitely many) possible linear dependence relations among v 1, v 2, and v 3. Slide 1.7-7

LINEAR INDEPENDENCE OF MATRIX COLUMNS Suppose that we begin with a matrix instead of a set of vectors. A =! a 1 a "# n $ %& The matrix equation A x= 0 xa + x a +... + x a 0 1 1 2 2 n n = can be written as. Each linear dependence relation among the columns of A corresponds to a nontrivial solution of. A x= 0 Thus, the columns of matrix A are linearly independent if and only if the equation A x= 0 has only the trivial solution. Slide 1.7-8

SETS OF ONE OR TWO VECTORS A set containing only one vector say, v is linearly independent if and only if v is not the zero vector. This is because the vector equation the trivial solution when. v 0 x v= 0 1 has only The zero vector is linearly dependent because has many nontrivial solutions. x 0= 0 1 Slide 1.7-9

SETS OF ONE OR TWO VECTORS A set of two vectors {v 1, v 2 } is linearly dependent if at least one of the vectors is a multiple of the other. The set is linearly independent if and only if neither of the vectors is a multiple of the other. Slide 1.7-10

SETS OF TWO OR MORE VECTORS Theorem 7: Characterization of Linearly Dependent Sets S = {v,...,v } An indexed set 1 p of two or more vectors is linearly dependent if and only if at least one of the vectors in S is a linear combination of the others. v 0 In fact, if S is linearly dependent and 1, then some v j (with j >1 ) is a linear combination of the preceding vectors, v 1,,. v j 1 Slide 1.7-11

SETS OF TWO OR MORE VECTORS Proof: If some v j in S equals a linear combination of the other vectors, then v j can be subtracted from both sides of the equation, producing a linear dependence relation with a nonzero weight ( 1) on v j. [For instance, if v = c v + c v 1 2 2 3 3 Thus S is linearly dependent., then 0 = ( 1)v 1 + c 2 v 2 + c 3 v 3 + 0v 4 +...+ 0v p Conversely, suppose S is linearly dependent. If v 1 is zero, then it is a (trivial) linear combination of the other vectors in S..] Slide 1.7-12

SETS OF TWO OR MORE VECTORS v 0 Otherwise, 1, and there exist weights c 1,, c p, not all zero, such that c c c 1 1 2 2 p p v + v +... + v = 0 Let j be the largest subscript for which. j = 1 c v = 0 1 1 v 0. If, then, which is impossible because 1 j. c 0 Slide 1.7-13

SETS OF TWO OR MORE VECTORS So j >1, and cv +... + c v + 0v + 0v +... + 0v = 0 1 1 j j j j+ 1 p c v = cv... c v j j 1 1 j 1 j 1 c c 1 j 1 v = v... v. j + + 1 j 1 c c j j Slide 1.7-14

SETS OF TWO OR MORE VECTORS Theorem 7 does not say that every vector in a linearly dependent set is a linear combination of the preceding vectors. A vector in a linearly dependent set may fail to be a linear combination of the other vectors. 3 u = 1 0 1 v= 6 0 Example 2: Let and. Describe the set spanned by u and v, and explain why a vector w is in Span {u, v} if and only if {u, v, w} is linearly dependent. Slide 1.7-15

SETS OF TWO OR MORE VECTORS Solution: The vectors u and v are linearly independent because neither vector is a multiple of the other, and so they span a plane in 3. Span {u, v} is the x 1 x 2 -plane (with x = ). 3 0 If w is a linear combination of u and v, then {u, v, w} is linearly dependent, by Theorem 7. Conversely, suppose that {u, v, w} is linearly dependent. By theorem 7, some vector in {u, v, w} is a linear combination of the preceding vectors (since ). u 0 That vector must be w, since v is not a multiple of u. Slide 1.7-16

SETS OF TWO OR MORE VECTORS So w is in Span {u, v}. See the figures given below. Example 2 generalizes to any set {u, v, w} in u and v linearly independent. 3 The set {u, v, w} will be linearly dependent if and only if w is in the plane spanned by u and v. with Slide 1.7-17

SETS OF TWO OR MORE VECTORS Theorem 8: If a set contains more vectors than there are entries in each vector, then the set is linearly dependent. That is, any set {v 1,, v p } in n is linearly dependent if p> n. Proof: Let A =! v 1 v $ "# p %&. Then A is n p, and the equation A x= 0 corresponds to a system of n equations in p unknowns. p > n If, there are more variables than equations, so there must be a free variable. Slide 1.7-18

SETS OF TWO OR MORE VECTORS A x= 0 Hence has a nontrivial solution, and the columns of A are linearly dependent. See the figure below for a matrix version of this theorem. Theorem 8 says nothing about the case in which the number of vectors in the set does not exceed the number of entries in each vector. Slide 1.7-19

SETS OF TWO OR MORE VECTORS S = {v,...,v } n Theorem 9: If a set 1 p in contains the zero vector, then the set is linearly dependent. Proof: By renumbering the vectors, we may suppose v = 0 1. Then the equation that S in linearly dependent. 1v + 0v +... + 0v = 0 1 2 p shows Slide 1.7-20

2 Matrix Algebra 2.1 MATRIX OPERATIONS

MATRIX OPERATIONS If A is an m n matrix that is, a matrix with m rows and n columns then the scalar entry in the ith row and jth column of A is denoted by a ij and is called the (i, j)-entry of A. See the figure below. Each column of A is a list of m real numbers, which identifies a vector in. m Slide 2.1-2

MATRIX OPERATIONS The columns are denoted by a 1,, a n, and the matrix A is written as A =! a 1 a 2 a "# The number a ij is the ith entry (from the top) of the jth column vector a j. m n The diagonal entries in an matrix are a 11, a 22, a 33,, and they form the main diagonal of A. A diagonal matrix is a sequence nondiagonal entries are zero. n n An example is the identity matrix, I n. n $ %& A aij n m. = matrix whose Slide 2.1-3

SUMS AND SCALAR MULTIPLES An m n matrix whose entries are all zero is a zero matrix and is written as 0. The two matrices are equal if they have the same size (i.e., the same number of rows and the same number of columns) and if their corresponding columns are equal, which amounts to saying that their corresponding entries are equal. m n A+ B If A and B are matrices, then the sum is the m n matrix whose columns are the sums of the corresponding columns in A and B. Slide 2.1-4

SUMS AND SCALAR MULTIPLES Since vector addition of the columns is done entrywise, each entry in A+ B is the sum of the corresponding entries in A and B. The sum same size. A+ B Example 1: Let C is defined only when A and B are the 4 0 5 1 1 1 A=, B=, 1 3 2 3 5 7 2 3 = 0 1 A+ B A+ C and. Find and. Slide 2.1-5

SUMS AND SCALAR MULTIPLES 5 1 6 A+ B= 2 8 9 A+ C Solution: but is not defined because A and C have different sizes. If r is a scalar and A is a matrix, then the scalar multiple ra is the matrix whose columns are r times the corresponding columns in A. Theorem 1: Let A, B, and C be matrices of the same size, and let r and s be scalars. a. A+ B= B+ A Slide 2.1-6

SUMS AND SCALAR MULTIPLES b. c. d. e. f. ( A+ B) + C = A+ ( B+ C) A+ 0 = A r( A + B) = ra + rb ( r + s) A = ra + sa r( sa) = ( rs) A Each quantity in Theorem 1 is verified by showing that the matrix on the left side has the same size as the matrix on the right and that corresponding columns are equal. Slide 2.1-7

MATRIX MULTIPLICATION When a matrix B multiplies a vector x, it transforms x into the vector Bx. If this vector is then multiplied in turn by a matrix A, the resulting vector is A (Bx). See the Fig. below. Thus A (Bx) is produced from x by a composition of mappings the linear transformations. Slide 2.1-8

MATRIX MULTIPLICATION Our goal is to represent this composite mapping as multiplication by a single matrix, denoted by AB, so that. See the figure below. AB ( x)=(ab)x m n n p p If A is, B is, and x is in, denote the columns of B by b 1,, bp and the entries in x by x 1,, x p. Slide 2.1-9

MATRIX MULTIPLICATION Then Bx= xb +... + x b 1 1 p p By the linearity of multiplication by A, AB ( x) = Ax ( b ) +... + Ax ( b ) 1 1 = xab +... + x Ab 1 1 The vector A (Bx) is a linear combination of the vectors Ab 1,, Ab p, using the entries in x as weights. In matrix notation, this linear combination is written as A(Bx) =! Ab 1 Ab 2 Ab "# p p p p p $. %& x Slide 2.1-10

MATRIX MULTIPLICATION Thus multiplication by! "# transforms x into A (Bx). Ab 1 Ab 2 Ab p $ %& m n n p Definition: If A is an matrix, and if B is an matrix with columns b 1,, b p, then the product AB is the matrix whose columns are Ab 1,, Ab p. m p That is, AB = A! "# b 1 b 2 b Multiplication of matrices corresponds to composition of linear transformations. p $ %& =! Ab Ab 1 2 Ab "# p Slide 2.1-11 $ %&

MATRIX MULTIPLICATION Example 2: Compute AB, where B 4 3 9 = 1 2 3. A 2 3 = 1 5 and Solution: Write B = [ b b b ] 1 2 3, and compute: Slide 2.1-12

Ab 1 MATRIX MULTIPLICATION 2 3 4 = 1 5 1 11 = 1 2 3 3, Ab =, 2 1 5 2 0 = 13 Ab 3 2 3 6 = 1 5 3 21 = 9 Then AB [ b b b ] 11 0 21 = A = 1 2 3 1 13 9 Ab 1 Ab 2 Ab 3 Slide 2.1-13

MATRIX MULTIPLICATION Each column of AB is a linear combination of the columns of A using weights from the corresponding column of B. Row column rule for computing AB If a product AB is defined, then the entry in row i and column j of AB is the sum of the products of corresponding entries from row i of A and column j of B. If (AB) ij denotes the (i, j)-entry in AB, and if A is an m nmatrix, then. ( AB) = a b +... + a b ij i1 1j in nj Slide 2.1-14

PROPERTIES OF MATRIX MULTIPLICATION m n Theorem 2: Let A be an matrix, and let B and C have sizes for which the indicated sums and products are defined. ABC ( ) = ( ABC ) a. (associative law of multiplication) AB ( + C) = AB+ AC ( B + C) A = BA + CA r( AB) = ( ra) B = A( rb) I A= A= AI b. (left distributive law) c. (right distributive law) d. for any scalar r e. (identity for matrix m n multiplication) Slide 2.1-15

PROPERTIES OF MATRIX MULTIPLICATION Proof: Property (a) follows from the fact that matrix multiplication corresponds to composition of linear transformations (which are functions), and it is known that the composition of functions is associative. Let C =! c 1 c $ "# p%& By the definition of matrix multiplication, BC =! Bc 1 Bc $ "# p %& A(BC) =! A(Bc 1 ) A(Bc p ) "# $ %& Slide 2.1-16

PROPERTIES OF MATRIX MULTIPLICATION AB ( x) = ( AB)x The definition of AB makes for all x, so A(BC) =! (AB)c 1 (AB)c $ "# p %& = (AB)C The left-to-right order in products is critical because AB and BA are usually not the same. Because the columns of AB are linear combinations of the columns of A, whereas the columns of BA are constructed from the columns of B. The position of the factors in the product AB is emphasized by saying that A is right-multiplied by B or that B is left-multiplied by A. Slide 2.1-17

PROPERTIES OF MATRIX MULTIPLICATION AB = BA If, we say that A and B commute with one another. Warnings: AB BA 1. In general,. 2. The cancellation laws do not hold for matrix multiplication. That is, if AB = AC, then it is not true in general that. 3. If a product AB is the zero matrix, you cannot conclude in general that either or. B = C A = 0 B = 0 Slide 2.1-18

POWERS OF A MATRIX n n If A is an matrix and if k is a positive integer, then A k denotes the product of k copies of A: A k = A A If A is nonzero and if x is in, then A k x is the result of left-multiplying x by A repeatedly k times. k n If k = 0, then A 0 x should be x itself. Thus A 0 is interpreted as the identity matrix. Slide 2.1-19

THE TRANSPOSE OF A MATRIX m n n m Given an matrix A, the transpose of A is the matrix, denoted by A T, whose columns are formed from the corresponding rows of A. Theorem 3: Let A and B denote matrices whose sizes are appropriate for the following sums and products. a. b. c. For any scalar r, d. ( T T A ) = A ( A+ B) T = A T + B T ( ra) T = ra T ( AB) T = B T A T Slide 2.1-20

THE TRANSPOSE OF A MATRIX The transpose of a product of matrices equals the product of their transposes in the reverse order. Slide 2.1-21