Announcements Wednesday, September 20

Similar documents
Announcements Monday, September 17

Announcements Monday, September 18

Announcements Wednesday, November 15

Section 1.7. Linear Independence

Announcements Monday, November 13

Announcements Monday, November 13

Announcements Wednesday, August 30

Announcements Wednesday, August 30

Math 1553 Introduction to Linear Algebra. School of Mathematics Georgia Institute of Technology

Announcements Wednesday, September 05

Announcements Wednesday, October 04

Announcements Wednesday, October 25

Announcements August 31

Announcements Wednesday, September 27

Announcements Wednesday, November 01

Announcements Monday, November 19

Announcements Monday, November 19

Announcements Monday, October 29

Linear Independence Reading: Lay 1.7

Announcements Monday, September 25

Announcements Wednesday, November 15

Math 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix

Announcements Wednesday, November 7

Announcements Wednesday, November 01

Review for Chapter 1. Selected Topics

MAT 1302B Mathematical Methods II

Linear Algebra MATH20F Midterm 1

Announcements Wednesday, November 7

Midterm 1 Review. Written by Victoria Kala SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015

Math 1553 Introduction to Linear Algebra. School of Mathematics Georgia Institute of Technology

Section 1.5. Solution Sets of Linear Systems

Announcements Monday, November 20

Announcements September 19

Announcements Monday, November 26

Section 3.3. Matrix Equations

Announcements Monday, November 26

MATH 1553, JANKOWSKI MIDTERM 2, SPRING 2018, LECTURE A

Lecture 03. Math 22 Summer 2017 Section 2 June 26, 2017

Solutions to Math 51 Midterm 1 July 6, 2016

Topic 14 Notes Jeremy Orloff

Math 54 First Midterm Exam, Prof. Srivastava September 23, 2016, 4:10pm 5:00pm, 155 Dwinelle Hall.

Basic Linear Algebra Ideas. We studied linear differential equations earlier and we noted that if one has a homogeneous linear differential equation

Shorts

MATH240: Linear Algebra Review for exam #1 6/10/2015 Page 1

Announcements Wednesday, October 10

Announcements Monday, October 02

Chapter 6. Orthogonality and Least Squares

MATH 1553, SPRING 2018 SAMPLE MIDTERM 2 (VERSION B), 1.7 THROUGH 2.9

MATH 1553, FALL 2018 SAMPLE MIDTERM 2: 3.5 THROUGH 4.4

Week #4: Midterm 1 Review

MAT 1302B Mathematical Methods II

APPM 2360 Exam 2 Solutions Wednesday, March 9, 2016, 7:00pm 8:30pm

Math 1553 Introduction to Linear Algebra. School of Mathematics Georgia Institute of Technology

Announcements Wednesday, September 06

Linear Independence x

Welcome to Math 102 Section 102

Exam 2 Solutions. (a) Is W closed under addition? Why or why not? W is not closed under addition. For example,

Math 301 Test I. M. Randall Holmes. September 8, 2008

Math 290, Midterm II-key

Math 21b: Linear Algebra Spring 2018

LECTURES 14/15: LINEAR INDEPENDENCE AND BASES

Announcements Monday, November 06

EXAM. Exam #2. Math 2360 Summer II, 2000 Morning Class. Nov. 15, 2000 ANSWERS

Section Instructors: by now you should be scheduled into one of the following Sections:

Announcements: Nov 10

MATH 152 Exam 1-Solutions 135 pts. Write your answers on separate paper. You do not need to copy the questions. Show your work!!!

1 Linear systems, existence, uniqueness

Chapter 1: Linear Equations

MA 242 LINEAR ALGEBRA C1, Solutions to First Midterm Exam

Math 2802 Applications of Linear Algebra. School of Mathematics Georgia Institute of Technology

Chapter 1: Linear Equations

Worksheet for Lecture 15 (due October 23) Section 4.3 Linearly Independent Sets; Bases

Section 6.1. Inner Product, Length, and Orthogonality

Math 308 Midterm, Feb 11th 2015 Winter Form Bonus. Possible (6) 63

ASTR 101L: Motion of the Sun Take Home Lab

Derivatives and series FTW!

School of Mathematics Georgia Institute of Technology. Math 1553 Introduction to Linear Algebra

Math 31 Lesson Plan. Day 5: Intro to Groups. Elizabeth Gillaspy. September 28, 2011

Linear algebra and differential equations (Math 54): Lecture 10

Math 220: Summer Midterm 1 Questions

MATH 3330 INFORMATION SHEET FOR TEST 3 SPRING Test 3 will be in PKH 113 in class time, Tues April 21

Math 240, 4.3 Linear Independence; Bases A. DeCelles. 1. definitions of linear independence, linear dependence, dependence relation, basis

Fall 2016 MATH*1160 Final Exam

( v 1 + v 2 ) + (3 v 1 ) = 4 v 1 + v 2. and ( 2 v 2 ) + ( v 1 + v 3 ) = v 1 2 v 2 + v 3, for instance.

if b is a linear combination of u, v, w, i.e., if we can find scalars r, s, t so that ru + sv + tw = 0.

MAT188H1S LINEAR ALGEBRA: Course Information as of February 2, Calendar Description:

Math 250B Midterm I Information Fall 2018

LECTURE 1- ROTATION. Phys 124H- Honors Analytical Physics IB Chapter 10 Professor Noronha-Hostler

Linear Algebra. Introduction. Marek Petrik 3/23/2017. Many slides adapted from Linear Algebra Lectures by Martin Scharlemann

Math 369 Exam #2 Practice Problem Solutions

b for the linear system x 1 + x 2 + a 2 x 3 = a x 1 + x 3 = 3 x 1 + x 2 + 9x 3 = 3 ] 1 1 a 2 a

2018 Fall 2210Q Section 013 Midterm Exam I Solution

Linear Equations in Linear Algebra

Section 4.5. Matrix Inverses

MATH 1553-C MIDTERM EXAMINATION 3

Chapter 5. Eigenvalues and Eigenvectors

Determine whether the following system has a trivial solution or non-trivial solution:

Problem set #4. Due February 19, x 1 x 2 + x 3 + x 4 x 5 = 0 x 1 + x 3 + 2x 4 = 1 x 1 x 2 x 4 x 5 = 1.

Announcements Tuesday, April 17

Math 2940: Prelim 1 Practice Solutions

Transcription:

Announcements Wednesday, September 20 WeBWorK 1.4, 1.5 are due on Wednesday at 11:59pm. The first midterm is on this Friday, September 22. Midterms happen during recitation. The exam covers through 1.5. About half the problems will be conceptual, and the other half computational. There is a practice midterm posted on the website. It is identical in format to the real midterm (although there may be ±1 2 problems). Study tips: There are lots of problems at the end of each section in the book, and at the end of the chapter, for practice. Make sure to learn the theorems and learn the definitions, and understand what they mean. There is a reference sheet on the website. Sit down to do the practice midterm in 50 minutes, with no notes. Come to office hours! Double office hours this week: Tuesday, 9 11am; Wednesday, 1 3pm; Thursday, 9 11am and 2 4pm (note none on Monday). TA review sessions: check your email.

Section 1.7 Linear Independence

Motivation Sometimes the span of a set of vectors is smaller than you expect from the number of vectors. Span{v, w} w v Span{u, v, w} w v u This means that (at least) one of the vectors is redundant: you re using too many vectors to describe the span. Notice in each case that one vector in the set is already in the span of the others so it doesn t make the span bigger. Today we will formalize this idea in the concept of linear (in)dependence.

Definition A set of vectors {v 1, v 2,..., v p} in R n is linearly independent if the vector equation x 1v 1 + x 2v 2 + + x pv p = 0 has only the trivial solution x 1 = x 2 = = x p = 0. The set {v 1, v 2,..., v p} is linearly dependent otherwise. In other words, {v 1, v 2,..., v p} is linearly dependent if there exist numbers x 1, x 2,..., x p, not all equal to zero, such that x 1v 1 + x 2v 2 + + x pv p = 0. This is called a linear dependence relation or an equation of linear dependence. Like span, linear (in)dependence is another one of those big vocabulary words that you absolutely need to learn. Much of the rest of the course will be built on these concepts, and you need to know exactly what they mean in order to be able to answer questions on quizzes and exams (and solve real-world problems later on).

Definition A set of vectors {v 1, v 2,..., v p} in R n is linearly independent if the vector equation x 1v 1 + x 2v 2 + + x pv p = 0 has only the trivial solution x 1 = x 2 = = x p = 0. The set {v 1, v 2,..., v p} is linearly dependent otherwise. Note that linear (in)dependence is a notion that applies to a collection of vectors, not to a single vector, or to one vector in the presence of some others.

Checking Linear Independence Question: 1 1 3 Is 1, 1, 1 linearly independent? 1 2 4 Equivalently, does the (homogeneous) the vector equation 1 1 3 0 x 1 + y 1 + z 1 = 0 1 2 4 0 have a nontrivial solution? How do we solve this kind of vector equation? 1 1 3 1 1 1 row reduce 1 0 2 0 1 1 1 2 4 0 0 0 So x = 2z and y = z. So the vectors are linearly dependent, and an equation of linear dependence is (taking z = 1) 1 1 3 0 2 1 1 + 1 = 0. 1 2 4 0 [interactive]

Checking Linear Independence Question: 1 1 3 Is 1, 1, 1 linearly independent? 2 2 4 Equivalently, does the (homogeneous) the vector equation 1 1 3 0 x 1 + y 1 + z 1 = 0 2 2 4 0 have a nontrivial solution? 1 1 3 1 1 1 row reduce 1 0 0 0 1 0 2 2 4 0 0 1 x 0 The trivial solution y = 0 is the unique solution. So the vectors are z 0 linearly independent. [interactive]

and Matrix Columns In general, {v 1, v 2,..., v p} is linearly independent if and only if the vector equation x 1v 1 + x 2v 2 + + x pv p = 0 has only the trivial solution, if and only if the matrix equation Ax = 0 has only the trivial solution, where A is the matrix with columns v 1, v 2,..., v p: A = v 1 v 2 v p. This is true if and only if the matrix A has a pivot in each column. Important The vectors v 1, v 2,..., v p are linearly independent if and only if the matrix with columns v 1, v 2,..., v p has a pivot in each column. Solving the matrix equation Ax = 0 will either verify that the columns v 1, v 2,..., v p of A are linearly independent, or will produce a linear dependence relation.

Criterion Suppose that one of the vectors {v 1, v 2,..., v p} is a linear combination of the other ones (that is, it is in the span of the other ones): v 3 = 2v 1 1 v2 + 6v4 2 Then the vectors are linearly dependent: 2v 1 1 v2 v3 + 6v4 = 0. 2 Conversely, if the vectors are linearly dependent 2v 1 1 v2 + 6v4 = 0. 2 then one vector is a linear combination of (in the span of) the other ones: v 2 = 4v 1 + 12v 4. Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if one of the vectors is in the span of the other ones.

Another criterion Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if one of the vectors is in the span of the other ones. Equivalently: Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if you can remove one of the vectors without shrinking the span. Indeed, if v 2 = 4v 1 + 12v 3, then a linear combination of v 1, v 2, v 3 is x 1v 1 + x 2v 2 + x 3v 3 = x 1v 1 + x 2(4v 1 + 12v 3) + x 3v 3 = (x 1 + 4x 2)v 1 + (12x 2 + x 3)v 3, which is already in Span{v 1, v 3}. Conclution: v 2 was redundant.

Increasing span criterion Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if one of the vectors is in the span of the other ones. Better Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if there is some j such that v j is in Span{v 1, v 2,..., v j 1 }. Equivalently, {v 1, v 2,..., v p} is linearly independent if for every j, the vector v j is not in Span{v 1, v 2,..., v j 1 }. This means Span{v 1, v 2,..., v j } is bigger than Span{v 1, v 2,..., v j 1 }. Translation A set of vectors is linearly independent if and only if, every time you add another vector to the set, the span gets bigger.

Increasing span criterion: justification Better Theorem A set of vectors {v 1, v 2,..., v p} is linearly dependent if and only if there is some j such that v j is in Span{v 1, v 2,..., v j 1 }. Why? Take the largest j such that v j is in the span of the others. Then v j is in the span of v 1, v 2,..., v j 1. Why? If not (j = 3): v 3 = 2v 1 1 v2 + 6v4 2 Rearrange: (2v 1 12 v2 v3 ) v 4 = 1 6 so v 4 works as well, but v 3 was supposed to be the last one that was in the span of the others.

Pictures in R 2 One vector {v}: Linearly independent if v 0. v Span{v} [interactive 2D: 2 vectors] [interactive 2D: 3 vectors]

Pictures in R 2 Span{w} w v One vector {v}: Linearly independent if v 0. Two vectors {v, w}: Linearly independent Neither is in the span of the other. Span got bigger. Span{v} [interactive 2D: 2 vectors] [interactive 2D: 3 vectors]

Pictures in R 2 Span{w} Span{v, w} v w u Span{v} [interactive 2D: 2 vectors] [interactive 2D: 3 vectors] One vector {v}: Linearly independent if v 0. Two vectors {v, w}: Linearly independent Neither is in the span of the other. Span got bigger. Three vectors {v, w, u}: Linearly dependent: u is in Span{v, w}. Span didn t get bigger after adding u. Can remove u without shrinking the span. Also v is in Span{u, w} and w is in Span{u, v}.

Pictures in R 2 Span{v} w v Two collinear vectors {v, w}: Linearly dependent: w is in Span{v}. Can remove w without shrinking the span. Span didn t get bigger when we added w. Observe: Two vectors are linearly dependent if and only if they are collinear. [interactive 2D: 2 vectors] [interactive 2D: 3 vectors]

Pictures in R 2 Three vectors {v, w, u}: Linearly dependent: w is in Span{u, v}. v Can remove w without shrinking the span. Span{v} w u Span didn t get bigger when we added w. Observe: If a set of vectors is linearly dependent, then so is any larger set of vectors! [interactive 2D: 2 vectors] [interactive 2D: 3 vectors]

Pictures in R 3 Span{w} w v Two vectors {v, w}: Linearly independent: Neither is in the span of the other. Span got bigger when we added w. Span{v} [interactive 3D: 2 vectors] [interactive 3D: 3 vectors]

Pictures in R 3 Span{v, w} Span{w} u v Three vectors {v, w, u}: Linearly independent: span got bigger when we added u. w Span{v} [interactive 3D: 2 vectors] [interactive 3D: 3 vectors]

Pictures in R 3 Span{v, w} Span{w} Three vectors {v, w, x}: Linearly dependent: x v x is in Span{v, w}. w Can remove x without shrinking the span. Span didn t get bigger when we added x. Span{v} [interactive 3D: 2 vectors] [interactive 3D: 3 vectors]

Poll Poll Are there four vectors u, v, w, x in R 3 which are linearly dependent, but such that u is not a linear combination of v, w, x? If so, draw a picture; if not, give an argument. Yes: actually the pictures on the previous slides provide such an example. Linear dependence of {v 1,..., v p} means some v i is a linear combination of the others, not any.

Linear Dependence and Free Variables Theorem Let v 1, v 2,..., v p be vectors in R n, and consider the matrix A = v 1 v 2 v p. Then you can delete the columns of A without pivots (the columns corresponding to free variables), without changing Span{v 1, v 2,..., v p}. The pivot columns are linearly independent, so you can t delete any more columns. This means that each time you add a pivot column, then the span increases. Upshot Let d be the number of pivot columns in the matrix A above. If d = 1 then Span{v 1, v 2,..., v p} is a line. If d = 2 then Span{v 1, v 2,..., v p} is a plane. If d = 3 then Span{v 1, v 2,..., v p} is a 3-space. Etc.

Linear Dependence and Free Variables Justification Why? If the matrix is in RREF: 1 0 2 0 A = 0 1 3 0 0 0 0 1 then the column without a pivot is in the span of the pivot columns: 2 1 0 0 3 = 2 0 + 3 1 + 0 0 0 0 0 1 and the pivot columns are linearly independent: 0 1 0 0 x 1 0 = x 1 0 + x 2 1 + x 4 0 = x 2 = x 1 = x 2 = x 4 = 0. 0 0 0 1 x 4

Linear Dependence and Free Variables Justification Why? If the matrix is not in RREF, then row reduce: 1 7 23 3 A = 2 4 16 0 RREF 1 0 2 0 0 1 3 0 1 2 8 4 0 0 0 1 The following vector equations have the same solution set: 1 7 23 3 x 1 2 + x 2 4 + x 3 16 + x 4 0 = 0 1 2 8 4 1 0 2 0 x 1 0 + x 2 1 + x 3 3 + x 4 0 = 0 0 0 0 1 We conclude that and that x 1 1 2 1 23 16 = 2 8 + x 2 1 2 1 + x 4 7 4 2 + 3 7 4 2 3 + 0 0 4 3 0 = 0 has only the trivial solution. 4

Labeling the span Theorem Suppose that {v 1, v 2,..., v p} is linearly independent. Then every vector w in Span{v 1, v 2,..., v p} can be written in exactly one way as a linear combination w = a 1v 1 + a 2v 2 + + a pv p. Why? If you could write w in two different ways: w = a 1v 1 + a 2v 2 + a 3v 3 = b 1v 1 + b 2v 2 + b 3v 3 then you can subtract the second equation from the first to obtain 0 = (a 1 b 1)v 1 + (a 2 b 2)v 2 + (a 3 b 3)v 3. But the only solution to this vector equation is the trivial solution: a 1 b 1 = 0 a 2 b 2 = 0 a 3 b 3 = 0 which implies a 1 = b 1, a 2 = b 2, a 3 = b 3.

Labeling the span Theorem Suppose that {v 1, v 2,..., v p} is linearly independent. Then every vector w in Span{v 1, v 2,..., v p} can be written in exactly one way as a linear combination w = a 1v 1 + a 2v 2 + + a pv p. This means that we can use R p to label the points of Span{v 1, v 2,..., v p}: (a 1, a 2,..., a p) labels a 1v 1 + a 2v 2 + + a pv p. Span{v, w} w v [interactive] In the picture, Span{v, w} looks like R 2, with v and w playing the roles of (1, 0) and (0, 1), respectively. This plane is not R 2, but we are using the elements of R 2 to label the points on the plane. More on this later (the B-basis).

Two more facts Fact 1: Say v 1, v 2,..., v n are in R m. If n > m then {v 1, v 2,..., v n} is linearly dependent: the matrix A = v 1 v 2 v n. cannot have a pivot in each column (it is too wide). This says you can t have 4 linearly independent vectors in R 3, for instance. A wide matrix can t have linearly independent columns. Fact 2: If one of v 1, v 2,..., v n is zero, then {v 1, v 2,..., v n} is linearly dependent. For instance, if v 1 = 0, then is a linear dependence relation. 1 v 1 + 0 v 2 + 0 v 3 + + 0 v n = 0 A set containing the zero vector is linearly dependent.

Summary A set of vectors is linearly independent if removing one of the vectors shrinks the span; otherwise it s linearly dependent. There are several other criteria for linear (in)dependence which lead to pretty pictures. The columns of a matrix are linearly independent if and only if the RREF of the matrix has a pivot in every column. The pivot columns of a matrix A are linearly independent, and you can delete the non-pivot columns (the free columns) without changing the column span. Wide matrices cannot have linearly independent columns. Warning These are not the official definitions!