Problem Set # 1 Solution, 18.06

Similar documents
v = ( 2)

18.06 Problem Set 1 Solutions Due Thursday, 11 February 2010 at 4 pm in Total: 100 points

Conceptual Questions for Review

4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1

Math Linear Algebra Final Exam Review Sheet

Lecture 7. Econ August 18

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Matrix Operations: Determinant

Contents. 1 Vectors, Lines and Planes 1. 2 Gaussian Elimination Matrices Vector Spaces and Subspaces 124

Math 240 Calculus III

MAC1105-College Algebra. Chapter 5-Systems of Equations & Matrices

6-3 Solving Systems by Elimination

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

22A-2 SUMMER 2014 LECTURE Agenda

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Computational Methods. Systems of Linear Equations

Extra Problems for Math 2050 Linear Algebra I

Linear Algebra Primer

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Lecture 7: Introduction to linear systems

COMP 175 COMPUTER GRAPHICS. Lecture 04: Transform 1. COMP 175: Computer Graphics February 9, Erik Anderson 04 Transform 1

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Name: MATH 3195 :: Fall 2011 :: Exam 2. No document, no calculator, 1h00. Explanations and justifications are expected for full credit.

x + 2y + 3z = 8 x + 3y = 7 x + 2z = 3

Honors Advanced Mathematics Determinants page 1

September 23, Chp 3.notebook. 3Linear Systems. and Matrices. 3.1 Solve Linear Systems by Graphing

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication.

MTAEA Vectors in Euclidean Spaces

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 4377/6308 Advanced Linear Algebra

Lecture 2: Vector-Vector Operations

Math 2331 Linear Algebra

7.6 The Inverse of a Square Matrix

System of Linear Equations

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

AMS526: Numerical Analysis I (Numerical Linear Algebra)

MTH 215: Introduction to Linear Algebra

General Physics I, Spring Vectors

y b where U. matrix inverse A 1 ( L. 1 U 1. L 1 U 13 U 23 U 33 U 13 2 U 12 1

Lecture 2 Systems of Linear Equations and Matrices, Continued

THE NULLSPACE OF A: SOLVING AX = 0 3.2

2. Every linear system with the same number of equations as unknowns has a unique solution.

Basic Concepts in Linear Algebra

POLI270 - Linear Algebra

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

web: HOMEWORK 1

6: Introduction to Elimination. Lecture

22A-2 SUMMER 2014 LECTURE 5

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Review of Basic Concepts in Linear Algebra

MATH Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product.

MAC Module 1 Systems of Linear Equations and Matrices I

MATH2210 Notebook 2 Spring 2018

(1) for all (2) for all and all

Ch. 7.3, 7.4: Vectors and Complex Numbers

Maths for Signals and Systems Linear Algebra for Engineering Applications

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

An overview of key ideas

11.1 Vectors in the plane

Linear Equations and Matrix

EE5120 Linear Algebra: Tutorial 1, July-Dec Solve the following sets of linear equations using Gaussian elimination (a)

Unit 1 Matrices Notes Packet Period: Matrices

4.5 Integration of Rational Functions by Partial Fractions

Linear Algebra I Lecture 8

Section 5.3 Systems of Linear Equations: Determinants

Linear Algebra and Matrix Inversion

2x + 5 = x = x = 4

Chapter 7. Tridiagonal linear systems. Solving tridiagonal systems of equations. and subdiagonal. E.g. a 21 a 22 a A =

Updated: January 16, 2016 Calculus II 7.4. Math 230. Calculus II. Brian Veitch Fall 2015 Northern Illinois University

Linear Algebra. The Manga Guide. Supplemental Appendixes. Shin Takahashi, Iroha Inoue, and Trend-Pro Co., Ltd.

10.1 Vectors. c Kun Wang. Math 150, Fall 2017

Volume in n Dimensions

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

[ Here 21 is the dot product of (3, 1, 2, 5) with (2, 3, 1, 2), and 31 is the dot product of

Solving Consistent Linear Systems

Lecture 6 & 7. Shuanglin Shao. September 16th and 18th, 2013

MATH 3511 Lecture 1. Solving Linear Systems 1

1 Last time: determinants

Exercises Chapter II.

Math for ML: review. CS 1675 Introduction to ML. Administration. Lecture 2. Milos Hauskrecht 5329 Sennott Square, x4-8845

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Ack: 1. LD Garcia, MTH 199, Sam Houston State University 2. Linear Algebra and Its Applications - Gilbert Strang

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

YOU CAN BACK SUBSTITUTE TO ANY OF THE PREVIOUS EQUATIONS

Applied Numerical Linear Algebra. Lecture 8

Chapter 4. Solving Systems of Equations. Chapter 4

4.3 Row operations. As we have seen in Section 4.1 we can simplify a system of equations by either:

pset2-sol September 7, 2017 a 2 3

Section 5.6. LU and LDU Factorizations

1.1 Single Variable Calculus versus Multivariable Calculus Rectangular Coordinate Systems... 4

MATH 22A: LINEAR ALGEBRA Chapter 1

1 Last time: linear systems and row operations

Lecture 22: Section 4.7

Section 13.4 The Cross Product

Dylan Zwick. Fall Ax=b. EAx=Eb. UxrrrEb

Matrices Gaussian elimination Determinants. Graphics 2009/2010, period 1. Lecture 4: matrices

MATH 304 Linear Algebra Lecture 20: Review for Test 1.

Lecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II)

Transcription:

Problem Set # 1 Solution, 1.06 For grading: Each problem worths 10 points, and there is points of extra credit in problem. The total maximum is 100. 1. (10pts) In Lecture 1, Prof. Strang drew the cone (infinite triangle) that comes from all combinations cv + dw with c 0 and d 0. Which c and d would give that triangle cut off by a top line from v to w? Which c and d give the parallelogram that starts with sides v and w? The points on the line connecting v and w, by vector addition, is represented by w + t(v w) = (1 t)w + tv, with 0 t 1 therefore the top boundary is characterized by c + d = 1, 0 c, 0 d. Together with the other boundary (v and w), the triangle is given by c + d 1, c 0, d 0. For the parallelogram, each point could be reached by starting from cv with 0 c 1 then go the other direction dw with 0 d 1. Therefore the answer is 0 c 1, 0 d 1. 2. (10pts) The length of v is v = v v = v 2 1 +... + v2 n. The dot product v w equals v w times the cosine of the angle between v and w. If v = and w =, what are the smallest and largest possible values of the dot product v w and of v w? From definition, v w = v w cos θ = 1 cos θ, where θ is the angle between v and w. Since cos θ ranges between 1 and 1, the smallest value of v w is 1, largest is 1. The smallest value is achieved when v and w are parallel and pointing to opposite direction, while the largest is when they are parallel and pointing to the same direction. For v w, we can use the triangle inequality (side 1 side 2 + side ): v w v w v + w which gives 2 v w. 1

Alternatively we can also use the definition of dot product: v w 2 = (v w) (v w) = v v 2v w+w w = v 2 2v w+ w 2 = 9 2v w+2 From the discussion above we know 1 v w 1, which gives us v w 2 6 that is 2 v w.. (10pts) The column vectors u = (1, 1, 2), v = (1, 2, ) and w = (,, ) are in a plane because w is what combination of u and v? Find two combinations of u, v, w that produce b = (0, 0, 0) and two combinations that produce b = (1, 1, c). What is the only possible number c that gives a vector on the plane? Solve xu + yv = w to get From above we know w = u + 2v. u + 2v w = (0, 0, 0) and multiplication by any constant gives the same result, for example 2u + v 2w = (0, 0, 0) (In fact, multiples of (1, 2, 1) are th only solutions we have.) To get the combination for b = (1, 1, c), by observation we get u 2v + 0w = (1, 1, 0) adding a copy of u + 2v w = (0, 0, 0) we get another copy u + 0v w = (1, 1, 0) (And (, 2, 0) + k(1, 2, 1) are all the solutions.) c = 0 is the only possible number to make (1, 1, c) lie on the plane. These problems come from Introduction to Linear Algebra (th edition). (10pts) Problem 19 on page 2 E = E 1 = 1 0 1 1 0 1 2

Let x = (,, ) Ex = 1 0 1 = then multiplying by E 1 E 1 (Ex) =. (10pts) Problem on page 1 0 1 = = x For a Sudoku matrix S, and x = (1, 1,..., 1), Sx is a column vector with 9 elements, all equal to : Sx = (,,...) There are 6 permutations of three numbers: (1,2,), (1,,2), (2,1,), (2,,1), (,1,2), (,2,1) as mentioned in Section 2.7. Group each three rows, i.e. row 1-, -6 and 7-9, and we can do row permutations inside each group, which would still give us Sudoku matrix. This in total gives 6 6 6 ways of creating new matrices. And exchange the order of the three row blocks will also give us Sudoku matrix. This gives another 6 ways of permutation. Combined with the row permutation inside each group, in total, we have 6 = 1296 orders of 9 rows that stay Sudoku. 6. (10pts) Problem on page 2 When k = we have x + y = 6 and x + y = 6, so after first step of elimination we ll have x + y = 6 and 0 = 12. Here elimination breaks down and we have 0 solutions. If k = we have x + y = 6 and x y = 6, so after first step we ll have x + y = 6 and 0 = 0. Here we have infinite number of solutions. When k = 0, we need to do a row exchange before elimination, and it gives one solution. 7. (10pts) Problem 11 on page (a) We could easily see that 1 2 (x + X, y + Y, z + Z) is also a solution. In fact, any combination of c(x, y, z) + d(x, Y, Z) with c + d is a solution. (b) The 2 planes also meet on the line containing the two points.. (10pts + Extra credit pts) Problem 26 on page (matrices with given row and column sums)

Adding the two equations of the 1st column gives a+b+c+d = 12, then subtract a + c = 2, we have s = b + d = 10. Two examples of matrices with these sums [ ] [ 1 0, 1 7 2 6 Extra credit: the system Ax = b is 1 1 Making A triangular: 1 1 1 a b c d ] = 2 10 So we can see that the elimination breaks down at the last step. There is a solution only if d = a + c b. The four columns of A lie in a D hyperplane. 9. (10pts) Problem 2 on page (a) Some linear combination of the 100 rows is the row of 100 zeros. (b) Some linear combination of the 100 columns is the column of zeros. (c) A very singular matrix has all ones. A better example has 99 random rows. The 100th row could be the sum of the first 99 rows (or any other combination of those rows with no zeros). (d) The row picture has 100 planes meeting along a common line through 0. The column picture has 100 vectors all in the same 99-dimensional hyperplane. 10. (10pts) Problem 29 on page 66 To eliminate column 1, we multiply by the following matrix E 1 = 1 1 0 0 0 0

The diagonal entries of E are all 1 s, and the only other nonzero entries are the 1st column. Here row 1 is subtracted from all other rows. Then we reduce the second column to all 1 s by the following matrix (considering block operations that only changes the bottom right by submatrix) And third column to E 2 = E = 0 0 1 0 1 0 0 All together, multiply E, E 2 and E 1 we get matrix E: E = E E 2 E 1 = So what E really does is that each row is subtracted from the next row. Similarly, for the smaller Pascal matrix, we use the following matrix to eliminate the second column: E = 0 Then the third column E = 0 0 All together we get M with MA = I. Then M must be the inverse of the Pascal matrix. M = E E E = 1 2 1 0 1 1