x 1 + 2x 2 + 3x 3 = 0 x 1 + 2x 2 + 3x 3 = 0, x 2 + x 3 = 0 x 3 3 x 3 1
|
|
- Claribel Summers
- 5 years ago
- Views:
Transcription
1 . Orthogonal Complements and Projections In this section we discuss orthogonal complements and orthogonal projections. The orthogonal complement of a subspace S is the complement that is orthogonal to S. Then the orthogonal projection onto S is the projection onto S with respect to its orthogonal complement. Consider any set of vectors E R n. What happens if we look at all of the vectors orthogonal to each vector in E denoted E (pronounced E perp )? Definition.. For E R n let E = { x R n x y = 0 for all y E}. When S is a subspace we call S the orthogonal complement of S. and For example 2 = x x x x 2 = 0 = 0 2 x x 2 x x = 0 = span 0. 0 x x x 0 = 2 = 0 = 0 x x x x = x x x = 0 + x = 0 = span. Notice that E is a subspace in these two examples which is not a coincidence. Proposition.2. For any E R n E is a subspace. Proof. We show that E satisfies the three properties of being a subspace. (a) We have 0 E since 0 y = 0 for all y E. Thus 0 E. (b) Suppose x E. Then for all y E ( x + ) y = x y + y = = 0 since x E. Therefore x + y 2 E and E is closed under addition. (c) Suppose x E and c is a scalar. Then for all y E (c x) y = c( x y) = c(0) = 0 since x E. Therefore c x E and E is closed under scalar multiplication. However if E is a nonzero subspace we have to solve an infinite set of equations to find E according to the definition. How can we possibly find E for a nonzero subspace? We need to realize that there is a lot of redudancy in these equations: For example if x y then x c y for all scalars c. In fact we can reduce the set of equations to just being orthogonal to a basis for S. Proposition.. For any { v... v k } R n (span{ v... v k }) = { v... v k }.
2 2 Proof. Let S = span{ v... v k }. First if x S then x v = = x v k = 0 because v... v k S. Thus x { v... v k } and S { v... v k }. On the other hand if x { v... v k } then x v = = x v k = 0 Now for any y S we can write y = c v + + c k v k so x y = x (c v + + c k v k ) = c ( x v ) + + c k ( x v k ) = c (0) + + c k (0) = 0. Thus x S and { v... v k } S. This shows S = { v... v k }. We will use Proposition?? shortly to describe how to find the orthogonal complement of a subspace more systematically. Recall that the column space of a matrix A denote col(a) is the span of its columns. It now becomes of use to us to also look at the span of the rows of a matrix. Let row(a) denote the span of the rows of matrix A. For example row ([ 2 ]) 0 0 = span 2. Finally let the transpose A of a matrix A denote A with its rows and columns interchanged. For example [ ] 0 2 = 2 0 Notice that col(a) = row(a ) and that (A ) = A. Now we can formalize how to find the orthogonal complement of s subspace. Theorem.4. For any matrix A row(a) = ker(a) and col(a) = ker(a ). Proof. Let v... v m denote the rows of m n matrix. Then row(a) = span{ v... v m } and row(a) = (span{ v... v m }) = { v... v m } = { x R n x v = 0... x v m = 0}. Now notice that x v = 0... x v m = 0 are exactly the same questions we wolve when we are finding ker(a). For example if v = 2 then x x v = 0 2 = 0 x x = 0 x which is exactly the equation row represents in finding ker(a). Thus ker(a) = { x R n x v = 0... x v m = 0}. as well. Hence row(a) = ker(a). Finally notice that col(a) = row(a ) = ker(a ). Examples : Find bases for the following orthogonal complements. (a) span 2 4
3 0 (b) span 0 2 Solutions: (a) (b) span = row ([ 2 4 ]) ([ ]) = ker 2 4 = span span 2 = row = ker ([ ]) ([ ]) = ker ([ 0 4 ]) = span In fact as you may have guessed from the name S has a relationship to S that we have already seen namely that of a subspace complement. {[ As ]} opposed to {[ just ]} any subspace {[ ]} complement the orthogonal complement to S is unique. For example span span span are all subspace complements {[ {[ {[ 0 0 of span but only span is the orthogonal complement of span. 0]} ]} ]} S S Theorem.5. For any subspace S R n S and S are complements in R n. Proof. First we show that S S = 0. Suppose x S S. Then since x S and x S x must be orthogonal to itself: x x = 0 = x = 0 by the definiteness of the dot product. Therefore S S = 0. Next we show that S + S = n using dimension. Let { v... v k } be a basis for S so that dim(s) = k. Then S = (span { v... v k }) = { v... v k } = { x R n x v = 0... x v k = 0}.
4 4 So S is the set of solutions to a homogeneous system of k equations in n variables which means there are at most k leading variables and so at least n k free variables Hence dim(s ( n k. Then dim(s + S ) = dim(s) + dim(s ) dim(s S ) = k + dim(s ) 0 k + (n k) = n. But S + S R n so dim(s + S ) n as well. This forces dim(s + S ) = n and thus S + S = R n. Now that we know S and S are complements we can discuss the orthogonal projection π SS onto S with respect to S. For shorthand let π S := π SS. In order to describe a formula for the orthogonal projection we will use the following fact to be proven in the next section: Theorem.6. Every subspace of R n has an orthogonal basis. Proof. See next section. Proposition.7. Let S R n be an orthogonal basis for R n and { u... u k } be an orthogonal basis for S. Then x u π S ( x) = u 2 u + + u k 2 u k Proof. Let { v... v n k } be an orthogonal basis for S. As you will see for homeowork this makes { u... u k v... v n k } an orthogonal basis for R n. This means for any x R n x span{ u... u k v... v n k }. Hence by Proposition?? ( ) ( ) x u x v x vn k x = u 2 u + + u k 2 u k + v 2 v + + v n k 2 v n k. Since and we can conclude that x u u 2 u + + u k 2 u k S x v x vn k v 2 v + + v n k 2 v n k S π S ( x) = π SS ( x) = x u u 2 u + + u k 2 u k. Exercises: () Find bases for the following orthogonal complements. 2 (a) span (b) span (c) span 5 7
5 6 (d) span (e) span (2) Find formulae for π S ( x) for each of the following subspaces S. (a) S = span 2. (b) S = span 2. (c) S = span. 2 2 (d) S = span (e) S = span Problems:. (2) Suppose S R n is a subspace { u... u k } is an orthogonal basis for S and { v... v n k } is an orthogonal basis for S. Show that { u... u k v... v n k } is orthogonal basis for R n. 2. (2) Fix x R n and a subspace S R n. Show that π S ( x) is the unique vector in R n so that (i) π S ( x) S (ii) x π S ( x) S.. () Let S be a subspace of R n. Show that for all x y R n x y = π S ( x) π S ( y) + π S ( x) π S ( y) 4. () Show that (S ) = S for any subspace S of R n. 5. () (a) Show that if A C (A and C differ by EROs) then row(a) = row(c). (b) Show that the nonzero rows in RREF(A) form a basis for row(a). 6. () Show that for any matrix A rank(a) = rank(a ). 7. (2) Prove or give a counterexample: For any matrix A nullity(a) = nullity(a ). 8. (2) Show that if the rows of m n matrix A are linearly independent then rank(a) = m and nullity(a) = n m.
6 6 9. () Suppose { a... a n } is a basis for R n and b... b n are scalars. Show that there exists a unique x R n so that x a = b... x a n = b n. 0. () Suppose { u... u k } is orthonormal in R n and let S = span{ u... u k }. Then for x R n show that x S if and only if x 2 = ( x u ) ( x u k ) 2.. (4) Fix x R n and a subspace S R n. Show that the minimum of x y 2 over all y S is acheived uniquely when y = π S ( x). 2. (4) Suppose P : R n R n is a projection map which satisfies Show that P is an orthogonal projection. P ( x) x for all x R n.
Solutions: We leave the conversione between relation form and span form for the reader to verify. x 1 + 2x 2 + 3x 3 = 0
6.2. Orthogonal Complements and Projections In this section we discuss orthogonal complements and orthogonal projections. The orthogonal complement of a subspace S is the set of all vectors orthgonal to
More informationSolution: (a) S 1 = span. (b) S 2 = R n, x 1. x 1 + x 2 + x 3 + x 4 = 0. x 4 Solution: S 5 = x 2. x 3. (b) The standard basis vectors
.. Dimension In this section, we introduce the notion of dimension for a subspace. For a finite set, we can measure its size by counting its elements. We are interested in a measure of size on subspaces
More informationThe Fundamental Theorem of Linear Algebra
The Fundamental Theorem of Linear Algebra Nicholas Hoell Contents 1 Prelude: Orthogonal Complements 1 2 The Fundamental Theorem of Linear Algebra 2 2.1 The Diagram........................................
More informationWorksheet for Lecture 23 (due December 4) Section 6.1 Inner product, length, and orthogonality
Worksheet for Lecture (due December 4) Name: Section 6 Inner product, length, and orthogonality u Definition Let u = u n product or dot product to be and v = v v n be vectors in R n We define their inner
More informationMarch 27 Math 3260 sec. 56 Spring 2018
March 27 Math 3260 sec. 56 Spring 2018 Section 4.6: Rank Definition: The row space, denoted Row A, of an m n matrix A is the subspace of R n spanned by the rows of A. We now have three vector spaces associated
More informationElementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.
Elementary Linear Algebra Review for Exam Exam is Monday, November 6th. The exam will cover sections:.4,..4, 5. 5., 7., the class notes on Markov Models. You must be able to do each of the following. Section.4
More informationPRACTICE PROBLEMS FOR THE FINAL
PRACTICE PROBLEMS FOR THE FINAL Here are a slew of practice problems for the final culled from old exams:. Let P be the vector space of polynomials of degree at most. Let B = {, (t ), t + t }. (a) Show
More informationSolutions to Math 51 First Exam April 21, 2011
Solutions to Math 5 First Exam April,. ( points) (a) Give the precise definition of a (linear) subspace V of R n. (4 points) A linear subspace V of R n is a subset V R n which satisfies V. If x, y V then
More informationof A in U satisfies S 1 S 2 = { 0}, S 1 + S 2 = R n. Examples 1: (a.) S 1 = span . 1 (c.) S 1 = span, S , S 2 = span 0 (d.
. Complements and Projection Maps In this section, we explore the notion of subspaces being complements. Then, the unique decomposition of vectors in R n into two pieces associated to complements lets
More informationReview Notes for Linear Algebra True or False Last Updated: February 22, 2010
Review Notes for Linear Algebra True or False Last Updated: February 22, 2010 Chapter 4 [ Vector Spaces 4.1 If {v 1,v 2,,v n } and {w 1,w 2,,w n } are linearly independent, then {v 1 +w 1,v 2 +w 2,,v n
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More informationMath 4A Notes. Written by Victoria Kala Last updated June 11, 2017
Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...
More informationSolutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015
Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See
More information2018 Fall 2210Q Section 013 Midterm Exam II Solution
08 Fall 0Q Section 0 Midterm Exam II Solution True or False questions points 0 0 points) ) Let A be an n n matrix. If the equation Ax b has at least one solution for each b R n, then the solution is unique
More informationNotice that the set complement of A in U satisfies
Complements and Projection Maps In this section, we explore the notion of subspaces being complements Then, the unique decomposition of vectors in R n into two pieces associated to complements lets us
More information(v, w) = arccos( < v, w >
MA322 F all206 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: Commutativity:
More informationELE/MCE 503 Linear Algebra Facts Fall 2018
ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2
More informationSUMMARY OF MATH 1600
SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You
More informationMath 123, Week 5: Linear Independence, Basis, and Matrix Spaces. Section 1: Linear Independence
Math 123, Week 5: Linear Independence, Basis, and Matrix Spaces Section 1: Linear Independence Recall that every row on the left-hand side of the coefficient matrix of a linear system A x = b which could
More informationMath 102, Winter 2009, Homework 7
Math 2, Winter 29, Homework 7 () Find the standard matrix of the linear transformation T : R 3 R 3 obtained by reflection through the plane x + z = followed by a rotation about the positive x-axes by 6
More informationLecture 3q Bases for Row(A), Col(A), and Null(A) (pages )
Lecture 3q Bases for Row(A), Col(A), and Null(A) (pages 57-6) Recall that the basis for a subspace S is a set of vectors that both spans S and is linearly independent. Moreover, we saw in section 2.3 that
More informationOverview. Motivation for the inner product. Question. Definition
Overview Last time we studied the evolution of a discrete linear dynamical system, and today we begin the final topic of the course (loosely speaking) Today we ll recall the definition and properties of
More informationMath Linear Algebra
Math 220 - Linear Algebra (Summer 208) Solutions to Homework #7 Exercise 6..20 (a) TRUE. u v v u = 0 is equivalent to u v = v u. The latter identity is true due to the commutative property of the inner
More informationAnswers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3
Answers in blue. If you have questions or spot an error, let me know. 3 4. Find all matrices that commute with A =. 4 3 a b If we set B = and set AB = BA, we see that 3a + 4b = 3a 4c, 4a + 3b = 3b 4d,
More informationFinal Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015
Final Review Written by Victoria Kala vtkala@mathucsbedu SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015 Summary This review contains notes on sections 44 47, 51 53, 61, 62, 65 For your final,
More informationLINEAR ALGEBRA REVIEW
LINEAR ALGEBRA REVIEW SPENCER BECKER-KAHN Basic Definitions Domain and Codomain. Let f : X Y be any function. This notation means that X is the domain of f and Y is the codomain of f. This means that for
More informationx 2 For example, Theorem If S 1, S 2 are subspaces of R n, then S 1 S 2 is a subspace of R n. Proof. Problem 3.
.. Intersections and Sums of Subspaces Until now, subspaces have been static objects that do not interact, but that is about to change. In this section, we discuss the operations of intersection and addition
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationChapter 6. Orthogonality and Least Squares
Chapter 6 Orthogonality and Least Squares Section 6.1 Inner Product, Length, and Orthogonality Orientation Recall: This course is about learning to: Solve the matrix equation Ax = b Solve the matrix equation
More informationSection 6.1. Inner Product, Length, and Orthogonality
Section 6. Inner Product, Length, and Orthogonality Orientation Almost solve the equation Ax = b Problem: In the real world, data is imperfect. x v u But due to measurement error, the measured x is not
More information(v, w) = arccos( < v, w >
MA322 Sathaye Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v
More informationWorksheet for Lecture 25 Section 6.4 Gram-Schmidt Process
Worksheet for Lecture Name: Section.4 Gram-Schmidt Process Goal For a subspace W = Span{v,..., v n }, we want to find an orthonormal basis of W. Example Let W = Span{x, x } with x = and x =. Give an orthogonal
More informationMATH 15a: Linear Algebra Practice Exam 2
MATH 5a: Linear Algebra Practice Exam 2 Write all answers in your exam booklet. Remember that you must show all work and justify your answers for credit. No calculators are allowed. Good luck!. Compute
More informationMATH 240 Spring, Chapter 1: Linear Equations and Matrices
MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear
More information(v, w) = arccos( < v, w >
MA322 F all203 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v,
More informationTypical Problem: Compute.
Math 2040 Chapter 6 Orhtogonality and Least Squares 6.1 and some of 6.7: Inner Product, Length and Orthogonality. Definition: If x, y R n, then x y = x 1 y 1 +... + x n y n is the dot product of x and
More informationv = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :
Length, Angle and the Inner Product The length (or norm) of a vector v R 2 (viewed as connecting the origin to a point (v 1,v 2 )) is easily determined by the Pythagorean Theorem and is denoted v : v =
More information7. Dimension and Structure.
7. Dimension and Structure 7.1. Basis and Dimension Bases for Subspaces Example 2 The standard unit vectors e 1, e 2,, e n are linearly independent, for if we write (2) in component form, then we obtain
More informationLINEAR ALGEBRA: THEORY. Version: August 12,
LINEAR ALGEBRA: THEORY. Version: August 12, 2000 13 2 Basic concepts We will assume that the following concepts are known: Vector, column vector, row vector, transpose. Recall that x is a column vector,
More informationMATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.
MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:
More informationMATH Linear Algebra
MATH 304 - Linear Algebra In the previous note we learned an important algorithm to produce orthogonal sequences of vectors called the Gramm-Schmidt orthogonalization process. Gramm-Schmidt orthogonalization
More informationLecture 9: Vector Algebra
Lecture 9: Vector Algebra Linear combination of vectors Geometric interpretation Interpreting as Matrix-Vector Multiplication Span of a set of vectors Vector Spaces and Subspaces Linearly Independent/Dependent
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More information2. (10 pts) How many vectors are in the null space of the matrix A = 0 1 1? (i). Zero. (iv). Three. (ii). One. (v).
Exam 3 MAS 3105 Applied Linear Algebra, Spring 2018 (Clearly!) Print Name: Apr 10, 2018 Read all of what follows carefully before starting! 1. This test has 7 problems and is worth 110 points. Please be
More informationMath 3191 Applied Linear Algebra
Math 191 Applied Linear Algebra Lecture 1: Inner Products, Length, Orthogonality Stephen Billups University of Colorado at Denver Math 191Applied Linear Algebra p.1/ Motivation Not all linear systems have
More informationDefinitions for Quizzes
Definitions for Quizzes Italicized text (or something close to it) will be given to you. Plain text is (an example of) what you should write as a definition. [Bracketed text will not be given, nor does
More informationLecture 13: Row and column spaces
Spring 2018 UW-Madison Lecture 13: Row and column spaces 1 The column space of a matrix 1.1 Definition The column space of matrix A denoted as Col(A) is the space consisting of all linear combinations
More informationChapter 6. Orthogonality
6.4 The Projection Matrix 1 Chapter 6. Orthogonality 6.4 The Projection Matrix Note. In Section 6.1 (Projections), we projected a vector b R n onto a subspace W of R n. We did so by finding a basis for
More informationVector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then
1 Def: V be a set of elements with a binary operation + is defined. F be a field. A multiplication operator between a F and v V is also defined. The V is called a vector space over the field F if: V is
More informationSolutions to Final Exam
Solutions to Final Exam. Let A be a 3 5 matrix. Let b be a nonzero 5-vector. Assume that the nullity of A is. (a) What is the rank of A? 3 (b) Are the rows of A linearly independent? (c) Are the columns
More informationChapter SSM: Linear Algebra. 5. Find all x such that A x = , so that x 1 = x 2 = 0.
Chapter Find all x such that A x : Chapter, so that x x ker(a) { } Find all x such that A x ; note that all x in R satisfy the equation, so that ker(a) R span( e, e ) 5 Find all x such that A x 5 ; x x
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More information22m:033 Notes: 6.1 Inner Product, Length and Orthogonality
m:033 Notes: 6. Inner Product, Length and Orthogonality Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman April, 00 The inner product Arithmetic is based on addition and
More informationMath 1180, Notes, 14 1 C. v 1 v n v 2. C A ; w n. A and w = v i w i : v w = i=1
Math 8, 9 Notes, 4 Orthogonality We now start using the dot product a lot. v v = v v n then by Recall that if w w ; w n and w = v w = nx v i w i : Using this denition, we dene the \norm", or length, of
More informationConsider a subspace V = im(a) of R n, where m. Then,
5.4 LEAST SQUARES AND DATA FIT- TING ANOTHER CHARACTERIZATION OF ORTHOG- ONAL COMPLEMENTS Consider a subspace V = im(a) of R n, where A = [ ] v 1 v 2... v m. Then, V = { x in R n : v x = 0, for all v in
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationChapter 4 Euclid Space
Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,
More informationOrthogonal Complements
Orthogonal Complements Definition Let W be a subspace of R n. If a vector z is orthogonal to every vector in W, then z is said to be orthogonal to W. The set of all such vectors z is called the orthogonal
More informationDefinition Suppose S R n, V R m are subspaces. A map U : S V is linear if
.6. Restriction of Linear Maps In this section, we restrict linear maps to subspaces. We observe that the notion of linearity still makes sense for maps whose domain and codomain are subspaces of R n,
More informationChapter 2 Subspaces of R n and Their Dimensions
Chapter 2 Subspaces of R n and Their Dimensions Vector Space R n. R n Definition.. The vector space R n is a set of all n-tuples (called vectors) x x 2 x =., where x, x 2,, x n are real numbers, together
More informationMatrix invertibility. Rank-Nullity Theorem: For any n-column matrix A, nullity A +ranka = n
Matrix invertibility Rank-Nullity Theorem: For any n-column matrix A, nullity A +ranka = n Corollary: Let A be an R C matrix. Then A is invertible if and only if R = C and the columns of A are linearly
More information4.3 - Linear Combinations and Independence of Vectors
- Linear Combinations and Independence of Vectors De nitions, Theorems, and Examples De nition 1 A vector v in a vector space V is called a linear combination of the vectors u 1, u,,u k in V if v can be
More informationLinear Algebra: Homework 7
Linear Algebra: Homework 7 Alvin Lin August 6 - December 6 Section 3.5 Exercise x Let S be the collection of vectors in R y that satisfy the given property. In each case, either prove that S forms a subspace
More informationy 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.
6.. Length, Angle, and Orthogonality In this section, we discuss the defintion of length and angle for vectors and define what it means for two vectors to be orthogonal. Then, we see that linear systems
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationMTH 2310, FALL Introduction
MTH 2310, FALL 2011 SECTION 6.2: ORTHOGONAL SETS Homework Problems: 1, 5, 9, 13, 17, 21, 23 1, 27, 29, 35 1. Introduction We have discussed previously the benefits of having a set of vectors that is linearly
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationFirst of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if
5 Matrices in Different Coordinates In this section we discuss finding matrices of linear maps in different coordinates Earlier in the class was the matrix that multiplied by x to give ( x) in standard
More informationMath 308 Practice Test for Final Exam Winter 2015
Math 38 Practice Test for Final Exam Winter 25 No books are allowed during the exam. But you are allowed one sheet ( x 8) of handwritten notes (back and front). You may use a calculator. For TRUE/FALSE
More informationis injective because f(a) = f(b) only when a = b. On the other hand, 1
2.. Injective maps, Kernel, and Linear Independence In this section, we introduce the concept of injectivity for maps. Being injective is a very important property for maps and it appears everywhere in
More informationKernel and range. Definition: A homogeneous linear equation is an equation of the form A v = 0
Kernel and range Definition: The kernel (or null-space) of A is ker A { v V : A v = 0 ( U)}. Theorem 5.3. ker A is a subspace of V. (In particular, it always contains 0 V.) Definition: A is one-to-one
More informationMATH 260 LINEAR ALGEBRA EXAM III Fall 2014
MAH 60 LINEAR ALGEBRA EXAM III Fall 0 Instructions: the use of built-in functions of your calculator such as det( ) or RREF is permitted ) Consider the table and the vectors and matrices given below Fill
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationLinear Algebra Highlights
Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to
More informationGlossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB
Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the
More informationSSEA Math 51 Track Final Exam August 30, Problem Total Points Score
Name: This is the final exam for the Math 5 track at SSEA. Answer as many problems as possible to the best of your ability; do not worry if you are not able to answer all of the problems. Partial credit
More informationMath 3C Lecture 25. John Douglas Moore
Math 3C Lecture 25 John Douglas Moore June 1, 2009 Let V be a vector space. A basis for V is a collection of vectors {v 1,..., v k } such that 1. V = Span{v 1,..., v k }, and 2. {v 1,..., v k } are linearly
More informationThe Gram Schmidt Process
u 2 u The Gram Schmidt Process Now we will present a procedure, based on orthogonal projection, that converts any linearly independent set of vectors into an orthogonal set. Let us begin with the simple
More informationThe Gram Schmidt Process
The Gram Schmidt Process Now we will present a procedure, based on orthogonal projection, that converts any linearly independent set of vectors into an orthogonal set. Let us begin with the simple case
More informationWorksheet for Lecture 15 (due October 23) Section 4.3 Linearly Independent Sets; Bases
Worksheet for Lecture 5 (due October 23) Name: Section 4.3 Linearly Independent Sets; Bases Definition An indexed set {v,..., v n } in a vector space V is linearly dependent if there is a linear relation
More information6.4 BASIS AND DIMENSION (Review) DEF 1 Vectors v 1, v 2,, v k in a vector space V are said to form a basis for V if. (a) v 1,, v k span V and
6.4 BASIS AND DIMENSION (Review) DEF 1 Vectors v 1, v 2,, v k in a vector space V are said to form a basis for V if (a) v 1,, v k span V and (b) v 1,, v k are linearly independent. HMHsueh 1 Natural Basis
More informationChapter 2. General Vector Spaces. 2.1 Real Vector Spaces
Chapter 2 General Vector Spaces Outline : Real vector spaces Subspaces Linear independence Basis and dimension Row Space, Column Space, and Nullspace 2 Real Vector Spaces 2 Example () Let u and v be vectors
More informationMATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION
MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether
More informationMath 21b: Linear Algebra Spring 2018
Math b: Linear Algebra Spring 08 Homework 8: Basis This homework is due on Wednesday, February 4, respectively on Thursday, February 5, 08. Which of the following sets are linear spaces? Check in each
More informationLinear Algebra. Grinshpan
Linear Algebra Grinshpan Saturday class, 2/23/9 This lecture involves topics from Sections 3-34 Subspaces associated to a matrix Given an n m matrix A, we have three subspaces associated to it The column
More informationMath 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections
Math 6 Lecture Notes: Sections 6., 6., 6. and 6. Orthogonal Sets and Projections We will not cover general inner product spaces. We will, however, focus on a particular inner product space the inner product
More informationLecture 14: Orthogonality and general vector spaces. 2 Orthogonal vectors, spaces and matrices
Lecture 14: Orthogonality and general vector spaces 1 Symmetric matrices Recall the definition of transpose A T in Lecture note 9. Definition 1.1. If a square matrix S satisfies then we say S is a symmetric
More informationSpring 2015 Midterm 1 03/04/15 Lecturer: Jesse Gell-Redman
Math 0 Spring 05 Midterm 03/04/5 Lecturer: Jesse Gell-Redman Time Limit: 50 minutes Name (Print): Teaching Assistant This exam contains pages (including this cover page) and 5 problems. Check to see if
More informationExam in TMA4110 Calculus 3, June 2013 Solution
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 8 Exam in TMA4 Calculus 3, June 3 Solution Problem Let T : R 3 R 3 be a linear transformation such that T = 4,
More informationAnswer Key for Exam #2
. Use elimination on an augmented matrix: Answer Key for Exam # 4 4 8 4 4 4 The fourth column has no pivot, so x 4 is a free variable. The corresponding system is x + x 4 =, x =, x x 4 = which we solve
More informationMAT 242 CHAPTER 4: SUBSPACES OF R n
MAT 242 CHAPTER 4: SUBSPACES OF R n JOHN QUIGG 1. Subspaces Recall that R n is the set of n 1 matrices, also called vectors, and satisfies the following properties: x + y = y + x x + (y + z) = (x + y)
More informationWorksheet for Lecture 15 (due October 23) Section 4.3 Linearly Independent Sets; Bases
Worksheet for Lecture 5 (due October 23) Name: Section 4.3 Linearly Independent Sets; Bases Definition An indexed set {v,..., v n } in a vector space V is linearly dependent if there is a linear relation
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationy 1 y 2 . = x 1y 1 + x 2 y x + + x n y n y n 2 7 = 1(2) + 3(7) 5(4) = 4.
. Length, Angle, and Orthogonality In this section, we discuss the defintion of length and angle for vectors. We also define what it means for two vectors to be orthogonal. Then we see that linear systems
More informationHomework 5. (due Wednesday 8 th Nov midnight)
Homework (due Wednesday 8 th Nov midnight) Use this definition for Column Space of a Matrix Column Space of a matrix A is the set ColA of all linear combinations of the columns of A. In other words, if
More informationS09 MTH 371 Linear Algebra NEW PRACTICE QUIZ 4, SOLUTIONS Prof. G.Todorov February 15, 2009 Please, justify your answers.
S09 MTH 37 Linear Algebra NEW PRACTICE QUIZ 4, SOLUTIONS Prof. G.Todorov February, 009 Please, justify your answers. 3 0. Let A = 0 3. 7 Determine whether the column vectors of A are dependent or independent.
More informationMATH 2210Q MIDTERM EXAM I PRACTICE PROBLEMS
MATH Q MIDTERM EXAM I PRACTICE PROBLEMS Date and place: Thursday, November, 8, in-class exam Section : : :5pm at MONT Section : 9: :5pm at MONT 5 Material: Sections,, 7 Lecture 9 8, Quiz, Worksheet 9 8,
More informationMidterm #2 Solutions
Naneh Apkarian Math F Winter Midterm # Solutions Here is a solution key for the second midterm. The solutions presented here are more complete and thorough than your responses needed to be - in order to
More informationRow Space, Column Space, and Nullspace
Row Space, Column Space, and Nullspace MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Introduction Every matrix has associated with it three vector spaces: row space
More information