Solutions: We leave the conversione between relation form and span form for the reader to verify. x 1 + 2x 2 + 3x 3 = 0
|
|
- Emil Ramsey
- 5 years ago
- Views:
Transcription
1 6.2. Orthogonal Complements and Projections In this section we discuss orthogonal complements and orthogonal projections. The orthogonal complement of a subspace S is the set of all vectors orthgonal to S denoted S. We will show that S is in fact a subspace complementary to S. Then the orthogonal projection onto S is the projection onto S with respect to S. Consider any set of vectors E R n. What happens if we look at all of the vectors orthogonal to each vector in E denoted E (pronounced E perp )? Definition For any E R n define E = { x R n x y = for all y E}. When S is a subspace we call S the orthogonal complement of S. Example Express the following as the span of a finite set of vectors. (a) 2 (b) 2 Solutions: We leave the conversione between relation form and span form for the reader to verify. (a) x x x 2 = 2 = x = 2 x x x x = = span. (b) 2 = x x 2 = x = x x x x = x x x = + x = = span. Notice that E is a subspace in these two examples which is not a coincidence. Proposition For any E R n E is a subspace. Proof. We show that E satisfies the three properties of being a subspace. (a) We have E since y = for all y E. Thus E. (b) Suppose x E. Then for all y E ( x + ) y = x y + y = + = since x E. Therefore x + E and E is closed under addition. (c) Suppose x E and c is a scalar. Then for all y E (c x) y = c( x y) = c() = since x E. Therefore c x E and E is closed under scalar multiplication.
2 2 However if E is an infinte set like a subspace it looks like we have to solve an infinite set of equations to find E. How can we possibly find E if E is a nonzero subspace? We need to realize that there is a lot of redudancy in these equations. For example if x y then x c y for all scalars c. In fact we can reduce the set of equations to just being orthogonal to a basis for S. Proposition For any v... v k R n (span( v... v k )) = { v... v k }. Proof. Let S = span( v... v k ). First if x S then x v =... x v k = because v... v k S. Thus x { v... v k } so S { v... v k }. On the other hand if x { v... v k } then x v =... x v k = Now for any y S we can write y = c v + + c k v k so x y = x (c v + + c k v k ) = c ( x v ) + + c k ( x v k ) = c () + + c k () =. Thus x S and { v... v k } S. This shows S = { v... v k }. We will use Proposition shortly to describe how to find the orthogonal complement of a subspace more systematically. Recall that range(a) is the span of the columns of A. It now becomes of use to us to also look at the span of the rows of a matrix. Definition Let row(a) denote the span of the rows of matrix A. For example ([ ]) 2 row = span 2. Let the transpose A of a matrix A denote A with its rows and columns interchanged. For example [ ] 2 = 2 Notice that range(a) = row(a ) and that (A ) = A. Now we can formalize how to find the orthogonal complement of a subspace. Theorem For any matrix A row(a) = ker(a) and range(a) = ker(a ). Proof. Suppose A is an m n matrix with rows v... v m from top to bottom. Then row(a) = span( v... v m ) and () row(a) = (span( v... v m )) = { v... v m } = { x R n x v =... x v m = }. Now notice that x v =... x v m = are exactly the same equations we solve to find ker(a). For example if v = 2 then x x v = 2 = x x = x which is the equation the row [ 2 ] gives when finding ker(a). Thus so row(a) = ker(a) by (). Finally notice that ker(a) = { x R n x v =... x v m = } range(a) = row(a ) = ker(a ).
3 Example Find bases for the following orthogonal complements. (a) span 2 4 (b) span 2 Solutions: (a) span 2 4 (b). 2 4 = row ([ 2 4 ]) ([ ]) = ker 2 4 = span. span 2 = row = ker ([ ]) ([ ]) = ker 2 2 ([ ]) 4 4 = span 2 2. In fact as you may have already guessed from the name S is in fact a subspace complementary to S. As opposed ([ to just([ any complementary ([ subspace the orthogonal complement ([ of S is unique. For example ([ 5 2 span span span are all subspaces complementary of span but only span ]) ]) ]) ]) ]) ([ is the orthogonal complement of span. The reader may verify that the only subspace complementary ]) to S that is also orthogonal to S is the orthogonal complement of S Problem 4. After proving that S S are complementary subspaces we mention other properties that orthogonal complements have. S S Theorem For any subspace S R n S and S are complementary subspaces in R n. Proof. First we show that S S = { }. Suppose x S S. Then since x S and x S x must be orthogonal to itself: x x = = x =
4 4 by Proposition 6..2(iii). Therefore S S = { }. Next we show that S + S = R n using dimension. Let ( v... v k ) be a basis for S so that dim(s) = k. Then S = (span( v... v k )) = { v... v k } = { x R n x v =... x v k = }. So S is the set of solutions to a homogeneous system of k equations in n variables which means there are at most k leading variables and so at least n k free variables. Hence dim(s ) n k. Then dim(s + S ) = dim(s) + dim(s ) dim(s S ) = k + dim(s ) k + (n k) = n. But S + S R n so dim(s + S ) n as well. This forces dim(s + S ) = n which forces S + S = R n so S and S are complementary subspaces of R n by Definition.5.2. Proposition For any subspaces S S S 2 R n (a) dim(s ) = n dim(s). (b) If S S 2 then S2 S (c) ((S) ) = S (d) (S + S 2 ) = S S2 (e) (S S 2 ) = S + S2. Proof. Problem 5. Now that we know S and S are complementary subspaces we can discuss the orthogonal projection π SS onto S with respect to S which we abbreviate π S. Definition Suppose S R n is a subspace. Then is the projection onto S with respect to S. Using Proposition (c) we also have Thus for all x R n (2) π S = π SS π S = π S (S ) = π S S. x = π S ( x) + π S ( x) In order to describe a formula for the orthogonal projection we will need the following Theorem to be proven in Section 6.. Theorem Every subspace of R n has an orthogonal basis. Proof. See Theorem 6... But that doesn t mean you can t try it yourself first. Proposition Suppose S R n is a subspace with orthogonal basis ( u... u k ). Then x u x uk π S ( x) = u 2 u + + u k 2 u k Proof. Let ( v... v n k ) be an orthogonal basis for S which exists by Then ( u... u k v... v n k ) is an orthogonal basis for R n Problem. This means for any x R n x span( u... u k v... v n k ). Hence by Proposition 6..2 x = (( x u u 2 ) u + + ( ) ) x uk u k 2 u k + ( ) x v x vn k v 2 v + + v n k 2 v n k.
5 5 Since we can conclude that x u x uk u 2 u + + u k 2 u k S x v x vn k v 2 v + + v n k 2 v n k S x u x uk π S ( x) = π SS ( x) = u 2 u + + u k 2 u k x v x vn k π S ( x) = π S S( x) = v 2 v + + v n k 2 v n k. Exercises:. Find bases for the following orthogonal complements. 2 (a) span (b) span (c) span (d) span 5 4 (e) span Find formulas for π S ( x) for each of the following subspaces S. (a) S = span 2 (b) S = span 2 (c) S = span (d) S = span 4 5 2
6 6 (e) S = span 2. Problems:. (2) Suppose S R n is a subspace ( u... u k ) is an orthogonal basis for S and ( v... v n k ) is an orthogonal basis for S. Show that ( u... u k v... v n k ) is orthogonal basis for R n. 2. (2) Suppose ( u... u k v... v n k ) is orthogonal basis for R n. Show that span( u... u k ) = span( v... v n k ).. (2+) Fix x R n and a subspace S R n. Show that π S ( x) is the unique vector in R n so that (i) π S ( x) S (ii) x π S ( x) S. 4. (2+) Suppose S R n is a subspace. Show that for all x y R n 5. () Prove Proposition x y = π S ( x) π S ( y) + π S ( x) π S ( y) 6. (a) (2) Show that if A C (A and C differ by EROs) then row(a) = row(c). (b) () Show that the nonzero rows in RREF(A) form a basis for row(a). 7. () Prove or give a counterexample: For any matrix A rank(a) = rank(a ). 8. () Prove or give a counterexample: For any matrix A nullity(a) = nullity(a ). 9. (2) Show that if the rows of m n matrix A are linearly independent then rank(a) = m and nullity(a) = n m.. ) Show that for any n n matrix A det(a) = det(a ).. () Suppose ( u... u n ) is a basis for R n and b... b n are scalars. Show that there exists a unique x R n so that x u = b... x u n = b n. 2. (+) Suppose a... a n R m and b... b n are scalars. Let Show that there exists x R n so that if and only if ker(a) ker(b). A = [ a... a n ] B = [ b... b n ]. x u = b... x u n = b n.. () Suppose ( u... u k ) is orthonormal in R n and let S = span( u... u k ). Then for x R n show that x S if and only if x 2 = ( x u ) ( x u k ) () Suppose S S 2 R n are complementary subspaces so that Show that S 2 = S. 5. Suppose S S 2 R n are subspaces with S S 2. x y = for all x S y S 2.
7 7 (a) () Show that for all y S π S2 ( y) S. (b) () Show that for each x R n there exists unique u S v S 2 S w S2 x = u + v + w. (c) () Show that π S π S2 = π S. so that 6. (4) Fix x R n and a subspace S R n. Show that the minimum of x y 2 over all y S is acheived uniquely when y = π S ( x). 7. (4) Suppose P : R n R n is a projection map which satisfies Show that P is an orthogonal projection. P ( x) x for all x R n.
x 1 + 2x 2 + 3x 3 = 0 x 1 + 2x 2 + 3x 3 = 0, x 2 + x 3 = 0 x 3 3 x 3 1
. Orthogonal Complements and Projections In this section we discuss orthogonal complements and orthogonal projections. The orthogonal complement of a subspace S is the complement that is orthogonal to
More informationSolution: (a) S 1 = span. (b) S 2 = R n, x 1. x 1 + x 2 + x 3 + x 4 = 0. x 4 Solution: S 5 = x 2. x 3. (b) The standard basis vectors
.. Dimension In this section, we introduce the notion of dimension for a subspace. For a finite set, we can measure its size by counting its elements. We are interested in a measure of size on subspaces
More informationThe Fundamental Theorem of Linear Algebra
The Fundamental Theorem of Linear Algebra Nicholas Hoell Contents 1 Prelude: Orthogonal Complements 1 2 The Fundamental Theorem of Linear Algebra 2 2.1 The Diagram........................................
More informationNotice that the set complement of A in U satisfies
Complements and Projection Maps In this section, we explore the notion of subspaces being complements Then, the unique decomposition of vectors in R n into two pieces associated to complements lets us
More informationof A in U satisfies S 1 S 2 = { 0}, S 1 + S 2 = R n. Examples 1: (a.) S 1 = span . 1 (c.) S 1 = span, S , S 2 = span 0 (d.
. Complements and Projection Maps In this section, we explore the notion of subspaces being complements. Then, the unique decomposition of vectors in R n into two pieces associated to complements lets
More informationx 2 For example, Theorem If S 1, S 2 are subspaces of R n, then S 1 S 2 is a subspace of R n. Proof. Problem 3.
.. Intersections and Sums of Subspaces Until now, subspaces have been static objects that do not interact, but that is about to change. In this section, we discuss the operations of intersection and addition
More informationSolutions to Math 51 First Exam April 21, 2011
Solutions to Math 5 First Exam April,. ( points) (a) Give the precise definition of a (linear) subspace V of R n. (4 points) A linear subspace V of R n is a subset V R n which satisfies V. If x, y V then
More informationAnswers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3
Answers in blue. If you have questions or spot an error, let me know. 3 4. Find all matrices that commute with A =. 4 3 a b If we set B = and set AB = BA, we see that 3a + 4b = 3a 4c, 4a + 3b = 3b 4d,
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More informationELE/MCE 503 Linear Algebra Facts Fall 2018
ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2
More informationMath 4A Notes. Written by Victoria Kala Last updated June 11, 2017
Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...
More informationMath 123, Week 5: Linear Independence, Basis, and Matrix Spaces. Section 1: Linear Independence
Math 123, Week 5: Linear Independence, Basis, and Matrix Spaces Section 1: Linear Independence Recall that every row on the left-hand side of the coefficient matrix of a linear system A x = b which could
More informationMATH 15a: Linear Algebra Practice Exam 2
MATH 5a: Linear Algebra Practice Exam 2 Write all answers in your exam booklet. Remember that you must show all work and justify your answers for credit. No calculators are allowed. Good luck!. Compute
More informationLecture 9: Vector Algebra
Lecture 9: Vector Algebra Linear combination of vectors Geometric interpretation Interpreting as Matrix-Vector Multiplication Span of a set of vectors Vector Spaces and Subspaces Linearly Independent/Dependent
More informationReview Notes for Linear Algebra True or False Last Updated: February 22, 2010
Review Notes for Linear Algebra True or False Last Updated: February 22, 2010 Chapter 4 [ Vector Spaces 4.1 If {v 1,v 2,,v n } and {w 1,w 2,,w n } are linearly independent, then {v 1 +w 1,v 2 +w 2,,v n
More informationPRACTICE PROBLEMS FOR THE FINAL
PRACTICE PROBLEMS FOR THE FINAL Here are a slew of practice problems for the final culled from old exams:. Let P be the vector space of polynomials of degree at most. Let B = {, (t ), t + t }. (a) Show
More information2018 Fall 2210Q Section 013 Midterm Exam II Solution
08 Fall 0Q Section 0 Midterm Exam II Solution True or False questions points 0 0 points) ) Let A be an n n matrix. If the equation Ax b has at least one solution for each b R n, then the solution is unique
More informationMath 2174: Practice Midterm 1
Math 74: Practice Midterm Show your work and explain your reasoning as appropriate. No calculators. One page of handwritten notes is allowed for the exam, as well as one blank page of scratch paper.. Consider
More informationLINEAR ALGEBRA REVIEW
LINEAR ALGEBRA REVIEW SPENCER BECKER-KAHN Basic Definitions Domain and Codomain. Let f : X Y be any function. This notation means that X is the domain of f and Y is the codomain of f. This means that for
More informationMarch 27 Math 3260 sec. 56 Spring 2018
March 27 Math 3260 sec. 56 Spring 2018 Section 4.6: Rank Definition: The row space, denoted Row A, of an m n matrix A is the subspace of R n spanned by the rows of A. We now have three vector spaces associated
More informationDefinition Suppose S R n, V R m are subspaces. A map U : S V is linear if
.6. Restriction of Linear Maps In this section, we restrict linear maps to subspaces. We observe that the notion of linearity still makes sense for maps whose domain and codomain are subspaces of R n,
More informationElementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.
Elementary Linear Algebra Review for Exam Exam is Monday, November 6th. The exam will cover sections:.4,..4, 5. 5., 7., the class notes on Markov Models. You must be able to do each of the following. Section.4
More informationOverview. Motivation for the inner product. Question. Definition
Overview Last time we studied the evolution of a discrete linear dynamical system, and today we begin the final topic of the course (loosely speaking) Today we ll recall the definition and properties of
More informationMath Linear Algebra
Math 220 - Linear Algebra (Summer 208) Solutions to Homework #7 Exercise 6..20 (a) TRUE. u v v u = 0 is equivalent to u v = v u. The latter identity is true due to the commutative property of the inner
More informationSolutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015
Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See
More informationFirst of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if
5 Matrices in Different Coordinates In this section we discuss finding matrices of linear maps in different coordinates Earlier in the class was the matrix that multiplied by x to give ( x) in standard
More informationMath 415 Exam I. Name: Student ID: Calculators, books and notes are not allowed!
Math 415 Exam I Calculators, books and notes are not allowed! Name: Student ID: Score: Math 415 Exam I (20pts) 1. Let A be a square matrix satisfying A 2 = 2A. Find the determinant of A. Sol. From A 2
More informationMath 102, Winter 2009, Homework 7
Math 2, Winter 29, Homework 7 () Find the standard matrix of the linear transformation T : R 3 R 3 obtained by reflection through the plane x + z = followed by a rotation about the positive x-axes by 6
More informationWorksheet for Lecture 23 (due December 4) Section 6.1 Inner product, length, and orthogonality
Worksheet for Lecture (due December 4) Name: Section 6 Inner product, length, and orthogonality u Definition Let u = u n product or dot product to be and v = v v n be vectors in R n We define their inner
More informationFinal Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015
Final Review Written by Victoria Kala vtkala@mathucsbedu SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015 Summary This review contains notes on sections 44 47, 51 53, 61, 62, 65 For your final,
More informationMATH 240 Spring, Chapter 1: Linear Equations and Matrices
MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear
More informationLINEAR ALGEBRA: THEORY. Version: August 12,
LINEAR ALGEBRA: THEORY. Version: August 12, 2000 13 2 Basic concepts We will assume that the following concepts are known: Vector, column vector, row vector, transpose. Recall that x is a column vector,
More informationSUMMARY OF MATH 1600
SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You
More information(v, w) = arccos( < v, w >
MA322 F all206 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: Commutativity:
More informationSSEA Math 51 Track Final Exam August 30, Problem Total Points Score
Name: This is the final exam for the Math 5 track at SSEA. Answer as many problems as possible to the best of your ability; do not worry if you are not able to answer all of the problems. Partial credit
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationDefinitions for Quizzes
Definitions for Quizzes Italicized text (or something close to it) will be given to you. Plain text is (an example of) what you should write as a definition. [Bracketed text will not be given, nor does
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationLinear Algebra. Grinshpan
Linear Algebra Grinshpan Saturday class, 2/23/9 This lecture involves topics from Sections 3-34 Subspaces associated to a matrix Given an n m matrix A, we have three subspaces associated to it The column
More informationWorksheet for Lecture 25 Section 6.4 Gram-Schmidt Process
Worksheet for Lecture Name: Section.4 Gram-Schmidt Process Goal For a subspace W = Span{v,..., v n }, we want to find an orthonormal basis of W. Example Let W = Span{x, x } with x = and x =. Give an orthogonal
More informationChapter 2. General Vector Spaces. 2.1 Real Vector Spaces
Chapter 2 General Vector Spaces Outline : Real vector spaces Subspaces Linear independence Basis and dimension Row Space, Column Space, and Nullspace 2 Real Vector Spaces 2 Example () Let u and v be vectors
More informationChapter SSM: Linear Algebra. 5. Find all x such that A x = , so that x 1 = x 2 = 0.
Chapter Find all x such that A x : Chapter, so that x x ker(a) { } Find all x such that A x ; note that all x in R satisfy the equation, so that ker(a) R span( e, e ) 5 Find all x such that A x 5 ; x x
More informationLinear Algebra Highlights
Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to
More informationLecture 3q Bases for Row(A), Col(A), and Null(A) (pages )
Lecture 3q Bases for Row(A), Col(A), and Null(A) (pages 57-6) Recall that the basis for a subspace S is a set of vectors that both spans S and is linearly independent. Moreover, we saw in section 2.3 that
More informationChapter 6. Orthogonality and Least Squares
Chapter 6 Orthogonality and Least Squares Section 6.1 Inner Product, Length, and Orthogonality Orientation Recall: This course is about learning to: Solve the matrix equation Ax = b Solve the matrix equation
More information7. Dimension and Structure.
7. Dimension and Structure 7.1. Basis and Dimension Bases for Subspaces Example 2 The standard unit vectors e 1, e 2,, e n are linearly independent, for if we write (2) in component form, then we obtain
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationChapter 4 Euclid Space
Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,
More informationSection 6.1. Inner Product, Length, and Orthogonality
Section 6. Inner Product, Length, and Orthogonality Orientation Almost solve the equation Ax = b Problem: In the real world, data is imperfect. x v u But due to measurement error, the measured x is not
More informationConsider a subspace V = im(a) of R n, where m. Then,
5.4 LEAST SQUARES AND DATA FIT- TING ANOTHER CHARACTERIZATION OF ORTHOG- ONAL COMPLEMENTS Consider a subspace V = im(a) of R n, where A = [ ] v 1 v 2... v m. Then, V = { x in R n : v x = 0, for all v in
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationChapter 2 Subspaces of R n and Their Dimensions
Chapter 2 Subspaces of R n and Their Dimensions Vector Space R n. R n Definition.. The vector space R n is a set of all n-tuples (called vectors) x x 2 x =., where x, x 2,, x n are real numbers, together
More information2. (10 pts) How many vectors are in the null space of the matrix A = 0 1 1? (i). Zero. (iv). Three. (ii). One. (v).
Exam 3 MAS 3105 Applied Linear Algebra, Spring 2018 (Clearly!) Print Name: Apr 10, 2018 Read all of what follows carefully before starting! 1. This test has 7 problems and is worth 110 points. Please be
More informationMatrix invertibility. Rank-Nullity Theorem: For any n-column matrix A, nullity A +ranka = n
Matrix invertibility Rank-Nullity Theorem: For any n-column matrix A, nullity A +ranka = n Corollary: Let A be an R C matrix. Then A is invertible if and only if R = C and the columns of A are linearly
More informationv = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :
Length, Angle and the Inner Product The length (or norm) of a vector v R 2 (viewed as connecting the origin to a point (v 1,v 2 )) is easily determined by the Pythagorean Theorem and is denoted v : v =
More informationMATH Linear Algebra
MATH 304 - Linear Algebra In the previous note we learned an important algorithm to produce orthogonal sequences of vectors called the Gramm-Schmidt orthogonalization process. Gramm-Schmidt orthogonalization
More informationCheat Sheet for MATH461
Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A
More information(v, w) = arccos( < v, w >
MA322 Sathaye Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v
More informationGlossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB
Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the
More informationTypical Problem: Compute.
Math 2040 Chapter 6 Orhtogonality and Least Squares 6.1 and some of 6.7: Inner Product, Length and Orthogonality. Definition: If x, y R n, then x y = x 1 y 1 +... + x n y n is the dot product of x and
More informationRow Space and Column Space of a Matrix
Row Space and Column Space of a Matrix 1/18 Summary: To a m n matrix A = (a ij ), we can naturally associate subspaces of K n and of K m, called the row space of A and the column space of A, respectively.
More informationy 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.
6.. Length, Angle, and Orthogonality In this section, we discuss the defintion of length and angle for vectors and define what it means for two vectors to be orthogonal. Then, we see that linear systems
More information(v, w) = arccos( < v, w >
MA322 F all203 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v,
More informationb for the linear system x 1 + x 2 + a 2 x 3 = a x 1 + x 3 = 3 x 1 + x 2 + 9x 3 = 3 ] 1 1 a 2 a
Practice Exercises for Exam Exam will be on Monday, September 8, 7. The syllabus for Exam consists of Sections One.I, One.III, Two.I, and Two.II. You should know the main definitions, results and computational
More informationAlgorithms to Compute Bases and the Rank of a Matrix
Algorithms to Compute Bases and the Rank of a Matrix Subspaces associated to a matrix Suppose that A is an m n matrix The row space of A is the subspace of R n spanned by the rows of A The column space
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationis injective because f(a) = f(b) only when a = b. On the other hand, 1
2.. Injective maps, Kernel, and Linear Independence In this section, we introduce the concept of injectivity for maps. Being injective is a very important property for maps and it appears everywhere in
More informationKernel and range. Definition: A homogeneous linear equation is an equation of the form A v = 0
Kernel and range Definition: The kernel (or null-space) of A is ker A { v V : A v = 0 ( U)}. Theorem 5.3. ker A is a subspace of V. (In particular, it always contains 0 V.) Definition: A is one-to-one
More informationMath 308 Practice Test for Final Exam Winter 2015
Math 38 Practice Test for Final Exam Winter 25 No books are allowed during the exam. But you are allowed one sheet ( x 8) of handwritten notes (back and front). You may use a calculator. For TRUE/FALSE
More informationMath 3191 Applied Linear Algebra
Math 191 Applied Linear Algebra Lecture 1: Inner Products, Length, Orthogonality Stephen Billups University of Colorado at Denver Math 191Applied Linear Algebra p.1/ Motivation Not all linear systems have
More informationSolutions to Final Exam
Solutions to Final Exam. Let A be a 3 5 matrix. Let b be a nonzero 5-vector. Assume that the nullity of A is. (a) What is the rank of A? 3 (b) Are the rows of A linearly independent? (c) Are the columns
More informationAFFINE AND PROJECTIVE GEOMETRY, E. Rosado & S.L. Rueda 4. BASES AND DIMENSION
4. BASES AND DIMENSION Definition Let u 1,..., u n be n vectors in V. The vectors u 1,..., u n are linearly independent if the only linear combination of them equal to the zero vector has only zero scalars;
More information4.3 - Linear Combinations and Independence of Vectors
- Linear Combinations and Independence of Vectors De nitions, Theorems, and Examples De nition 1 A vector v in a vector space V is called a linear combination of the vectors u 1, u,,u k in V if v can be
More informationMath 3C Lecture 25. John Douglas Moore
Math 3C Lecture 25 John Douglas Moore June 1, 2009 Let V be a vector space. A basis for V is a collection of vectors {v 1,..., v k } such that 1. V = Span{v 1,..., v k }, and 2. {v 1,..., v k } are linearly
More informationChapter 6 - Orthogonality
Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/
More information6.4 BASIS AND DIMENSION (Review) DEF 1 Vectors v 1, v 2,, v k in a vector space V are said to form a basis for V if. (a) v 1,, v k span V and
6.4 BASIS AND DIMENSION (Review) DEF 1 Vectors v 1, v 2,, v k in a vector space V are said to form a basis for V if (a) v 1,, v k span V and (b) v 1,, v k are linearly independent. HMHsueh 1 Natural Basis
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationChapter 6. Orthogonality
6.4 The Projection Matrix 1 Chapter 6. Orthogonality 6.4 The Projection Matrix Note. In Section 6.1 (Projections), we projected a vector b R n onto a subspace W of R n. We did so by finding a basis for
More information1 Last time: determinants
1 Last time: determinants Let n be a positive integer If A is an n n matrix, then its determinant is the number det A = Π(X, A)( 1) inv(x) X S n where S n is the set of n n permutation matrices Π(X, A)
More informationVector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then
1 Def: V be a set of elements with a binary operation + is defined. F be a field. A multiplication operator between a F and v V is also defined. The V is called a vector space over the field F if: V is
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationThe definition of a vector space (V, +, )
The definition of a vector space (V, +, ) 1. For any u and v in V, u + v is also in V. 2. For any u and v in V, u + v = v + u. 3. For any u, v, w in V, u + ( v + w) = ( u + v) + w. 4. There is an element
More informationMath 21b: Linear Algebra Spring 2018
Math b: Linear Algebra Spring 08 Homework 8: Basis This homework is due on Wednesday, February 4, respectively on Thursday, February 5, 08. Which of the following sets are linear spaces? Check in each
More informationOrthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6
Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and
More informationSemicontinuity of rank and nullity and some consequences
Semicontinuity of rank and nullity and some consequences Andrew D Lewis 2009/01/19 Upper and lower semicontinuity Let us first define what we mean by the two versions of semicontinuity 1 Definition: (Upper
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More informationMath 1180, Notes, 14 1 C. v 1 v n v 2. C A ; w n. A and w = v i w i : v w = i=1
Math 8, 9 Notes, 4 Orthogonality We now start using the dot product a lot. v v = v v n then by Recall that if w w ; w n and w = v w = nx v i w i : Using this denition, we dene the \norm", or length, of
More informationMath 313 Chapter 5 Review
Math 313 Chapter 5 Review Howard Anton, 9th Edition May 2010 Do NOT write on me! Contents 1 5.1 Real Vector Spaces 2 2 5.2 Subspaces 3 3 5.3 Linear Independence 4 4 5.4 Basis and Dimension 5 5 5.5 Row
More informationMATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.
MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:
More informationMath 550 Notes. Chapter 2. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010
Math 550 Notes Chapter 2 Jesse Crawford Department of Mathematics Tarleton State University Fall 2010 (Tarleton State University) Math 550 Chapter 2 Fall 2010 1 / 20 Linear algebra deals with finite dimensional
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More information(a) only (ii) and (iv) (b) only (ii) and (iii) (c) only (i) and (ii) (d) only (iv) (e) only (i) and (iii)
. Which of the following are Vector Spaces? (i) V = { polynomials of the form q(t) = t 3 + at 2 + bt + c : a b c are real numbers} (ii) V = {at { 2 + b : a b are real numbers} } a (iii) V = : a 0 b is
More information18.06 Problem Set 3 - Solutions Due Wednesday, 26 September 2007 at 4 pm in
8.6 Problem Set 3 - s Due Wednesday, 26 September 27 at 4 pm in 2-6. Problem : (=2+2+2+2+2) A vector space is by definition a nonempty set V (whose elements are called vectors) together with rules of addition
More information1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?
. Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in
More informationMAT 242 CHAPTER 4: SUBSPACES OF R n
MAT 242 CHAPTER 4: SUBSPACES OF R n JOHN QUIGG 1. Subspaces Recall that R n is the set of n 1 matrices, also called vectors, and satisfies the following properties: x + y = y + x x + (y + z) = (x + y)
More informationChapter Two Elements of Linear Algebra
Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMath 24 Spring 2012 Questions (mostly) from the Textbook
Math 24 Spring 2012 Questions (mostly) from the Textbook 1. TRUE OR FALSE? (a) The zero vector space has no basis. (F) (b) Every vector space that is generated by a finite set has a basis. (c) Every vector
More informationPractice Exam. 2x 1 + 4x 2 + 2x 3 = 4 x 1 + 2x 2 + 3x 3 = 1 2x 1 + 3x 2 + 4x 3 = 5
Practice Exam. Solve the linear system using an augmented matrix. State whether the solution is unique, there are no solutions or whether there are infinitely many solutions. If the solution is unique,
More information