Chapter 2. Vectors and Vector Spaces

Similar documents
Chapter 2. Vectors and Vector Spaces

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Lecture Notes 1: Vector spaces

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Linear Normed Spaces (cont.) Inner Product Spaces

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

4.4. Orthogonality. Note. This section is awesome! It is very geometric and shows that much of the geometry of R n holds in Hilbert spaces.

The following definition is fundamental.

Spectral Theory, with an Introduction to Operator Means. William L. Green

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

Typical Problem: Compute.

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Definitions and Properties of R N

Another consequence of the Cauchy Schwarz inequality is the continuity of the inner product.

Elementary linear algebra

Lecture 23: 6.1 Inner Products

4.1 Distance and Length

Chapter 3 Transformations

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

Vector Geometry. Chapter 5

Vectors in Function Spaces

Math 3191 Applied Linear Algebra

This pre-publication material is for review purposes only. Any typographical or technical errors will be corrected prior to publication.

INNER PRODUCT SPACE. Definition 1

David Hilbert was old and partly deaf in the nineteen thirties. Yet being a diligent

Introduction to Real Analysis Alternative Chapter 1

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Axioms of Kleene Algebra

Final Exam Practice Problems Math 428, Spring 2017

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math Linear Algebra

Math 290, Midterm II-key

Hilbert spaces. 1. Cauchy-Schwarz-Bunyakowsky inequality

Chapter 6. Orthogonality

Systems of Linear Equations

Chapter 1 Preliminaries

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

Vector Spaces. Commutativity of +: u + v = v + u, u, v, V ; Associativity of +: u + (v + w) = (u + v) + w, u, v, w V ;

Handout #6 INTRODUCTION TO ALGEBRAIC STRUCTURES: Prof. Moseley AN ALGEBRAIC FIELD

Fourier and Wavelet Signal Processing

B553 Lecture 3: Multivariate Calculus and Linear Algebra Review

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India

Projection Theorem 1

Definitions for Quizzes

Extra Problems for Math 2050 Linear Algebra I

Section 3.2. Multiplication of Matrices and Multiplication of Vectors and Matrices

Abstract Vector Spaces

Chapter 2 Linear Transformations

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

CHAPTER VIII HILBERT SPACES

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH Linear Algebra

1 Inner Product Space

Functional Analysis MATH and MATH M6202

0 Sets and Induction. Sets

Section II.1. Free Abelian Groups

Linear Algebra Massoud Malek

Exercise Sheet 1.

There are two things that are particularly nice about the first basis

MIDTERM I LINEAR ALGEBRA. Friday February 16, Name PRACTICE EXAM SOLUTIONS

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

PROBLEMS. (b) (Polarization Identity) Show that in any inner product space

Section 3.9. Matrix Norm

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Lecture 20: 6.1 Inner Products

ECON 4117/5111 Mathematical Economics

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Inner product spaces. Layers of structure:

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

SUMMARY OF MATH 1600

Section 5.6. LU and LDU Factorizations

Introduction and Preliminaries

Vector Calculus. Lecture Notes

An introduction to some aspects of functional analysis

which has a check digit of 9. This is consistent with the first nine digits of the ISBN, since

(, ) : R n R n R. 1. It is bilinear, meaning it s linear in each argument: that is

Vector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then

1.2 LECTURE 2. Scalar Product

Math 117: Topology of the Real Numbers

y 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.

which are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar.

Abstract & Applied Linear Algebra (Chapters 1-2) James A. Bernhard University of Puget Sound

1. General Vector Spaces

Chapter 3. Basic Properties of Matrices

Review of Linear Algebra

1. Continuous Functions between Euclidean spaces

Analysis and Linear Algebra. Lectures 1-3 on the mathematical tools that will be used in C103

Mathematics Department Stanford University Math 61CM/DM Inner products

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation

LINEAR ALGEBRA REVIEW

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

Linear Algebra. Preliminary Lecture Notes

Linear Analysis Lecture 5

CHAPTER II HILBERT SPACES

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Supplement to Section 22. Topological Groups

l(y j ) = 0 for all y j (1)

Exercises for Unit I (Topics from linear algebra)

Applied Linear Algebra in Geoscience Using MATLAB

Transcription:

2.1. Operations on Vectors 1 Chapter 2. Vectors and Vector Spaces Section 2.1. Operations on Vectors Note. In this section, we define several arithmetic operations on vectors (especially, vector addition and scalar multiplication). We reintroduce much of the terminology associated with vectors from sophomore level Linear Algebra (MATH 2010). Definition. A (real) vector space is a set V of vectors along with an operation of addition + of vectors and multiplication of a vector by a scalar (real number), which satisfies the following. For all vectors x, y, z V and for all scalars a, b R: (A1) (x + y) + z = x + (y + z) (Associativity of Vector Addition) (A2) x + y = y + x (Commutivity of Vector Addition) (A3) There exists 0 V such that 0 + x = x (Additive Identity) (A4) x + ( x) = 0 (Additive Inverses) (S1) a(x+y) = ax+ay (Distribution of Scalar Multiplication over Vector Addition) (S2) (a+b)x = ax+bx (Distribution of Scalar Addition over Scalar Multiplication) (S3) a(bx) = (ab)x (Associativity) (S4) 1x = x (Preservation of Scale). Note. More generally, we can consider a vector space over a field. For details, see my online notes on Vector Spaces from Introduction to Modern Algebra 2: http://faculty.etsu.edu/gardnerr/4127/notes/vi-30.pdf. However, unless otherwise noted, we only consider finite dimensional vector spaces with real scalars.

2.1. Operations on Vectors 2 Definition. Two vectors in R n, x = [x 1,x 2,...,x n ] and y = [y 1,y 2,...,y n ] are equal if x i = y i for i = 1,2,...,n. The zero vector in R n is 0 n = 0 = [0,0,...,0]. The one vector in R n (also called the summing vector ) is 1 n = 1 = [1,1,...,1]. (Later when using 1 n to generate sums, we will treat it as a column vector.) Definition. Let x = [x 1,x 2,...,x n ] and y = [y 1,y 2,...,y n ] be vectors in R n. Let a R be a scalar. Define (1) x + y = [x 1 + y 1,x 2 + y 2,...,x n + y n ], and (2) ax = [ax 1,ax 2,...,ax n ]. Theorem 2.1.1. Properties of Vector Algebra in R n. Let x,y,z R n and let a,b R be scalars. Then: A1. (x + y) + z = x + (y + z) (Associativity of Vector Addition) A2. x + y = y + x (Commutivity of Vector Addition) A3. 0 + x = x (Additive Identity) A4. x + ( x) = 0 (Additive Inverses) S1. a(x+y) = ax+ay (Distribution of Scalar Multiplication over Vector Addition) S2. (a+b)x = ax+bx (Distribution of Scalar Addition over Scalar Multiplication) S3. a(bx) = (ab)x (Associativity) S4. 1x = x (Preservation of Scale)

2.1. Operations on Vectors 3 Note. The statement of Theorem 2.1.1 is as it is stated in Fraleigh and Beauregard s Linear Algebra, 3rd Edition, Addison-Wesley Publishing Company (1995). The proof is very elementary and is based on the definition of vector addition, scalar multiplication, and the field properties of R. Definition. With vectors v 1,v 2,...,v k R n and scalars a 1,a 2,...,a k R, we have the linear combination a 1 v 1 +a 2 v 2 + +a k v k R n. The set of vectors in R n, {v 1,v 2,...,v k } is linearly independent if the equation a 1 v 1 + a 2 v 2 + + a k v k = 0 implies that a 1 = a 2 = = a k = 0. A set of vectors in R n is linearly dependent if it is not linearly independent. Note. If the set of vectors in R n, {v 1,v 2,...,v k }, is linearly dependent then there are some scalars a 1,a 2,...,a k R not all zero such that a 1 v 1 +a 2 v 2 + +a k v k = 0. Note. In Exercise 2.1, the following is established: Exercise 2.1. The maximum number of vectors in R n that forms a linearly independent set is n. Definition. A set of n-vectors, V R n, is a vector space if it is closed under linear combinations. That is, if for any x,y V and any scalars a,b R we have ax + by V.

2.1. Operations on Vectors 4 Note. The above definition may seem unusually simple to you. Technically, we are basing all the properties in the traditional definition of vector space on Theorem 2.1.1 and the previous definition just justifies the fact that V is a subspace of R n. Note. In Exercise 2.1.B you are asked to prove: For vector space V (as defined above), if W 1 and W 2 are finite subsets of V which are linearly independent and of maximal size then W 1 and W 2 are the same size (or cardinality ); that is, W 1 = W 2. Notice that by Exercise 2.1, for linearly independent sets W 1 and W 2 we have W 1 n and W 2 n. This justifies the following definition. Definition. The maximum number of linearly independent n-vectors in a vector space V is the dimension of the vector space, denoted dim(v ). When V consists of n-vectors, n is the order of vector space V. Definition. Two vector spaces V 1 and V 2 both of order n are essentially disjoint if V 1 V 2 = {0}. Definition. The ith unit vector in R n, denoted e i, has a 1 in the ith position and 0 s in all other positions: e i = [0,0,...,0,1,0,...,0]. For x = [x 1,x 2,...,x n ] R n, the sign vector is defined as sign(x) = [sign(x 1 ),sign(x 2 ),...,sign(x n )] R n where 1 if x i > 0 sign(x i ) = 0 if x i = 0 1 if x i < 0.

2.1. Operations on Vectors 5 Note. An ordering on a set is a reflexive, antisymmetric, transitive binary operation. A total ordering on set A is an ordering on A such that for any a,b A either a b or b a. An ordering on set A which is not a total ordering is a partial ordering. Applications of partial orderings often involve the partial ordering of subset inclusion. For example, is a partial ordering on A = {{a}, {b}, {a, b}} since {a} {a,b}, {b} {a,b}, but neither {a} {b} nor {b} {a} holds. Definition. We define a partial ordering on R n as follows. We say x is greater than y denoted x > y if x i > y i for i = 1,2,...,n. Similarly, we say x is greater than or equal to y, denoted x y, if x i y i for i = 1,2,...,n. Definition. A subset of a vector space V that is itself a vector space is a subspace of V. Theorem 2.1.2. Let V 1 and V 2 be vector spaces of n-vectors. Then V 1 V 2 is a vector space. Note. A union of two vector spaces of n-vectors need not be a vector space, as you will show by example in Exercise 2.2. Note. The text refers to a set of vectors of the same order as a space of vectors. A better term would be a set of n-vectors.

2.1. Operations on Vectors 6 Definition. If V 1 and V 2 are sets of n-vectors, then the sum of V 1 and V 2 is V 1 + V 2 = {v 1 + v 2 v 1 V 1,v 2 V 2 }. Theorem 2.1.3. If V 1 and V 2 are vector spaces of n-vectors, then V 1 + V 2 is a vector space. Definition. If V 1 and V 2 are essentially disjoint vector spaces of n-vectors, then the vector space V 1 + V 2 is the direct sum of V 1 and V 2, denoted V 1 V 2. Theorem 2.1.4. If vector spaces V 1 and V 2 are essentially disjoint then every element of V 1 V 2 can be written as v 1 +v 2, where v 1 V 1 and v 2 V 2, in a unique way. Definition. A set of n-vectors that contains all positive scalar multiples of any vector in the set and contains the zero vector is a cone. A set of vectors C is a convex cone if for all v 1,v 2 C and all a,b 0, we have av 1 + bv 2 C. Example. In R 2, C 1 = {(x, 0),(0,y) x > 0,y > 0} {(0,0)} is a cone but not a convex cone. The set C 2 = {v R 2 v > 0} {(0,0)} (which includes all vectors in R 2 which when in standard position with their tail at the origin, have their head in the first quadrant of the Cartesian plane) is a convex cone. Similarly, a cone (not double-cone) with its vertex at the origin in R n is a cone in the sense defined here; this follows from the parallelogram law of addition of vectors.

2.1. Operations on Vectors 7 Definition. Let G be a subset of vector space V of n-vectors. If each vector in V is some linear combination of elements of G, then G is a spanning set (or generating set) of V. The set of all linear combinations of elements of G is the span of G, denoted span(g). If a spanning set G of V is linearly independent, then it is a basis of V. Theorem 2.1.5. If {v 1,v 2,...,v k } is a basis for a vector space V, then each element can be uniquely represented as a linear combination of the basis vectors. Note. If B 1 is a basis for V 1 and B 2 is a basis for V 2 where V 1 and V 2 are essentially disjoint vector space of n-vectors, then B 1 B 2 is a basis for V 1 V 2. This follows from Theorems 2.1.4 and 2.1.5. Definition. A set of vectors S = {v i i = 1,2,...,k} such that for any vector v in a cone C there exists scalars a i 0 for i = 1,2,...,k so that v = k i=1 a iv i and if for scalars b i 0 for i = 1,2,...,k and n i=1 b iv i = 0 then b i = 0 for i = 1,2,...,k is a spanning set (or generating set) for cone C. If a spanning set of a cone has a finite number of elements, the cone is a polyhedron. A spanning set of a cone consisting of a minimum number of vectors of any spanning set for that cone is a basis set for the cone. Example. In R 2, the cones C 1 and C 2 both have as a basis {[1,0],[0,1]}. Therefore, both are polyhedra.

2.1. Operations on Vectors 8 Definition. For two n-vectors x,y R n, define the inner product of x = [x 1,x 2,...,x n ] and y = [y 1,y 2,...,y n ] (also called dot product or scalar product) as x,y = n i=1 x iy i. A vector space with an inner product is an inner product space. Theorem 2.1.6. Properties of Inner Products. Let x,y,z R n and let a R. Then: 1. If x 0 then x,x > 0 and 0,x = x, 0 = 0,0 = 0 (Nonnegativity and Mapping of the Identity), 2. x, y = y, x (Commutivity of Inner Products), 3. a x, y = ax, y (Factoring of Scalar Multiplication in Inner Products), 4. x + y,z = x,z + y,z (Relation of Vector Addition to Addition of Inner Products). Note. We can combine the claim of Theorem 2.1.6 to show that the Inner Product is linear in the first and second entries: ax + by,z = a x,z + b y,z and x,ay + bz = a x,y + b x,z. Note. An important property of inner products is the Schwarz Inequality (or Cauchy-Schwarz Inequality ). It is used in the proof of the Triangle Inequality once we address norms. Theorem 2.1.7. Schwarz Inequality. For any x,y R n we have x,y x,x 1/2 y,y 1/2.

2.1. Operations on Vectors 9 Definition. A norm,, on a set S R n is a function from S to R satisfying: 1. If x 0 then x > 0, and 0 = 0 (Nonnegativity and Mapping of the Identity), 2. ax = a x for all x S and a R (Relation of Scalar Multiplication to Real Multiplication), 3. x + y x + y for all x,y S (Triangle Inequality). A vector space together with a norm is a normed linear space (or simply normed space). Example. The Euclidean norm on R n is the familiar norm x = [x 1,x 2,...,x n ] = { n k=1 (x k) 2} 1/2 = x,x 1/2. If x 0 then some x i 0 and so x x i > 0, and 0 = 0. So Nonnegativity and Mapping of the Identity holds. For x R n and a R, = { n } 1/2 ax = [ax 1,ax 2,...,ax n ] = (ax k ) 2 { a 2 k=1 } 1/2 { n n } 1/2 (x k ) 2 = a (x k ) 2 = a x, k=1 so Relation of Scalar Multiplication to Real Multiplication holds. For any x, y R n, k=1 x + y 2 = x + y,x + y = x,x + 2 x,y + y,y x,x + 2 x,x 1/2 y,y 1/2 + y,y by the Schwarz Inequality = x 2 + 2 x y + y 2 = ( x + y ) 2, and so x + y x + y and the Triangle Inequality holds. Therefore, the Euclidean norm is indeed a norm or R n.

2.1. Operations on Vectors 10 Note. Another norm on R n is the l p -norm, p, for p 1 defined as ( n ) 1/p x p = [x 1,x 2,...,x n ] p = x k p. We omit the details of the proof that the l p -norm is indeed a norm. Notice that the l 2 -norm is the same as the Euclidean norm. You are likely to encounter the l p - norm in the future in the setting of l p -spaces (as opposed to the space R n ) where the l p -spaces are examples of Banach spaces. The l -norm on R n is defined as x = max{ x k k = 1,2,...,n}. Gentle gives a (not very rigorous) argument on page 18 that lim p x p = x. Statisticians sometimes call the l -norm the Chebyshev norm. k=1 Definition. Let {v 1,v 2,...,v k } be a basis for vector space V. For any x = c 1 v 1 + c 2 v 2 + + c k v k V we define the basis norm ρ(x) = { k j=1 c2 j} 1/2. Theorem 2.1.8. The basis norm is indeed a norm for any basis {v 1,v 2,...,v k } of vector space V. Definition. Let a and b be norms on a vector space V. Then a is equivalent to b, denoted a = b, if there are r > 0 and s > 0 such that for all x V we have r x b x a s x b. Note. The equivalence of a and b does not reflect the fact that the norms give the same value for any particular vector. Equivalence is ultimately related to the topology on the vector space and the behavior of convergence sequences.

2.1. Operations on Vectors 11 Theorem 2.1.9. Equivalence of norms is an equivalence relation. That is, it is reflexive, symmetric, and transitive. Proof. The proof is left as Exercise 2.1.D. Theorem 2.1.10. Every norm on (finite dimensional vector space) V is equivalent to the basis norm ρ for some given basis {v 1,v 2,...,v k }. Therefore, any two norms on V are equivalent. Note. We are only considering finite dimensional spaces, so Theorem 2.1.10 applies to all vector spaces we consider. The result does not hold in infinite dimensions and a given infinite dimensional vector space can have two unequal norms. Definition. A sequence of vectors x 1,x 2,... in a normed vector space V (with norm ) converges to vector x V if for all ε > 0 there is M N such that if n M then x x n < ε. Definition. For nonzero vector x in normed vector space V (with norm ), the vector x = x/ x is the normalized vector associated with x.

2.1. Operations on Vectors 12 Definition. For S any set of elements, : S R is a metric if for all x,y,z S we have: 1. (x,y) > 0 if x y and (x,y) = 0 if x = y, 2. (x,y) = (y,x), 3. (x,y) (x,z) + (z,y). Note. We can think of a norm on a vector space as a way to associate a length with a vector. A metric on a vector space gives a way to associate a distance between two vectors. In a vector space, norms and metrics are intimately related. In fact, if is a norm on vector space V then defining (x,y) = x y for all x,y V implies that is a metric on V. Conversely, if is a metric on vector space V, then defining x = (0,x) is a norm on V. These claims are to be justified in a modified version of Exercise 2.7. Definition. Two vectors v 1,v 2 V, where V is a vector space with an inner product, are orthogonal if v 1,v 2 = 0. A set of vectors S = {v 1,v 2,...,v k } is an 0 if i j orthonormal set if for all v i,v j S we have v i,v j = 1 if i = j. Theorem 2.1.11. A set of nonzero vectors {v 1,v 2,...,v k } in a vector space with an inner product for which v i,v j = 0 for i j (the vectors are said to be mutually orthogonal) is a linearly independent set.

2.1. Operations on Vectors 13 Definition. Two spaces of n-vectors, V 1 and V 2, are orthogonal, denoted V 1 V 2, if v 1,v 2 = 0 for all v 1 V 1 and v 2 V 2. If V 1 V 2 and V 1 V 2 = R n, then V 2 is the orthogonal complement of V 1, denoted V 2 = V 1 (sometimes called the perp space of V 1 ). Note. By Exercise 2.9, if V 1 V 2 then V 1 V 2 = {0}. That is, if V 1 V 2 then V 1 and V 2 are essentially disjoint and so V 1 V 2 is defined. Note. If B 1 is a basis for V 1 and B 2 is a basis for V 2 where V 1 V 2 (and V 1,V 2 are finite dimensional), then B 1 B 2 is a basis for V 1 V 2, by Exercise 2.1.F. Since B 1 B 2 = (B 1 V 1, B 2 V 2, and V 1 V 2 = {0}, but 0 B 1 and 0 B 2 ). So the dimension of V 1 V 2 is B 1 B 2 = B 1 + B 2. That is, dim(v 1 V 2 ) = dim(v 1 ) + dim(v 2 ). Note. We introduced the vector 1 n R n as 1 n = [1,1,...,1] T above and referred to it as the summing vector. For any x R n we have, by treating 1 T n ( 1 n transpose ) as a 1 n matrix and x as a n 1 matrix, that as a matrix product 1 T nx = n i=1 x i. (Technically, the right hand side is a 1 1 matrix.) Definition. For x R n, the arithmetic mean (or simply mean) x is defined as x = 1 n 1T nx. The mean vector associated with x R n is x = [x,x,...,x] R n. Note. We have [x,x,...,x] 2 = nx 2. Gentle also uses the symbol x for both the mean of the entries in x and the mean vector of x. We follow his lead, but try to verbally express whether x represents a scalar or vector. Revised: 6/14/2018