Lecture 9. Econ August 20

Similar documents
Econ Slides from Lecture 8

Definition 3 A Hamel basis (often just called a basis) of a vector space X is a linearly independent set of vectors in X that spans X.

Econ Lecture 8. Outline. 1. Bases 2. Linear Transformations 3. Isomorphisms

Vector Spaces and Linear Transformations

(here, we allow β j = 0). Generally, vectors are represented as column vectors, not row vectors. β 1. β n. crd V (x) = R n

NORMS ON SPACE OF MATRICES

Lecture 10. Econ August 21

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

Lecture 2. Econ August 11

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Elementary linear algebra

A = 1 6 (y 1 8y 2 5y 3 ) Therefore, a general solution to this system is given by

Linear Algebra Lecture Notes-I

Econ Slides from Lecture 7

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

Scientific Computing: Dense Linear Systems

This last statement about dimension is only one part of a more fundamental fact.

Dot Products, Transposes, and Orthogonal Projections

Announcements Wednesday, November 01

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Course Summary Math 211

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

Final EXAM Preparation Sheet

ECON2285: Mathematical Economics

Lecture 2: Linear Algebra

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

Convex Sets. Prof. Dan A. Simovici UMB

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

Solutions to Final Exam

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Lecture Summaries for Linear Algebra M51A

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Announcements Monday, October 29

Lecture 7. Econ August 18

Elements of Convex Optimization Theory

Linear Algebra Highlights

A course in low-dimensional geometry

Lecture 3. Econ August 12

The converse is clear, since

MATH 33A LECTURE 2 SOLUTIONS 1ST MIDTERM

Linear Algebra Lecture Notes-II

Math 21b. Review for Final Exam

Announcements Wednesday, November 01

Lecture 7: Positive Semidefinite Matrices

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)

MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix.

Linear Algebra. Preliminary Lecture Notes

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

Vector Spaces. (1) Every vector space V has a zero vector 0 V

Math 344 Lecture # Linear Systems

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Math113: Linear Algebra. Beifang Chen

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

Chapter 2 Linear Transformations

Definition of adjoint

Answers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3

x y + z = 3 2y z = 1 4x + y = 0

Math 113 Midterm Exam Solutions

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

REPRESENTATION THEORY WEEK 7

4 Vector Spaces. 4.1 Basic Definition and Examples. Lecture 10

Linear Algebra. Preliminary Lecture Notes

Structural Properties of Utility Functions Walrasian Demand

Math 341: Convex Geometry. Xi Chen

7. Dimension and Structure.

Online Exercises for Linear Algebra XM511

Definitions for Quizzes

235 Final exam review questions

2. Every linear system with the same number of equations as unknowns has a unique solution.

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Math 323 Exam 2 Sample Problems Solution Guide October 31, 2013

Math 291-2: Lecture Notes Northwestern University, Winter 2016

MATH 304 Linear Algebra Lecture 15: Linear transformations (continued). Range and kernel. Matrix transformations.

Review of Some Concepts from Linear Algebra: Part 2

Linear algebra and differential equations (Math 54): Lecture 10

2: LINEAR TRANSFORMATIONS AND MATRICES

Methods of Mathematical Physics X1 Homework 2 - Solutions

Vector Space Concepts

Sometimes the domains X and Z will be the same, so this might be written:

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

ECON 5111 Mathematical Economics

Math 309 Notes and Homework for Days 4-6

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

First of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if

The definition of a vector space (V, +, )

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

AFFINE AND PROJECTIVE GEOMETRY, E. Rosado & S.L. Rueda 4. BASES AND DIMENSION

Math Linear Algebra II. 1. Inner Products and Norms

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Linear Vector Spaces

Introduction to Matrix Algebra

ECON0702: Mathematical Methods in Economics

MATH 235. Final ANSWERS May 5, 2015

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Linear Algebra. Session 12

MTH 464: Computational Linear Algebra

1. Let r, s, t, v be the homogeneous relations defined on the set M = {2, 3, 4, 5, 6} by

Math 54 HW 4 solutions

Chapter -1: Di erential Calculus and Regular Surfaces

Transcription:

Lecture 9 Econ 2001 2015 August 20

Lecture 9 Outline 1 Linear Functions 2 Linear Representation of Matrices 3 Analytic Geometry in R n : Lines and Hyperplanes 4 Separating Hyperplane Theorems Back to vector spaces so that we can illustrate the formal connection between linear functions and matrices. This matters for a lot calculus since derivatives are linear functions. Then some useful geometry in R n. Announcements: - Test 2 will be tomorrow at 3pm, in WWPH 4716, and there will be recitation at 1pm. The exam will last an hour.

Linear Transformations Definition Let X and Y be two vector spaces. We say T : X Y is a linear transformation if T (α 1 x 1 + α 2 x 2 ) = α 1 T (x 1 ) + α 2 T (x 2 ) x 1, x 2 X, α 1, α 2 R Let L(X, Y ) denote the set of all linear transformations from X to Y. L(X, Y ) is a vector space. Example The function T : R n R m defined by T (x) = Ax, with A is linear: m n T (ax + by) = A(ax + by) = aax + bay = at (x) + bt (y)

Linear Functions An equivalent characterization of linearity l : R n R m is linear if and only if 1 for all x and y in R n, l(x + y) = l(x) + l(y) and 2 for all scalars λ, l(λx) = λl(x) Linear functions are additive and exhibit constant returns to scale (econ jargon). If l is linear, then l(0) = 0 and, more generally, l(x) = l( x).

Compositions of Linear Transformations The compositions of two linear functions is a linear function Given R L(X, Y ) and S L(Y, Z), define S R : X Z. (S R)(αx 1 + βx 2 ) = S(R(αx 1 + βx 2 )) = S(αR(x 1 ) + βr(x 2 )) = αs(r(x 1 )) + βs(r(x 2 )) = α(s R)(x 1 ) + β(s R)(x 2 ) so S R L(X, Z). Think about statement and proof for the equivalent characterization of linearity.

Kernel and Rank Definition Let T L(X, Y ). The image of T is Im T = T (X ) = {y Y : y = T (x) for some x X } The kernel of T is ker T = {x X : T (x) = 0} The rank of T is Rank T = dim(im T ) dim X, is the cardinality of any basis of X. The kernel is the set of solutions to T (x) = 0 The image is the set of vectors in Y for which T (x) = y has at least one solution. Exercise: prove that T L(X, Y ) is one-to-one if and only if ker T = {0}. Theorem (The Rank-Nullity Theorem) Let X be a finite-dimensional vector space, T L(X, Y ). Then Im T and ker T are vector subspaces of Y and X respectively, and dim X = dim ker T + Rank T

Invertible Linear Functions Definition T L(X, Y ) is invertible if there exists a function S : Y X such that S (T (x)) = x x X T (S (y)) = y y Y Denote S by T 1. In other words, T is invertible if and only if it is one-to-one and onto. This is the usual condition for the existence of an inverse function. The linearity of the inverse function follows from the linearity of T. Theorem If T L(X, Y ) is invertible, then T 1 L(Y, X ); that is, T 1 is linear. Proof. Problem Set 9.

From Linear Functions to Matrices l : R n R m is linear if and only if (i) l(x + y) = l(x) + l(y) and l(λx) = λl(x). Any linear function is described by a matrix. Compute l(e i ), where e i is the ith component of the standard basis in R n. For each e i, l(e i ) is a vector in R m, call it a i. Form A as the square matrix with ith column equal to a i. By construction, A has n columns and m rows and therefore f (x) = Ax for all x. Next, apply this procedure to any linear basis of a vector space. Using this observation, we derive a conection between composition of linear functions and multiplication of matrices.

Linear Transformations and Bases If X and Y are vector spaces, to define a linear transformation T from X to Y, it is suffi cient to define T for every element of a basis for X. Why? Theorem Let V be a basis for X. Then every vector x X has a unique representation as a linear combination of a finite number of elements of V. Let X and Y be two vector spaces, and let V = {v λ : λ Λ} be a basis for X. Then a linear function T L(X, Y ) is completely determined by its values on V, that is: 1 Given any set {y λ : λ Λ} Y, T L(X, Y ) such that T (v λ ) = y λ λ Λ 2 If S, T L(X, Y ) and S(v λ ) = T (v λ ) for all λ Λ, then S = T.

Linear Transformations and Bases Proof. Proof of 1. If x X, x has a unique representation of the form n x = α i v λi α i 0 i = 1,..., n i=1 Define T (x) = n i=1 α i y λi Then T (x) Y and you will show that T is linear as an exercise. Proof of 2. Suppose S(v λ ) = T (v λ ) for all λ Λ. We need to show that S = T. Given x X, ( n ) S(x) = S α i v λi = so S = T. = T (x) i=1 n α i S (v λi ) = i=1 ( n n ) α i T (v λi ) = T α i v λi i=1 i=1

Isomorphisms Definitions Two vector spaces X and Y are isomorphic if there is an invertible T L(X, Y ) T is then called an isomorphism. Isomorphic vector spaces are essentially indistinguishable If T is an isomorphism it is one-to-one and onto: x 1, x 2 X, there exist unique y 1, y 2 Y s.t. y 1 = T (x 1 ) and y 2 = T (x 2 ) y 1, y 2 Y, there exist unique x 1, x 2 X s.t. x 1 = T 1 (y 1 ) and x 2 = T 1 (y 2 ) moreover, by linearity T (αx 1 + βx 2 ) = αy 1 + βy 2 and T 1 (αy 1 + βy 2 ) = αx 1 + βx 2 Theorem Two vector spaces X and Y are isomorphic if and only if dim X = dim Y.

Coordinate Representation of Vectors By the previous theorem, any vector space of dimension n is isomorphic to R n. What s the isomorphism? Let X be a finite-dimensional vector space over R with dim X = n. Fix a basis V = {v 1,..., v n } of X, then any x X has a unique representation x = n j=1 β j v j (so the β j are unique real numbers). Define the coordinate representation of x with respect to the basis V as By construction crd V (v 1 ) = 1 0. 0 0 crd V (x) = crd V (v 2 ) = β 1. β n 0 1. 0 0 R n crd V is an isomorphism from X to R n (verify this).... crd V (v n ) = 0 0. 0 1

Matrix Representation Suppose T L(X, Y ), where X and Y have dimension n and m respecively. Fix bases V = {v 1,..., v n } of X and W = {w 1,..., w m } of Y. Since T (v j ) Y, we can write Define T (v j ) = m α ij w i i=1 α 11 α 1n Mtx W,V (T ) =..... α m1 α mn Each column is given by crd W (T (v j )), the coordinates of T (v j ) with respect to W. Thus any linear transformation from X to Y is equivalent to a matrix once one fixes the two bases.

Matrix Representation Example X = Y = R 2, V = {(1, 0), (0, 1)}, W = {(1, 1), ( 1, 1)}; Let T be the identity: T (x) = x for each x. ( ) 1 0 Notice that Mtx W,V (T ). Mtx 0 1 W,V (T ) is the matrix that changes basis from V to W. v 1 = (1, 0) = α 11 (1, 1) + α 21 ( 1, 1) α 11 α 21 = 1 and α 11 + α 21 = 0 2α How do we compute it? 11 = 1, α 11 = 1 2 hence α 21 = 1 2 v 2 = (0, 1) = α 12 (1, 1) + α 22 ( 1, 1) α 12 α 22 = 0 and α 12 + α 22 = 1 2α 12 = 1, α 12 = 1 2 hence α 22 = 1 2 ( ) 1/2 1/2 So Mtx W,V (id) = 1/2 1/2

Matrices and Vectors Given T L(X, Y ), where dim X = n and dim Y = m, and V = {v 1,..., v n } and W = {w 1,..., w m } are bases of X and Y respectively. α 11 α 1n We built: Mtx W,V (T ) =..... α m1 α mn Clearly: α 11 α 1n..... α m1 α mn 1 0. 0 and crd V (v 1 ) = = α 11. α m1 Thus Mtx W,V (T ) crd V (v j ) = crd W (T (v j )) and more generally Mtx W,V (T ) crd V (x) = crd W (T (x)) x X REMARK: Multiplying a vector by a matrix does two things 1 Computes the action of the linear function T 2 Accounts for the change in basis. 1 0. 0.

Matrices and Linear Transofrmations are Isomorphic Theorem Let X and Y be vector spaces over R, with dim X = n, dim Y = m. Then L(X, Y ), the space of linear transformations from X to Y, is isomorphic to M m n, the vector space of m n matrices over R. If V = {v 1,..., v n } is a basis for X and W = {w 1,..., w m } is a basis for Y, then Mtx W,V L(L(X, Y ), M m n ) and Mtx W,V is an isomorphism from L(X, Y ) to M m n. This is mostly the consequence of things we already know about dimensionality once you realize the relationship between matrices and linear transformations. Note that X and Y are general vector spaces, but they are isomorphic to matrices of real numbers.

M m=1 Summary and Illustration Let T L(V, W ), where V and W are vectors spaces over R with dim V = N and dim W = M, and fix bases {v 1,..., v N } of V and {w 1,..., w M } of W. Any v V can be uniquely represented as v = N n=1 x nv n (the x n s are unique). Any T (v) W is uniquely represented as T (v) = M m=1 y mw m (the y m s are unique). Since T (v n ) W for each n, T (v n ) = M m=1 a nmw m (where the a nm s are unique). Thus: N N N M M N y m w m = T (v) = T ( x n v n ) = x n T (v n ) = x n a nm w m = ( x n a mn )w m Therefore, n=1 n=1 y m = n=1 m=1 N x n a nm for all m n=1 m=1 n=1 Let y = (y 1,.., y M ) R M, x = (x 1,.., x N ) R N, and A = [a mn ] (the m n matrix with m, n entry given by the real number a mn ); then, we have y = Ax In words: the matrix of real numbers A represents the linear transformation T given the bases {v 1,..., v N } and {w 1,..., w M }. there is one and only one linear transformation T corresponding to A. there is one and only one matrix A corresponding to T.

Matrix Product as Composite Transformation Suppose we have another linear transformation S L(W, Q) where Q is a vector space with basis {q 1,..., q J }. Let B = [b jm ] be the J M matrix representing S with respect to the bases {w 1,..., w M } and {q 1,..., q J } for W and Q, respectively. Then, for all m, S(w m ) = J j=1 b jmq j. S T : V Q is the linear transformation S T (v) = S(T (v)). Using the same logic of the previous slide: M S T (v n ) = S(T (v n )) = S( a nm w m ) Therefore where = m=1 m=1 M M a nm S(w m ) = c jn = m=1 S T (v n ) = a nm J b jm q j = j=1 J c jn q j j=1 M b jm a mn for j = 1,..., J m=1 J M ( b jm a mn )q j j=1 m=1 Thus the J N matrix C = [c jn ] represents S T. What is C? The product BA (check the definition of matrix product).

Analytic Geomtetry: Lines How do we talk about points, lines, planes,... in R n? Definition A line in R n is described by a point x and a direction v. It can be represented as {z R n : there exists t R such that z = x + tv} If t [0, 1], this is the line segment connecting x to x + v. REMARK Even in R n two points still determine a line: the line connecting x to y is the line containing x in the direction v. Check that this is the same as the line through y in the direction v.

Analytic Geomtetry: Hyperplanes Definition A hyperplane is a set described by a point x 0 R n and a normal direction of the plane p R n, p 0. It can be represented as {z R n : p (z x 0 ) = 0}. A hyperplane consists of all the z such that the direction z x 0 is orthogonal to p. Hyperplanes can equivalently be written as p z = p x 0 (and if x 0 = 0 this is p z = 0) In R 2 lines are also hyperplanes. In R 3 hyperplanes are ordinary planes. REMARK Lines and hyperplanes are two kinds of flat subsets of R n. Lines are subsets of dimension one. Hyperplanes are subsets of dimension n 1 or co-dimension one. One can have flat subsets of any dimension less than n.

Linear Manifolds Lines and hyperplanes are not subspaces because they do not contain the origin. (what is a subspace anyhow?) They are obtained by translating a subspace: adding the same constant to all of its elements. Definition A linear manifold of R n is a set S such that there is a subspace V on R n and x 0 R n with S = V + {x 0 } {y : y = v + x 0 for some v V } Lines and hyperplanes are linear manifolds (not linear subspaces).

Lines Here is another way to describe a line. Given two points x and y in R n, the line that passes through these points is: {z R n : z = x + t(y x) for some t}. This is called parametric representation (it defines n equations). Example Given y, x R 2, we find a line through them by solving: z 1 = x 1 + t(y 1 x 1 ) and z 2 = x 2 + t(y 2 x 2 ). Solve the first equation for t and substite out: z 2 = x 2 + (y 2 x 2 ) (z 1 x 1 ) or z 2 x 2 = y 2 x 2 (z 1 x 1 ). y 1 x 1 y 1 x 1 This is the standard way to represent the equation of a line (in the plane) through (x 1, x 2 ) with slope (y 2 x 2 )(y 1 x 1 ).

Lines Given two points x and y in R n, the line that passes through these points is: {z R n : z = x + t(y x) for some t}. This is called parametric representation (it defines n equations). The parametric representation is almost equivalent to the standard representation in R 2. Why almost? It is more general since it allows for lines parallel to the axes. One needs two pieces of information to describe a line: Point and direction, or Two points (you get the direction by subtracting the points).

Hyperplanes How does one describe an hyperplane? One way is to use a point and a (normal) direction {z R n : p (z x 0 ) = 0} Another way is to use n points in R n, provided these are in general position. You can go from points to the normal solving a linear system of equations.

How to Find an Hyperplane when n=3 In R 3 an hyperplane is a plane and one can describe it using three points. Example Given (1, 2, 3), (0, 1, 1), (2, 1, 1) find A, B, C, D such that Ax 1 + Bx 2 + Cx 3 = D A + 2B 3C = D Solve the following system of equations: B + C = D 2A + B + C = D This yields (A, B, C, D) = (0,.8D,.2D, D). Hence, if we find one set of coeffi cients, any non-zero multiple will also work. Hence an equation for the plane is: 4x 2 + x 3 = 5 (check that the three points actually satisfy this equation).

Describing Hyperplanes with Normals Given some points, look for a normal direction. A normal direction is a direction that is orthogonal to all directions in the plane. A direction in the plane is a direction of a line in the plane. Hence, we can get such a direction by subtracting any two points in the plane. Given (1, 2, 3), (0, 1, 1), (2, 1, 1) find a two dimensional hyperplane It has two independent directions. One direction comes from the difference between the first two points: (1, 2, 3) (0, 1, 1) = (1, 1, 4). The other can come from the difference between the second and third points (0, 1, 1) (2, 1, 1) = ( 2, 0, 0). We can now find a normal to both of them. That is, a p such that p 0 and p (1, 1, 4) = p ( 2, 0, 0) = 0 any multiple of (0, 4, 1) solves this system of two equations and three unknowns. Hence, the equation for the hyperplane is (0, 4, 1) (x 1 1, x 2 2, x 3 + 3) = 0

Separating Hyperplanes and Convexity Consider two sets X and Y in R n which do not interserct. A hyperplane in R n is expressed as p x = c where x R n, p 0 is in R n, and c R. Under what conditions can we find an hyperplane that divides the space in two, each side containing only one of those sets? Draw a few pictures. The following is a minimal condition sets have to satisfy to make separation possible. Definition X R n is convex if x, y X and α [0, 1] we have αx + (1 α)y X

Separating Hyperplane Theorem Start with the simpler case of separate a point from a set. Theorem (Separating Hyperplane) Given a nonempty, closed, convex set X R n and x R n, with x / X, There exists a p R n, p 0, and a c R such that and That is, X {y R n : p y c} p x < c p y = c defines a separating hyperplane for X : it leaves all of X on one side and x on the other. Without loss of generality, one can normalize the normal to the separating hyperplane. That is, we can assume that p = 1.

Proof of the separating hyperplane theorem in 9(+1 for you) steps Consider the problem of minimizing the distance between x and X. That is: find a vector that solves min y X x y. 1 X is nonempty, so we can find some element z X. While X is not necessarily bounded, without loss of generality we can replace X by {y X : y x z x }. This set is compact because X is closed. 2 The norm is a continuous function. Hence there is a solution y. 3 Let p = y x. Since x / X, p 0. 4 Let c = p y. Since c p x = p (y x) = p 2, therefore c > p x. 5 We need to show if y X, then p y c. 6 Notice how this inequality is equivalent to (y x) (y y ) 0. 7 Since X is convex and y solves the minimzation problem, it must be that ty + (1 t)y x 2 is minimized when t = 0. 8 Since the derivative of ty + (1 t)y x 2 is non-negative at t = 0, the first order condition will do it. 9 Check that differentiating ty + (1 t)y x 2 (as a function of t) and simplifying yields the desired inequality.

Extensions If x is in the boundary of X, then you can approximate x by a sequence x k such that each x k / X. This yields a sequence of p k, which can be taken to be unit vectors, that satisfy the conclusion of the theorem. A subsequence of the p k must converge. The limit point p will satisfy the conclusion of the theorem (except we can only guarantee that c p x rather than the strict equality). The closure of any convex set is convex. Given a convex set X and a point x not in the interior of the set, we can separate x from the closure of X. Putting these things together...

A Stronger Separating Hyperplane Theorem Theorem (Supporting Hyperplane) Given a convex set X R n and x R n, x. If x is not in the interior of X, then there exists p R n, p 0, and c R such that and X {y R n p y c} p x c Other general versions separate two convex sets. The easiest case is when the sets do not have any point in common, and the slightly harder is the one in which no point of one set is in the interior of the other set. Let A, B R n be nonempty, disjoint convex sets. Then there exists a nonzero vector p R n such that p a p b a A, b B The separating hyperplane theorem can be used to provide intuition for the way we solve constrained optimization problems. In the typical economic application, the separating hyperplane s normal is a price vector, and the separation property states that a particular vector costs more than vectors in a consumption set.

Tomorrow Calculus 1 Level Sets 2 Derivatives and Partial Derivatives 3 Differentiability 4 Tangents to Level Sets