0.1 Tangent Spaces and Lagrange Multipliers
|
|
- Laurence Barker
- 5 years ago
- Views:
Transcription
1 01 TANGENT SPACES AND LAGRANGE MULTIPLIERS 1 01 Tangent Spaces and Lagrange Multipliers If a differentiable function G = (G 1,, G k ) : E n+k E k then the surface S defined by S = { x G( x) = v} is called the level surface for G( x) = v Note that each of the functions G i : E n+k R If we denote by S i the level surface for the equation G i ( x) = v i, then S = k i=1 S i Suppose that x 0 = (x 0 1,, x 0 n+k ) S and that G ( x 0 ) L(E n+k, E k ) has rank k Let δ i,j = 1 if i = j and 0 if i j With respect to the standard basis { e j = (δ 1,j,, δ n+k,j ) j = 1, 2,, n + k} for E n+k and the analogous smaller basis for E k, we note that the matrix [ G ( x 0 )] k (n+k) has its whole set of k row vectors linearly independent, and these row vectors are the gradient vectors G 1 ( x 0 ), G 2 ( x 0 ),, G k ( x 0 ) Let φ : R S be a differentiable function for which φ(0) = x 0 Then we call the vector v = φ (0) a tangent vector to S at x 0 Definition 011 The tangent space T x 0(S) at x 0 S is the set of all tangent vectors to S at x 0 The translate x 0 + T x 0 = { x 0 + v v T x 0(S)} is called the tangent plane to the surface S, with point of tangency at x 0 A translate a + V = { a + x x V } of a vector subspace V of E n is called an affine subspace of E n An affine subspace is a vector subspace if and only if a V (See Exercise 1) Theorem 011 Let G : E n+k E k be a differentiable function Let x 0 S = { x G( x) = v} Suppose G ( x 0 ) has rank k Then the tangent space T x 0(S) is the vector subspace T x 0(S) = ( span R { G 1 ( x 0 ),, G k ( x 0 )} ) of dimension n In words, T x 0(S) is the orthogonal complement of the span of the k gradient vectors G 1 ( x 0 ),, G k ( x 0 )
2 2 Proof Suppose first that v = φ (0) T x 0(S) This implies that φ maps into each level surface G i ( x) = v i We will show that v G i ( x 0 ) for each i = 1,, k In fact, G i ( φ(t)) v i, a real constant We differentiate using the Chain Rule to find that G i( φ(0)) φ (0) = 0 In terms of the matrix representation of the left side of the latter equation, we have G i ( x 0 ) φ (0) = 0, so that v G i ( x 0 ) This shows that T x 0(S) G i ( x 0 ) for each i This implies that T x 0(S) ( span R { G 1 ( x 0 ),, G k ( x 0 )} ) The hypothesis that rank( G ( x 0 )) = k implies that dimt x 0(S) n If we can show that the tangent space is at least n-dimensional, then it will have to be the entire orthogonal complement of the span of the gradient vectors as claimed Thus it will suffice to produce a linearly independent set of n vectors in the tangent space Because the rank of a matrix is also the number of linearly independent column vectors, it follows that the matrix [ G ( x 0 )] has k independent columns We can rearrange the order of the n elements of the standard basis of E n+k to arrange that the first k columns are linearly independent By the Implicit Function Theorem, there exists an open set U E k containing (x 0 1,, x 0 k ) and an open set V En containing (x 0 k+1,, x0 k+n ) such that there are unique differentiable functions solving the equation x 1 = ψ 1 (x k+1,, x k+n ) x k = ψ k (x k+1,, x k+n ) G(ψ 1 (x k+1,, x k+n ),, ψ k (x k+1,, x k+n ), x k+1,, x k+n ) = v Next we define n differentiable curves on S by the equations φ 1 (t) = (ψ 1 (x 0 k+1 + t, x0 k+2,, x0 k+n ),, ψ k(x 0 k+1 + t, x0 k+2,, x0 k+n ), x0 k+1 + t, x0 k+2,, x0 k+n ) φ n (t) = (ψ 1 (x 0 k+1,, x0 n+k 1, x0 k+n + t),, ψ k(x 0 k+1,, x0 k+n 1, x0 k+n + t), x0 k+1, x0 k+n 1, x0 k+n + t) In comparing the vectors φ i(0) for i = 1,, n, observe that for each of these vectors the final n entries are all 0 except for a single entry which is 1 The location of the 1 is different for each of these vectors Thus the n vectors are independent and the theorem is proved
3 01 TANGENT SPACES AND LAGRANGE MULTIPLIERS 3 Corollary 011 Let G : E k+n E k be a differentiable function and let x 0 S = { x G( x) = v} Suppose x 0 is a local extreme point of a differentiable function f : S R and that G ( x 0 ) has rank k Then there exist numbers λ 1,, λ k such that f( x 0 ) = λ 1 G 1 ( x 0 ) + + λ k G k ( x 0 ) (1) The numbers λ 1,, λ k are called Lagrange multipliers Proof If φ : R S is a differentiable curve on S with φ(0) = x 0, let ψ(t) = f( φ(t)) Since this function has an extreme point at 0, we have ψ (0) = f( φ(0)) φ (0) = 0 It follows from Theorem 011 that f( x 0 ) is orthogonal to the tangent space T x 0(S) Since the co-dimension of T x 0(S) is k, it follows that f( x 0 ) lies in the span of the k vectors G 1 ( x 0 ),, G k ( x 0 ) This proves the corollary The method of Lagrange multipliers permits an optimization problem to be replaced by a problem of solving a system of equations From the k + n components of the vectors in Equation 1, we obtain a system of k + n equations in the n+2k unknowns x 1,, x k+n, λ 1,, λ k We get k additional equations from the k components of the equation G( x) = v Thus we obtain a system of n+2k equations in n+2k unknowns Although we have replaced a calculus problem with an algebraic problem, the algebraic problem can be challenging Nevertheless, the method of Lagrange multipliers is a powerful tool for optimization problems Example 011 We will begin with a three-dimensional example Consider the surface S defined by the equation x 4 + y 4 + z 4 = 1 in E 3, shown in Figure 1 We will find both the maximum and the minimum values of the function f( x) = x 2 +y 2 +z 2 on S (In effect, we are determining the closest and furthest distances from the origin on S) In this example, we denote x = (x, y, z) Observe that if we define G( x) = x 4 + y 4 + z 4 then S = G 1 ({1}) Hence S is closed because G is continuous S is also bounded (Why?) Hence the function f must achieve both a maximum and a minimum value somewhere on S Since S is smooth at all points and since G is non-vanishing on S,
4 4 y z x Figure 1: x 4 + y 4 + z 4 = 1 the extreme points must occur at those points for which f( x) = λ G( x) This yields the following system of equations x(1 2λx 2 ) = 0 y(1 2λy 2 ) = 0 z(1 2λz 2 ) = 0 x 4 + y 4 + z 4 = 1 The reader should check the following by making the necessary calculations If none of the three variables is zero, then x 2 = y 2 = z 2 = 1 2λ showing that λ = ± 3 2 This implies that f(x, y, z) = 3 If exactly one of the three variables is zero, then at a point satisfying the system of equations we must have f(x, y, z) = 2 If exactly two of the variables are zero, then at a point satisfying the system we must have f(x, y, z) = 1
5 01 TANGENT SPACES AND LAGRANGE MULTIPLIERS 5 It follows that the maximum value of f on S is 3 But the reader should be able to explain why at least one of the variables must be non-zero Thus the minimum value is 1 There is also an easy way to explain even from the outset why f(x, y, z) 1 everywhere on S Exercises 01 1 Prove that the tangent plane x 0 + T x 0(S) is a vector subspace of E n if and only if x 0 T x 0(S) 2 Describe both the tangent space and the tangent ( plane to the sphere S n 1 = { x E n x = 1} at the point x 0 1 = 1 n, n 1,, n ) 3 The sphere S 3 E 4 is defined by { S 3 = x } 4 x 2 i = 1 i=1 Define f : S 3 R by f( x) = 4 i=1 a ix i where a i is a constant for each i {1, 2, 3, 4} Show that the maximum and minimum values of f on S 3 are ± 4 i=1 a2 i 4 The group SL(2, R) of matrices was defined in Exercise???? Let f : SL(2, R) R be defined by f( x) = x x x x 2 4, where we identify the matrix ( ) x1 x X = 2 SL(2, R) x 3 x 4 with the vector x = (x 1, x 2, x 3, x 4 ) constrained to the surface S in E 4 defined by the equation x 1 x 4 x 2 x 3 = 1 (a) Prove that f achieves a minimum value on S but that it has no maximum (b) Use the method of Lagrange multipliers to find the minimum value of f on S
6 6 5 Let f( x) = 4 i=1 x2 i for all x E 4 Let S 1 be the surface in determined by x 1 x 4 x 2 x 3 = 1 and let S 2 be the surface defined by 4 i=1 x i = 2 Let S = S 1 S 2 (a) Prove that f( x) has a minimum value on S but no maximum (b) Find the minimum value of f( x) on S
Math Advanced Calculus II
Math 452 - Advanced Calculus II Manifolds and Lagrange Multipliers In this section, we will investigate the structure of critical points of differentiable functions. In practice, one often is trying to
More informationTangent spaces, normals and extrema
Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent
More information1 Lagrange Multiplier Method
1 Lagrange Multiplier Method Near a maximum the decrements on both sides are in the beginning only imperceptible. J. Kepler When a quantity is greatest or least, at that moment its flow neither increases
More informationMath Topology II: Smooth Manifolds. Spring Homework 2 Solution Submit solutions to the following problems:
Math 132 - Topology II: Smooth Manifolds. Spring 2017. Homework 2 Solution Submit solutions to the following problems: 1. Let H = {a + bi + cj + dk (a, b, c, d) R 4 }, where i 2 = j 2 = k 2 = 1, ij = k,
More information10. Smooth Varieties. 82 Andreas Gathmann
82 Andreas Gathmann 10. Smooth Varieties Let a be a point on a variety X. In the last chapter we have introduced the tangent cone C a X as a way to study X locally around a (see Construction 9.20). It
More informationAPPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of
CHAPTER III APPLICATIONS The eigenvalues are λ =, An orthonormal basis of eigenvectors consists of, The eigenvalues are λ =, A basis of eigenvectors consists of, 4 which are not perpendicular However,
More informationCourse Summary Math 211
Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.
More informationIn view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written
11.8 Inequality Constraints 341 Because by assumption x is a regular point and L x is positive definite on M, it follows that this matrix is nonsingular (see Exercise 11). Thus, by the Implicit Function
More informationChapter 7. Extremal Problems. 7.1 Extrema and Local Extrema
Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced
More informationIntroduction to Arithmetic Geometry Fall 2013 Lecture #17 11/05/2013
18.782 Introduction to Arithmetic Geometry Fall 2013 Lecture #17 11/05/2013 Throughout this lecture k denotes an algebraically closed field. 17.1 Tangent spaces and hypersurfaces For any polynomial f k[x
More informationMULTIVARIABLE CALCULUS, LINEAR ALGEBRA, AND DIFFERENTIAL EQUATIONS
T H I R D E D I T I O N MULTIVARIABLE CALCULUS, LINEAR ALGEBRA, AND DIFFERENTIAL EQUATIONS STANLEY I. GROSSMAN University of Montana and University College London SAUNDERS COLLEGE PUBLISHING HARCOURT BRACE
More informationMathematics Department Stanford University Math 61CM/DM Inner products
Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationcan only hit 3 points in the codomain. Hence, f is not surjective. For another example, if n = 4
.. Conditions for Injectivity and Surjectivity In this section, we discuss what we can say about linear maps T : R n R m given only m and n. We motivate this problem by looking at maps f : {,..., n} {,...,
More informationChapter 6. Orthogonality
6.4 The Projection Matrix 1 Chapter 6. Orthogonality 6.4 The Projection Matrix Note. In Section 6.1 (Projections), we projected a vector b R n onto a subspace W of R n. We did so by finding a basis for
More informationCSL361 Problem set 4: Basic linear algebra
CSL361 Problem set 4: Basic linear algebra February 21, 2017 [Note:] If the numerical matrix computations turn out to be tedious, you may use the function rref in Matlab. 1 Row-reduced echelon matrices
More informationMATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian.
MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian. Spanning set Let S be a subset of a vector space V. Definition. The span of the set S is the smallest subspace W V that contains S. If
More informationSchur s Triangularization Theorem. Math 422
Schur s Triangularization Theorem Math 4 The characteristic polynomial p (t) of a square complex matrix A splits as a product of linear factors of the form (t λ) m Of course, finding these factors is a
More informationORIE 6300 Mathematical Programming I August 25, Recitation 1
ORIE 6300 Mathematical Programming I August 25, 2016 Lecturer: Calvin Wylie Recitation 1 Scribe: Mateo Díaz 1 Linear Algebra Review 1 1.1 Independence, Spanning, and Dimension Definition 1 A (usually infinite)
More informationf(x, y) = 1 2 x y2 xy 3
Problem. Find the critical points of the function and determine their nature. We have We find the critical points We calculate f(x, y) = 2 x2 + 3 2 y2 xy 3 f x = x y 3 = 0 = x = y 3 f y = 3y 3xy 2 = 0
More informationLagrange Multipliers
Lagrange Multipliers (Com S 477/577 Notes) Yan-Bin Jia Nov 9, 2017 1 Introduction We turn now to the study of minimization with constraints. More specifically, we will tackle the following problem: minimize
More informationTranspose & Dot Product
Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:
More informationLinear Algebra, 4th day, Thursday 7/1/04 REU Info:
Linear Algebra, 4th day, Thursday 7/1/04 REU 004. Info http//people.cs.uchicago.edu/laci/reu04. Instructor Laszlo Babai Scribe Nick Gurski 1 Linear maps We shall study the notion of maps between vector
More informationMath 21b: Linear Algebra Spring 2018
Math b: Linear Algebra Spring 08 Homework 8: Basis This homework is due on Wednesday, February 4, respectively on Thursday, February 5, 08. Which of the following sets are linear spaces? Check in each
More information1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?
. Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in
More informationLinear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra
MTH6140 Linear Algebra II Notes 2 21st October 2010 2 Matrices You have certainly seen matrices before; indeed, we met some in the first chapter of the notes Here we revise matrix algebra, consider row
More informationTranspose & Dot Product
Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:
More informationMath 215B: Solutions 1
Math 15B: Solutions 1 Due Thursday, January 18, 018 (1) Let π : X X be a covering space. Let Φ be a smooth structure on X. Prove that there is a smooth structure Φ on X so that π : ( X, Φ) (X, Φ) is an
More information3.7 Constrained Optimization and Lagrange Multipliers
3.7 Constrained Optimization and Lagrange Multipliers 71 3.7 Constrained Optimization and Lagrange Multipliers Overview: Constrained optimization problems can sometimes be solved using the methods of the
More informationA geometric proof of the spectral theorem for real symmetric matrices
0 0 0 A geometric proof of the spectral theorem for real symmetric matrices Robert Sachs Department of Mathematical Sciences George Mason University Fairfax, Virginia 22030 rsachs@gmu.edu January 6, 2011
More informationUpon successful completion of MATH 220, the student will be able to:
MATH 220 Matrices Upon successful completion of MATH 220, the student will be able to: 1. Identify a system of linear equations (or linear system) and describe its solution set 2. Write down the coefficient
More information11 Block Designs. Linear Spaces. Designs. By convention, we shall
11 Block Designs Linear Spaces In this section we consider incidence structures I = (V, B, ). always let v = V and b = B. By convention, we shall Linear Space: We say that an incidence structure (V, B,
More informationReview 1 Math 321: Linear Algebra Spring 2010
Department of Mathematics and Statistics University of New Mexico Review 1 Math 321: Linear Algebra Spring 2010 This is a review for Midterm 1 that will be on Thursday March 11th, 2010. The main topics
More informationDefinition 2.3. We define addition and multiplication of matrices as follows.
14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row
More informationA SHORT SUMMARY OF VECTOR SPACES AND MATRICES
A SHORT SUMMARY OF VECTOR SPACES AND MATRICES This is a little summary of some of the essential points of linear algebra we have covered so far. If you have followed the course so far you should have no
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationMODULE 8 Topics: Null space, range, column space, row space and rank of a matrix
MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix Definition: Let L : V 1 V 2 be a linear operator. The null space N (L) of L is the subspace of V 1 defined by N (L) = {x
More informationImplicit Functions, Curves and Surfaces
Chapter 11 Implicit Functions, Curves and Surfaces 11.1 Implicit Function Theorem Motivation. In many problems, objects or quantities of interest can only be described indirectly or implicitly. It is then
More informationSystems of Linear Equations
LECTURE 6 Systems of Linear Equations You may recall that in Math 303, matrices were first introduced as a means of encapsulating the essential data underlying a system of linear equations; that is to
More informationPhysics 6010, Fall 2016 Constraints and Lagrange Multipliers. Relevant Sections in Text:
Physics 6010, Fall 2016 Constraints and Lagrange Multipliers. Relevant Sections in Text: 1.3 1.6 Constraints Often times we consider dynamical systems which are defined using some kind of restrictions
More informationClass Meeting # 12: Kirchhoff s Formula and Minkowskian Geometry
MATH 8.52 COURSE NOTES - CLASS MEETING # 2 8.52 Introduction to PDEs, Spring 207 Professor: Jared Speck Class Meeting # 2: Kirchhoff s Formula and Minkowskian Geometry. Kirchhoff s Formula We are now ready
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More information2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.
Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.
More informationMath Linear Algebra Final Exam Review Sheet
Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of
More informationLagrange multipliers. Portfolio optimization. The Lagrange multipliers method for finding constrained extrema of multivariable functions.
Chapter 9 Lagrange multipliers Portfolio optimization The Lagrange multipliers method for finding constrained extrema of multivariable functions 91 Lagrange multipliers Optimization problems often require
More informationDetermine whether the following system has a trivial solution or non-trivial solution:
Practice Questions Lecture # 7 and 8 Question # Determine whether the following system has a trivial solution or non-trivial solution: x x + x x x x x The coefficient matrix is / R, R R R+ R The corresponding
More informationName: Instructor: Lecture time: TA: Section time:
Math 2220 Prelim 1 February 17, 2009 Name: Instructor: Lecture time: TA: Section time: INSTRUCTIONS READ THIS NOW This test has 6 problems on 7 pages worth a total of 100 points. Look over your test package
More informationOHSX XM511 Linear Algebra: Multiple Choice Exercises for Chapter 2
OHSX XM5 Linear Algebra: Multiple Choice Exercises for Chapter. In the following, a set is given together with operations of addition and scalar multiplication. Which is not a vector space under the given
More informationBROUWER FIXED POINT THEOREM. Contents 1. Introduction 1 2. Preliminaries 1 3. Brouwer fixed point theorem 3 Acknowledgments 8 References 8
BROUWER FIXED POINT THEOREM DANIELE CARATELLI Abstract. This paper aims at proving the Brouwer fixed point theorem for smooth maps. The theorem states that any continuous (smooth in our proof) function
More informationLecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018
Lecture 13: Orthogonal projections and least squares (Section 3.2-3.3) Thang Huynh, UC San Diego 2/9/2018 Orthogonal projection onto subspaces Theorem. Let W be a subspace of R n. Then, each x in R n can
More informationMATH 4211/6211 Optimization Constrained Optimization
MATH 4211/6211 Optimization Constrained Optimization Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Constrained optimization
More information10-725/36-725: Convex Optimization Prerequisite Topics
10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the
More informationFunctional Analysis Exercise Class
Functional Analysis Exercise Class Week 9 November 13 November Deadline to hand in the homeworks: your exercise class on week 16 November 20 November Exercises (1) Show that if T B(X, Y ) and S B(Y, Z)
More informationM.6. Rational canonical form
book 2005/3/26 16:06 page 383 #397 M.6. RATIONAL CANONICAL FORM 383 M.6. Rational canonical form In this section we apply the theory of finitely generated modules of a principal ideal domain to study the
More informationLeast squares problems Linear Algebra with Computer Science Application
Linear Algebra with Computer Science Application April 8, 018 1 Least Squares Problems 11 Least Squares Problems What do you do when Ax = b has no solution? Inconsistent systems arise often in applications
More informationSolution to Homework 1
Solution to Homework Sec 2 (a) Yes It is condition (VS 3) (b) No If x, y are both zero vectors Then by condition (VS 3) x = x + y = y (c) No Let e be the zero vector We have e = 2e (d) No It will be false
More informationThere are two things that are particularly nice about the first basis
Orthogonality and the Gram-Schmidt Process In Chapter 4, we spent a great deal of time studying the problem of finding a basis for a vector space We know that a basis for a vector space can potentially
More informationQ1 Q2 Q3 Q4 Tot Letr Xtra
Mathematics 54.1 Final Exam, 12 May 2011 180 minutes, 90 points NAME: ID: GSI: INSTRUCTIONS: You must justify your answers, except when told otherwise. All the work for a question should be on the respective
More information(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.
1 Which of the following statements is always true? I The null space of an m n matrix is a subspace of R m II If the set B = {v 1,, v n } spans a vector space V and dimv = n, then B is a basis for V III
More informationMath 302 Outcome Statements Winter 2013
Math 302 Outcome Statements Winter 2013 1 Rectangular Space Coordinates; Vectors in the Three-Dimensional Space (a) Cartesian coordinates of a point (b) sphere (c) symmetry about a point, a line, and a
More informationHomework 5 M 373K Mark Lindberg and Travis Schedler
Homework 5 M 373K Mark Lindberg and Travis Schedler 1. Artin, Chapter 3, Exercise.1. Prove that the numbers of the form a + b, where a and b are rational numbers, form a subfield of C. Let F be the numbers
More informationLinear Algebra I Lecture 8
Linear Algebra I Lecture 8 Xi Chen 1 1 University of Alberta January 25, 2019 Outline 1 2 Gauss-Jordan Elimination Given a system of linear equations f 1 (x 1, x 2,..., x n ) = 0 f 2 (x 1, x 2,..., x n
More informationHOMEWORK 7 SOLUTIONS
HOMEWORK 7 SOLUTIONS MA11: ADVANCED CALCULUS, HILARY 17 (1) Using the method of Lagrange multipliers, find the largest and smallest values of the function f(x, y) xy on the ellipse x + y 1. Solution: The
More informationPh.D. Katarína Bellová Page 1 Mathematics 2 (10-PHY-BIPMA2) EXAM - Solutions, 20 July 2017, 10:00 12:00 All answers to be justified.
PhD Katarína Bellová Page 1 Mathematics 2 (10-PHY-BIPMA2 EXAM - Solutions, 20 July 2017, 10:00 12:00 All answers to be justified Problem 1 [ points]: For which parameters λ R does the following system
More information12.10 Lagrange Multipliers
.0 Lagrange Multipliers In the last two sections we were often solving problems involving maimizing or minimizing a function f subject to a 'constraint' equation g. For eample, we minimized the cost of
More informationAnalysis II: The Implicit and Inverse Function Theorems
Analysis II: The Implicit and Inverse Function Theorems Jesse Ratzkin November 17, 2009 Let f : R n R m be C 1. When is the zero set Z = {x R n : f(x) = 0} the graph of another function? When is Z nicely
More informationMATHEMATICAL ECONOMICS: OPTIMIZATION. Contents
MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4
More informationOrthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6
Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and
More informationALGEBRAIC GEOMETRY I - FINAL PROJECT
ALGEBRAIC GEOMETRY I - FINAL PROJECT ADAM KAYE Abstract This paper begins with a description of the Schubert varieties of a Grassmannian variety Gr(k, n) over C Following the technique of Ryan [3] for
More informationIr O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )
Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O
More informationMoreover this binary operation satisfies the following properties
Contents 1 Algebraic structures 1 1.1 Group........................................... 1 1.1.1 Definitions and examples............................. 1 1.1.2 Subgroup.....................................
More informationOptimization Theory. A Concise Introduction. Jiongmin Yong
October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization
More information11. Dimension. 96 Andreas Gathmann
96 Andreas Gathmann 11. Dimension We have already met several situations in this course in which it seemed to be desirable to have a notion of dimension (of a variety, or more generally of a ring): for
More informationConvex Functions and Optimization
Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized
More informationRow Space, Column Space, and Nullspace
Row Space, Column Space, and Nullspace MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Introduction Every matrix has associated with it three vector spaces: row space
More information1.8 Dual Spaces (non-examinable)
2 Theorem 1715 is just a restatement in terms of linear morphisms of a fact that you might have come across before: every m n matrix can be row-reduced to reduced echelon form using row operations Moreover,
More informationLinear Algebra Practice Problems
Linear Algebra Practice Problems Page of 7 Linear Algebra Practice Problems These problems cover Chapters 4, 5, 6, and 7 of Elementary Linear Algebra, 6th ed, by Ron Larson and David Falvo (ISBN-3 = 978--68-78376-2,
More informationLeast Sparsity of p-norm based Optimization Problems with p > 1
Least Sparsity of p-norm based Optimization Problems with p > Jinglai Shen and Seyedahmad Mousavi Original version: July, 07; Revision: February, 08 Abstract Motivated by l p -optimization arising from
More informationChapter 3: Theory Review: Solutions Math 308 F Spring 2015
Chapter : Theory Review: Solutions Math 08 F Spring 05. What two properties must a function T : R m R n satisfy to be a linear transformation? (a) For all vectors u and v in R m, T (u + v) T (u) + T (v)
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More informationSummary Notes on Maximization
Division of the Humanities and Social Sciences Summary Notes on Maximization KC Border Fall 2005 1 Classical Lagrange Multiplier Theorem 1 Definition A point x is a constrained local maximizer of f subject
More informationMA 265 FINAL EXAM Fall 2012
MA 265 FINAL EXAM Fall 22 NAME: INSTRUCTOR S NAME:. There are a total of 25 problems. You should show work on the exam sheet, and pencil in the correct answer on the scantron. 2. No books, notes, or calculators
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationChapter 6 - Orthogonality
Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/
More informationMA 0540 fall 2013, Row operations on matrices
MA 0540 fall 2013, Row operations on matrices December 2, 2013 This is all about m by n matrices (of real or complex numbers). If A is such a matrix then A corresponds to a linear map F n F m, which we
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationL p Spaces and Convexity
L p Spaces and Convexity These notes largely follow the treatments in Royden, Real Analysis, and Rudin, Real & Complex Analysis. 1. Convex functions Let I R be an interval. For I open, we say a function
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More information4 Hilbert spaces. The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan
The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan Wir müssen wissen, wir werden wissen. David Hilbert We now continue to study a special class of Banach spaces,
More informationELE/MCE 503 Linear Algebra Facts Fall 2018
ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2
More informationIntroduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder
Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Bastian Steder Reference Book Thrun, Burgard, and Fox: Probabilistic Robotics Vectors Arrays of numbers Vectors represent
More informationNOTES (1) FOR MATH 375, FALL 2012
NOTES 1) FOR MATH 375, FALL 2012 1 Vector Spaces 11 Axioms Linear algebra grows out of the problem of solving simultaneous systems of linear equations such as 3x + 2y = 5, 111) x 3y = 9, or 2x + 3y z =
More information3.1. Derivations. Let A be a commutative k-algebra. Let M be a left A-module. A derivation of A in M is a linear map D : A M such that
ALGEBRAIC GROUPS 33 3. Lie algebras Now we introduce the Lie algebra of an algebraic group. First, we need to do some more algebraic geometry to understand the tangent space to an algebraic variety at
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More informationLinear Algebra. Min Yan
Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................
More informationElementary maths for GMT
Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More information