Mechanical Systems II. Method of Lagrange Multipliers

Similar documents
Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers

MATH 4211/6211 Optimization Constrained Optimization

8. Constrained Optimization

Optimization using Calculus. Optimization of Functions of Multiple Variables subject to Equality Constraints

Lagrange Multipliers

Generalization to inequality constrained problem. Maximize

3.7 Constrained Optimization and Lagrange Multipliers

HOMEWORK 7 SOLUTIONS

Functional differentiation

Paul Schrimpf. October 17, UBC Economics 526. Constrained optimization. Paul Schrimpf. First order conditions. Second order conditions

Lagrange multipliers. Portfolio optimization. The Lagrange multipliers method for finding constrained extrema of multivariable functions.

MATH529 Fundamentals of Optimization Constrained Optimization I

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Numerical Optimization

2 Lecture Defining Optimization with Equality Constraints

ECE580 Partial Solution to Problem Set 3

25. Chain Rule. Now, f is a function of t only. Expand by multiplication:

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Lagrange Multipliers

1 Lagrange Multiplier Method

Partial Derivatives. w = f(x, y, z).

ECE580 Solution to Problem Set 6

Constrained optimization

MATH2070 Optimisation

Mathematical Economics. Lecture Notes (in extracts)

Constrained controllability of semilinear systems with delayed controls

Math 273a: Optimization Basic concepts

Math Advanced Calculus II

Nonlinear Optimization: What s important?

An Introduction to Model-based Predictive Control (MPC) by

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016

Optimization Methods: Optimization using Calculus - Equality constraints 1. Module 2 Lecture Notes 4

The Fundamental Theorem of Linear Inequalities

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Introduction to the Calculus of Variations

Ph.D. Katarína Bellová Page 1 Mathematics 2 (10-PHY-BIPMA2) EXAM - Solutions, 20 July 2017, 10:00 12:00 All answers to be justified.

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Optimality Conditions

September Math Course: First Order Derivative

The Karush-Kuhn-Tucker conditions

Calculus of Variations

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

Math 5311 Constrained Optimization Notes

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

Lagrangian Duality. Evelien van der Hurk. DTU Management Engineering

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

Lecture Notes: Geometric Considerations in Unconstrained Optimization

0.1 Tangent Spaces and Lagrange Multipliers

1.2 Derivation. d p f = d p f(x(p)) = x fd p x (= f x x p ). (1) Second, g x x p + g p = 0. d p f = f x g 1. The expression f x gx

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Lecture 4: Optimization. Maximizing a function of a single variable

Exercises * on Linear Algebra

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

Implicit Functions, Curves and Surfaces

Harmonic Oscillator I

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Course Summary Math 211

f(x) f(z) c x z > 0 1

ENGI Partial Differentiation Page y f x

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Calculus of Variations Summer Term 2014

Lecture 18: Optimization Programming

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

Math 234 Final Exam (with answers) Spring 2017

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

minimize x subject to (x 2)(x 4) u,

IE 5531: Engineering Optimization I

Tangent spaces, normals and extrema

Optimization Theory. Lectures 4-6

This exam will be over material covered in class from Monday 14 February through Tuesday 8 March, corresponding to sections in the text.

Math 118, Fall 2014 Final Exam

May 9, 2014 MATH 408 MIDTERM EXAM OUTLINE. Sample Questions

Support Vector Machines

Review for the Final Exam

Constrained Optimization

Math 212-Lecture Interior critical points of functions of two variables

Chap 3. Linear Algebra

Lecture Notes for Chapter 12

Vector Derivatives and the Gradient

Parametric Equations, Function Composition and the Chain Rule: A Worksheet

Review of Optimization Methods

Strong and Weak Augmentability in Calculus of Variations

Conjugate Gradient (CG) Method

Differentiation. f(x + h) f(x) Lh = L.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Calculus of Variations Summer Term 2016

39.1 Absolute maxima/minima

Solutions to Homework 7

1 Computing with constraints

Variational Methods & Optimal Control

ENGI Partial Differentiation Page y f x

The theory of manifolds Lecture 3

Image restoration: numerical optimisation

Second Order Optimality Conditions for Constrained Nonlinear Programming

Econ Slides from Lecture 14

Algorithms for constrained local optimization

Optimal control problems with PDE constraints

Functions of Several Variables

MATH 105: PRACTICE PROBLEMS FOR CHAPTER 3: SPRING 2010

Transcription:

Mechanical Systems II. Method of Lagrange Multipliers Rafael Wisniewski Aalborg University Abstract. So far our approach to classical mechanics was limited to finding a critical point of a certain functional. In this lecture I shall present a first-order necessary condition for an extremum problem with additional equality constraints. This is so called a method of Lagrange multipliers. 1 Lagrange Conditions We shall deal with a first-order necessary condition for an extremum problem with equality constraints. The main result presented in this chapter is the Lagrange Multiplier Theorem. To better understand the idea of the underlying theorem we shall limit ourself to functions of two variables and one equality constraint. I shall not focus on exact properties of the functions used below, I will merely assume that all considered functions are continuously differentiable. We shall regard a problem of finding the minimum x = 1, x ) of a function f : R R subject to a constraint equation h(x) = 0, h : R R. (1) It will be vital for further analysis to introduce a notion of a level set. The level set through a point x is the set {x R : h(x) = h )}. Let h ) = 0 (i.e. x satisfies the constraint equation) and the differential ( dh h(x1, x ) ) =, h(x ) 1, x ) 0 () x 1 x 1 x we then can parameterize the level set {x : h(x) = 0} in a neighborhood of x by a curve x : (a, b) R (Why is it possible? Think about Implicit Function Theorem from the first lecture): x(t) = (x 1 (t), x (t)), t (a, b), x(t ) = x, ẋ(t ) 0. (3) Lemma 1. dh ) is orthogonal to ẋ(t ) Proof. Since h is zero on the curve x(t), we have that for all t (a, b), h(x(t)) = 0. (4)

Rafael Wisniewski Differentiating with respect to t and applying the chain rule yield 0 = d h(x(t)) = dh(x(t)), ẋ(t), (5) dt where, means the scalar product in R. Therefore dh ) is orthogonal to ẋ(t ). Now suppose x is a minimizer of f : R R on the set {x R : h(x) = 0}. Lemma. df ) is orthogonal to ẋ(t ). Proof. Observe that the composite function φ(t) = f(x(t)) (6) achieves the minimum at t. Consequently, the necessary condition for the unconstrained extremum problem implies dφ dt (t ) = 0. (7) Applying the chain rule gives 0 = dφ dt (t ) = df(x(t )), ẋ(t ) = df ), ẋ(t ). (8) Thus df ) is orthogonal to ẋ(t ). We shall now combine Lemmas 1 and. That is, both df ) and dh ) are orthogonal to ẋ(t ). Therefore, the vectors df ) and dh ) are parallel. In other words df ) is a scalar multiple of dh ). The above observation allows us to formulate the Lagrange theorem for functions of two variables with one constraint. Theorem 1 (Lagrange s Theorem). Let the point x be a minimizer (maximizer) of f : R R subject to the constraint h(x) = 0, h : R R. Then df ) and dh ) are parallel. Thus, if dh ) 0, then there exists a scalar λ such that df ) + λ dh ) = 0. (9) In the above theorem, we refer to λ as the Lagrange multiplier. Note that the Lagrange condition is only necessary but not sufficient. We shall now generalize the Lagrange s theorem for the case when f : R n R and h : R n R m, m n. We shall need a notion of a regular point. A point x is a regular point of a map g : R n R m if rank Dg ) = rank dg dx (x ) = m (10)

Mechanical Systems II. Method of Lagrange Multipliers 3 Theorem (Lagrange Multiplier Theorem). Let x be a local minimizer (or maximizer) of f : R n R, subject to h(x) = 0, h : R n R m, m n. Assume that x is a regular point of h. Then there exists λ R m such that Df ) + λ T Dh ) = 0. (11) I shall not prove the Lagrange Multiplier Theorem in this lecture, merely state that the proof uses similar reasoning as given above for the case n =, m = 1 and refer you to (1), pp. 335-343. It is often convenient to introduce the so-called Lagrangian function l : R n R m R l(x, λ) f(x) + λ T h(x). (1) The necessary condition (11) in the Lagrange Multiplier Theorem for a local minimizer x can be reformulated as Dl, λ ) = 0. (13) In other words, the necessary condition in the Lagrange Multiplier Theorem is equivalent to the first-order necessary condition for unconstrained optimization applied to the Lagrange function in (1). This is seen by differentiating l [ ] l(x, λ) l(x, λ) 0 = Dl(x, λ) =,, (14) x λ but 0 = l(x,λ) x = Df(x)+λ T Dh(x) gives the condition (11) and 0 = l(x,λ) λ = h(x), which is the equation of constrains. Example 1. Consider the problem of finding the extrema of the function f : R R, f(x 1, x ) = x 1 + x (15) on the ellipse {(x 1, x ) R : h(x 1, x ) = x 1 x 1 = 0}. (16) The lagrange function is l(x 1, x, λ) = x 1 + x + λ(x 1 + x 1). (17) The necessary conditions for l to reach extremum at 1, x, λ ) are 0 = l(x 1, x, λ) x 1 = x 1 + λ x 1 (18) 1,x,λ ) 0 = l(x 1, x, λ) x = x + 4λ x (19) 1,x,λ ) 0 = l(x 1, x, λ) λ = 1) + ) 1. (0) 1,x,λ )

4 Rafael Wisniewski From (18) x 1 = 0 or λ = 1. If x 1 = 0, then x = ± and λ = 1. If λ = 1 then x = 0 and x 1 = ±1. In conclusion we have the following solutions ( ) ( ) x (1) = 0,, x () = 0,, x (3) = ( 1, 0), x (4) = (1, 0) (1) and f(x (1) ) = f(x () ) = 1 () f(x (3) ) = f(x (4) ) = 1. (3) Thus x (1) and x () correspond to minimum, whereas the function f reaches maximum at x (3) and x (4). Example. We shall consider an example from linear algebra. Maximize the quadratic form f : R n R, x x T Qx, where Q T = Q, subject to x T P x = 1, where P T = P and P is nonsingular. The function h in the Lagrange Multiplier Theorem is h : R n R, x 1 x T P x. (4) The Lagrange function is now l(x, λ) = x T Qx + λ ( 1 x T P x ). (5) As in the previous example we compute partial derivatives of l l x (x, λ ) = ) T Q λ ) T P (6) l λ (x, λ ) = 1 ) T P x. (7) Thus the necessary conditions are Qx = λp x (8) ) T P x = 1. (9) But since P is nonsingular the equation (8) is equivalent to P 1 Qx = λ x, (30) which is the equation for the eigenvalues λs of the matrix P 1 Q. On the other hand multiplying (30) by xp yields ) T Qx = λ ) T P x = λ (31) We conclude that the maximizer of x T Qx subject to x T P x = 1 is the eigenvector of the matrix P 1 Q corresponding to the maximal eigenvalue.

References [1] S. Zak E. K.P. Chong. An Introduction to Optimization. John Wiley, 1976.