The Kuhn-Tucker and Envelope Theorems

Similar documents
The Kuhn-Tucker and Envelope Theorems

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

CHAPTER 1-2: SHADOW PRICES

Problem Set 2 Solutions

Optimization. A first course on mathematics for economists

INTRODUCTORY MATHEMATICS FOR ECONOMICS MSCS. LECTURE 3: MULTIVARIABLE FUNCTIONS AND CONSTRAINED OPTIMIZATION. HUW DAVID DIXON CARDIFF BUSINESS SCHOOL.

The Necessity of the Transversality Condition at Infinity: A (Very) Special Case

Constrained Optimization

Midterm Exam - Solutions

CHAPTER 3: OPTIMIZATION

How to Characterize Solutions to Constrained Optimization Problems

The Kuhn-Tucker Problem

1.4 FOUNDATIONS OF CONSTRAINED OPTIMIZATION

Nonlinear Programming and the Kuhn-Tucker Conditions

Optimization Paul Schrimpf September 12, 2018

E 600 Chapter 4: Optimization

Properties of Walrasian Demand

Review of Optimization Basics

It is convenient to introduce some notation for this type of problems. I will write this as. max u (x 1 ; x 2 ) subj. to. p 1 x 1 + p 2 x 2 m ;

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag.

Convex Optimization Overview (cnt d)

September Math Course: First Order Derivative

z = f (x; y) f (x ; y ) f (x; y) f (x; y )

Econ Slides from Lecture 14

Mathematical Foundations -1- Constrained Optimization. Constrained Optimization. An intuitive approach 2. First Order Conditions (FOC) 7

Consider this problem. A person s utility function depends on consumption and leisure. Of his market time (the complement to leisure), h t

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION

The Fundamental Welfare Theorems

Chapter 2 THE MATHEMATICS OF OPTIMIZATION. Copyright 2005 by South-Western, a division of Thomson Learning. All rights reserved.

Microeconomic Theory -1- Introduction

1 Objective. 2 Constrained optimization. 2.1 Utility maximization. Dieter Balkenborg Department of Economics

Week 7: The Consumer (Malinvaud, Chapter 2 and 4) / Consumer November Theory 1, 2015 (Jehle and 1 / Reny, 32

Topic 2. Consumption/Saving and Productivity shocks

ECON 186 Class Notes: Optimization Part 2

Finite Dimensional Optimization Part I: The KKT Theorem 1

Lecture 7: Weak Duality

The Envelope Theorem

Maximum Value Functions and the Envelope Theorem

Chiang/Wainwright: Fundamental Methods of Mathematical Economics

Review of Optimization Methods

Introduction to Machine Learning Spring 2018 Note Duality. 1.1 Primal and Dual Problem

EC487 Advanced Microeconomics, Part I: Lecture 2

Tutorial 3: Optimisation

Recitation #2 (August 31st, 2018)

Macroeconomic Theory and Analysis Suggested Solution for Midterm 1

Microeconomic Theory. Microeconomic Theory. Everyday Economics. The Course:

ECONOMIC OPTIMALITY. Date: October 10, 2005.

Homework 3 Suggested Answers

Gi en Demand for Several Goods

OPTIMIZATION THEORY IN A NUTSHELL Daniel McFadden, 1990, 2003

November 13, 2018 MAT186 Week 8 Justin Ko

Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers

In economics, the amount of a good x demanded is a function of the price of that good. In other words,

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

Reexamination of the A J effect. Abstract

Hicksian Demand and Expenditure Function Duality, Slutsky Equation

UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013

Chapter 4. Maximum Theorem, Implicit Function Theorem and Envelope Theorem

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Econ 508-A FINITE DIMENSIONAL OPTIMIZATION - NECESSARY CONDITIONS. Carmen Astorne-Figari Washington University in St. Louis.

Basic mathematics of economic models. 3. Maximization

BEEM103 UNIVERSITY OF EXETER. BUSINESS School. January 2009 Mock Exam, Part A. OPTIMIZATION TECHNIQUES FOR ECONOMISTS solutions

Dynamic Macroeconomic Theory Notes. David L. Kelly. Department of Economics University of Miami Box Coral Gables, FL

c 2007 Je rey A. Miron

Nonlinear Programming (NLP)

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko

Test code: ME I/ME II, 2004 Syllabus for ME I. Matrix Algebra: Matrices and Vectors, Matrix Operations, Determinants,

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

Economics 101. Lecture 2 - The Walrasian Model and Consumer Choice

Traditional vs. Correct Transversality Conditions, plus answers to problem set due 1/28/99

Mathematics Review For GSB 420. Instructor: Tim Opiela

Recitation 2-09/01/2017 (Solution)

where u is the decision-maker s payoff function over her actions and S is the set of her feasible actions.

ECON 255 Introduction to Mathematical Economics

Monetary Economics: Solutions Problem Set 1

Lecture 6: Discrete-Time Dynamic Optimization

Problem Set #2: Overlapping Generations Models Suggested Solutions - Q2 revised

Sometimes the domains X and Z will be the same, so this might be written:

Economics 101 Spring 2001 Section 4 - Hallam Problem Set #5

Economics 101A (Lecture 3) Stefano DellaVigna

Marginal Functions and Approximation

First Variation of a Functional

Solow Growth Model. Michael Bar. February 28, Introduction Some facts about modern growth Questions... 4

Mathematical Appendix

Competitive Equilibrium and the Welfare Theorems

1 Recursive Competitive Equilibrium

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:

Maximum and Minimum Values

A Basic Course in Real Analysis Prof. P. D. Srivastava Department of Mathematics Indian Institute of Technology, Kharagpur

a. This is an inequality constrained maximization problem. The Lagrangian is µ 1 q 1 = 0 µ 2 q 2 = 0 1 (q 1 ) φ φ c 2 (q 2) p 1 c 1(q 1 ) = 0

Econ 201: Problem Set 3 Answers

Computational Optimization. Constrained Optimization Part 2

Constrained maxima and Lagrangean saddlepoints

Optimization Methods: Optimization using Calculus - Equality constraints 1. Module 2 Lecture Notes 4

Ordering the Extraction of Polluting Nonrenewable Resources

FIN 550 Exam answers. A. Every unconstrained problem has at least one interior solution.

Math 119 Main Points of Discussion

Applications I: consumer theory

Transcription:

The Kuhn-Tucker and Envelope Theorems Peter Ireland EC720.01 - Math for Economists Boston College, Department of Economics Fall 2010 The Kuhn-Tucker and envelope theorems can be used to characterize the solution to a wide range of constrained optimization problems: static or dynamic, under perfect foresight or featuring randomness and uncertainty. In addition, these same two results provide foundations for the work on the maimum principle and dynamic programming that we will do later on. For both of these reasons, the Kuhn-Tucker and envelope theorems provide the starting point for our analysis. Let s consider each in turn, first in fairly general or abstract settings and then applied to some economic eamples. 1 The Kuhn-Tucker Theorem References: Diit, Chapters 2 and 3. Simon-Blume, Chapter 18. Acemoglu, Appendi A. Consider a simple constrained optimization problem: R choice variable F : R R objective function, continuously differentiable c G() constraint, with c R and G : R R, also continuously differentiable. The problem can be stated as: ma F () subject to c G() Copyright c 2010 by Peter Ireland. Redistribution is permitted for educational and research purposes, so long as no changes are made. All copies must be provided free of charge and must include this copyright notice. 1

This problem is simple because it is static and contains no random or stochastic elements that would force decisions to be made under uncertainty. This problem is also simple because it has a single choice variable and a single constraint. All these simplifications will make our statement and proof of the Kuhn-Tucker theorem as clean and intuitive as possible. But the results can be generalized along all of these dimensions and, later, we will work through eamples that do so. Probably the easiest way to solve this problem is via the method of Lagrange multipliers. The mathematical foundations that allow for the application of this method are given to us by Lagrange s Theorem or, in its most general form, the Kuhn-Tucker Theorem. To prove this theorem, begin by defining the Lagrangian: for any R and λ R. L(, λ) = F () + λ[c G()] Theorem (Kuhn-Tucker) Suppose that maimizes F () subject to c G(), where F and G are both continuously differentiable, and suppose that G ( ) 0. Then there eists a value λ of λ such that and λ satisfy the following four conditions: L 1 (, λ ) = F ( ) λ G ( ) = 0, (1) and L 2 (, λ ) = c G( ) 0, (2) λ 0, (3) λ [c G( )] = 0. (4) Proof Consider two possible cases, depending on whether or not the constraint is binding at. Case 1: Nonbinding Constraint. If c > G( ), then let λ = 0. Clearly, (2)-(4) are satisfied, so it only remains to show that (1) must hold. With λ = 0, (1) holds if and only if F ( ) = 0. (5) We can show that (5) must hold using a proof by contradiction. Suppose that instead of (5), it turns out that F ( ) < 0. Then, by the continuity of F and G, there must eist an ε > 0 such that F ( ε) > F ( ) and c > G( ε). 2

But this result contradicts the assumption that maimizes F () subject to c G(). Similarly, if it turns out that F ( ) > 0, then by the continuity of F and G there must eist an ε > 0 such that F ( + ε) > F ( ) and c > G( + ε), But, again, this result contradicts the assumption that maimizes F () subject to c G(). This establishes that (5) must hold, completing the proof for case 1. Case 2: Binding Constraint. Notes: If c = G( ), then let λ = F ( )/G ( ). This is possible, given the assumption that G ( ) 0. Clearly, (1), (2), and (4) are satisfied, so it only remains to show that (3) must hold. With λ = F ( )/G ( ), (3) holds if and only if F ( )/G ( ) 0. (6) We can show that (6) must hold using a proof by contradiction. Suppose that instead of (6), it turns out that F ( )/G ( ) < 0. One way that this can happen is if F ( ) > 0 and G ( ) < 0. But if these conditions hold, then the continuity of F and G implies the eistence of an ε > 0 such that F ( + ε) > F ( ) and c = G( ) > G( + ε), which contradicts the assumption that maimizes F () subject to c G(). And if, instead, F ( )/G ( ) < 0 because F ( ) < 0 and G ( ) > 0, then the continuity of F and G implies the eistence of an ε > 0 such that F ( ε) > F ( ) and c = G( ) > G( ε), which again contradicts the assumption that maimizes F () subject to c G(). This establishes that (6) must hold, completing the proof for case 2. a) The theorem can be etended to handle cases with more than one choice variable and more than one constraint: see Diit, Simon-Blume, or Acemoglu. The proof for the more general case with n choice variables and m constraints presented in section 19.6 of Simon and Blume s book basically works through the same steps that we took for the simple case above: First, Simon and Blume distinguish between the e binding constraints and the m e nonbinding constraints, just as we did above. 3

Second, they set the multipliers on the nonbinding constraints equal to zero and observe that if all functions entering into the constraints are continuously differentiable, it is possible to make small adjustments to the choice variables away from the optimum so that those m e constraints will continue to be nonbinding. The rest of the analysis considers only adjustments of this kind. Again this echoes our arguments from above. Third, they observe that in the case with multiple binding constraints, our assumption that G ( ) 0 becomes a condition that requires the e n matri formed by differentiating each binding constraint by each choice variable to have maimal rank. A clever application of the implicit function theorem shows that because of this rank condition, it is possible to find a set of Lagrange multipliers for the binding constraints that satisfy the set of n firstorder conditions. This step parallels what we did above, when we solved the FOC for the case with the binding constraint for the value λ. Since, at this point, it is clear that all of the FOC hold as do all of the constraints and complementary slackness conditions, it only remains to show that the multipliers for the binding constraints are all nonnegative. The last steps to prove this parallel our last steps from above, which show that λ 0 for the case of a binding constraint. For the general case, this part of the proof uses another clever application of the implicit function theorem to show that when the choice variables are modified in a way that still satisfies all of the binding constraints, the value of the objective function must decrease. Hence, an epression analogous to our equation (6) must hold, which in turn implies that the multipliers are nonnegative. b) Equations (1)-(4) are necessary conditions: If is a solution to the optimization problem, then there eists a λ such that (1)-(4) must hold. But (1)-(4) are not sufficient conditions: if and λ satisfy (1)-(4), it does not follow automatically that is a solution to the optimization problem. Despite point (b) listed above, the Kuhn-Tucker theorem is etremely useful in practice. Suppose that we are looking for the solution to the constrained optimization problem ma F () subject to c G(). The theorem tells us that if we form the Lagrangian L(, λ) = F () + λ[c G()], then and the associated λ must satisfy the first-order condition (FOC) obtained by differentiating L by and setting the result equal to zero: In addition, we know that must satisfy the constraint: L 1 (, λ ) = F ( ) λ G ( ) = 0, (1) c G( ). (2) 4

We know that the Lagrange multiplier λ must be nonnegative: And finally, we know that the complementary slackness condition λ 0. (3) λ [c G( )] = 0, (4) must hold: If λ > 0, then the constraint must bind; if the constraint does not bind, then λ = 0. In searching for the value of that solves the constrained optimization problem, we only need to consider values of that satisfy (1)-(4). Two pieces of terminology: a) The etra assumption that G ( ) 0 is needed to guarantee the eistence of a multiplier λ satisfying (1)-(4). This etra assumption is called the constraint qualification, and almost always holds in practice. b) Note that (1) is a FOC for, while (2) is like a FOC for λ. In most applications, the second-order conditions (SOC) will imply that maimizes L(, λ), while λ minimizes L(, λ). For this reason, (, λ ) is typically a saddle-point of L(, λ). Thus, in solving the problem in this way, we are using the Lagrangian to turn a constrained optimization problem into an unconstrained optimization problem, where we seek to maimize L(, λ) rather than simply F (). One final note: Our general constraint, c G(), nests as a special case the nonnegativity constraint 0, obtained by setting c = 0 and G() =. So nonnegativity constraints can be introduced into the Lagrangian in the same way as all other constraints. If we consider, for eample, the etended problem ma F () subject to c G() and 0, then we can introduce a second multiplier µ, form the Lagrangian as L(, λ, µ) = F () + λ[c G()] + µ, and write the first order condition for the optimal as L 1 (, λ, µ ) = F ( ) λ G ( ) + µ = 0. (1 ) In addition, analogs to our earlier conditions (2)-(4) must also hold for the second constraint: 0, µ 0, and µ = 0. 5

Kuhn and Tucker s original statement of the theorem, however, does not incorporate nonnegativity constraints into the Lagrangian. Instead, even with the additional nonnegativity constraint 0, they continue to define the Lagrangian as L(, λ) = F () + λ[c G()]. If this case, the first order condition for must be modified to read L 1 (, λ ) = F ( ) λ G ( ) 0, with equality if > 0. (1 ) Of course, in (1 ), µ 0 in general and µ = 0 if > 0. So a close inspection reveals that these two approaches to handling nonnegativity constraints lead in the end to the same results. 2 The Envelope Theorem References: Diit, Chapter 5. Simon-Blume, Chapter 19. Acemoglu, Appendi A. In our discussion of the Kuhn-Tucker theorem, we considered an optimization problem of the form ma F () subject to c G() Now, let s generalize the problem by allowing the functions F and G to depend on a parameter θ R. The problem can now be stated as ma F (, θ) subject to c G(, θ) For this problem, define the maimum value function V : R R as V (θ) = ma F (, θ) subject to c G(, θ) Note that evaluating V requires a two-step procedure: First, given θ, find the value of that solves the constrained optimization problem. Second, substitute this value of, together with the given value of θ, into the objective function to obtain V (θ) = F (, θ) Now suppose that we want to investigate the properties of this function V. Suppose, in particular, that we want to take the derivative of V with respect to its argument θ. 6

As the first step in evaluating V (θ), consider solving the constrained optimization problem for any given value of θ by setting up the Lagrangian L(, λ) = F (, θ) + λ[c G(, θ)] We know from the Kuhn-Tucker theorem that the solution to the optimization problem and the associated value of the multiplier λ must satisfy the complementary slackness condition: λ [c G(, θ)] = 0 Use this last result to rewrite the epression for V as V (θ) = F (, θ) = F (, θ) + λ [c G(, θ)] So suppose that we tried to calculate V (θ) simply by differentiating both sides of this equation with respect to θ: V (θ) = F 2 (, θ) λ G 2 (, θ). But, in principle, this formula may not be correct. The reason is that and λ will themselves depend on the parameter θ, and we must take this dependence into account when differentiating V with respect to θ. However, the envelope theorem tells us that our formula for V (θ) is, in fact, correct. That is, the envelope theorem tells us that we can ignore the dependence of and λ on θ in calculating V (θ). To see why, for any θ, let (θ) denote the solution to the problem: ma F (, θ) subject to c G(, θ), and let λ (θ) be the associated Lagrange multiplier. Theorem (Envelope) Let F and G be continuously differentiable functions of and θ. For any given θ, let (θ) maimize F (, θ) subject to c G(, θ), and let λ (θ) be the value of the associated Lagrange multiplier. Suppose, further, that (θ) and λ (θ) are also continuously differentiable functions, and that the constraint qualification G 1 [ (θ), θ] 0 holds for all values of θ. Then the maimum value function defined by satisfies V (θ) = ma F (, θ) subject to c G(, θ) V (θ) = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ]. (7) Proof The Kuhn-Tucker theorem tells us that for any given value of θ, (θ) and λ (θ) must satisfy L 1 [ (θ), λ (θ)] = F 1 [ (θ), θ] λ (θ)g 1 [ (θ), θ] = 0, (1) and λ (θ){c G[ (θ), θ]} = 0. (4) 7

In light of (4), V (θ) = F [ (θ), θ] = F [ (θ), θ] + λ (θ){c G[ (θ), θ]} Differentiating both sides of this epression with respect to θ yields V (θ) = F 1 [ (θ), θ] (θ) + F 2 [ (θ), θ] +λ (θ){c G[ (θ), θ]} λ (θ)g 1 [ (θ), θ] (θ) λ (θ)g 2 [ (θ), θ] which shows that, in principle, we must take the dependence of and λ on θ into account when calculating V (θ). Note, however, that V (θ) = {F 1 [ (θ), θ] λ (θ)g 1 [ (θ), θ]} (θ) which by (1) reduces to +F 2 [ (θ), θ] + λ (θ){c G[ (θ), θ]} λ (θ)g 2 [ (θ), θ], V (θ) = F 2 [ (θ), θ] + λ (θ){c G[ (θ), θ]} λ (θ)g 2 [ (θ), θ] Thus, it only remains to show that λ (θ){c G[ (θ), θ]} = 0 (8) Clearly, (8) holds for any θ such that the constraint is binding. For θ such that the constraint is not binding, (4) implies that λ (θ) must equal zero. Furthermore, by the continuity of G and, if the constraint does not bind at θ, there eists a ε > 0 such that the constraint does not bind for all θ +ε with ε > ε. Hence, (4) also implies that λ (θ +ε) = 0 for all ε > ε. Using the definition of the derivative λ (θ) = lim ε 0 λ (θ + ε) λ (θ) ε it once again becomes apparent that (8) must hold. 0 = lim ε 0 ε = 0, Thus, V (θ) = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ] as claimed in the theorem. Once again, this theorem is useful because it tells us that we can ignore the dependence of and λ on θ in calculating V (θ). But what is the intuition for why the envelope theorem holds? To obtain some intuition, begin by considering the simpler, unconstrained optimization problem: ma F (, θ), where is the choice variable and θ is the parameter. 8

Associated with this unconstrained problem, define the maimum value function in the same way as before: V (θ) = ma F (, θ). To evaluate V for any given value of θ, use the same two-step procedure as before. First, find the value (θ) that solves the unconstrained maimization problem for that value of θ. Second,substitute that value of back into the objective function to obtain V (θ) = F [ (θ), θ]. Now differentiate both sides of this epression through by θ, carefully taking the dependence of on θ into account: V (θ) = F 1 [ (θ), θ] (θ) + F 2 [ (θ), θ]. But, if (θ) is the value of that maimizes F given θ, we know that (θ) must be a critical value of F : F 1 [ (θ), θ] = 0. Hence, for the unconstrained problem, the envelope theorem implies that V (θ) = F 2 [ (θ), θ], so that, again, we can ignore the dependence of on θ in differentiating the maimum value function. And this result holds not because fails to depend on θ: to the contrary, in fact, will typically depend on θ through the function (θ). Instead, the result holds because since is chosen optimally, (θ) is a critical point of F given θ. Now return to the constrained optimization problem ma F (, θ) subject to c G(, θ) and define the maimum value function as before: V (θ) = ma F (, θ) subject to c G(, θ). The envelope theorem for this constrained problem tells us that we can also ignore the dependence of on θ when differentiating V with respect to θ, but only if we start by adding the complementary slackness condition to the maimized objective function to first obtain V (θ) = F [ (θ), θ] + λ (θ){c G[ (θ), θ]}. In taking this first step, we are actually evaluating the entire Lagrangian at the optimum, instead of just the objective function. We need to take this first step because for the constrained problem, the Kuhn-Tucker condition (1) tells us that (θ) is a critical point, not of the objective function by itself, but of the entire Lagrangian formed by adding the product of the multiplier and the constraint to the objective function. 9

And what gives the envelope theorem its name? The envelope theorem refers to a geometrical presentation of the same result that we ve just worked through. To see where that geometrical interpretation comes from, consider again the simpler, unconstrained optimization problem: ma F (, θ), where is the choice variable and θ is a parameter. Following along with our previous notation, let (θ) denote the solution to this problem for any given value of θ, so that the function (θ) tells us how the optimal choice of depends on the parameter θ. Also, continue to define the maimum value function V in the same way as before: V (θ) = ma F (, θ). Now let θ 1 denote a particular value of θ, and let 1 denote the optimal value of associated with this particular value θ 1. That is, let 1 = (θ 1 ). After substituting this value of 1 into the function F, we can think about how F ( 1, θ) varies as θ varies that is, we can think about F ( 1, θ) as a function of θ, holding 1 fied. In the same way, let θ 2 denote another particular value of θ, with θ 2 > θ 1 let s say. And following the same steps as above, let 2 denote the optimal value of associated with this particular value θ 2, so that 2 = (θ 2 ). Once again, we can hold 2 fied and consider F ( 2, θ) as a function of θ. The geometrical presentation of the envelope theorem can be derived by thinking about the properties of these three functions of θ: V (θ), F ( 1, θ), and F ( 2, θ). One thing that we know about these three functions is that for θ = θ 1 : V (θ 1 ) = F ( 1, θ 1 ) > F ( 2, θ 1 ), where the first equality and the second inequality both follow from the fact that, by definition, 1 maimizes F (, θ 1 ) by choice of. Another thing that we know about these three functions is that for θ = θ 2 : V (θ 2 ) = F ( 2, θ 2 ) > F ( 1, θ 2 ), because again, by definition, 2 maimizes F (, θ 2 ) by choice of. 10

The Envelope Theorem V(!) F( 2,!) F( 1,!)! 1! 2!

On a graph, these relationships imply that: At θ 1, V (θ) coincides with F ( 1, θ), which lies above F ( 2, θ). At θ 2, V (θ) coincides with F ( 2, θ), which lies above F ( 1, θ). And we could find more and more values of V by repeating this procedure for more and more specific values of θ i, i = 1, 2, 3,... In other words: V (θ) traces out the upper envelope of the collection of functions F ( i, θ), formed by holding i = (θ i ) fied and varying θ. Moreover, V (θ) is tangent to each individual function F ( i, θ) at the value θ i of θ for which i is optimal, or equivalently: V (θ) = F 2 [ (θ), θ], which is the same analytical result that we derived earlier for the unconstrained optimization problem. To generalize these arguments so that they apply to the constrained optimization problem ma F (, θ) subject to c G(, θ), simply use the fact that in most cases (where the appropriate second-order conditions hold) the value (θ) that solves the constrained optimization problem for any given value of θ also maimizes the Lagrangian function so that L(, λ, θ) = F (, θ) + λ[c G(, θ)], V (θ) = ma F (, θ) subject to c G(, θ) = ma L(, λ, θ) Now just replace the function F with the function L in working through the arguments from above to conclude that V (θ) = L 3 [ (θ), λ (θ), θ] = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ], which is again the same result that we derived before for the constrained optimization problem. 11

3 Two Eamples 3.1 Utility Maimization A consumer has a utility function defined over consumption of two goods: U(c 1, c 2 ) Prices: p 1 and p 2 Income: I Budget constraint: I p 1 c 1 + p 2 c 2 = G(c 1, c 2 ) The consumer s problem is: ma c 1,c 2 U(c 1, c 2 ) subject to I p 1 c 1 + p 2 c 2 The Kuhn-Tucker theorem tells us that if we set up the Lagrangian: L(c 1, c 2, λ) = U(c 1, c 2 ) + λ(i p 1 c 1 p 2 c 2 ) Then the optimal consumptions c 1 and c 2 and the associated multiplier λ must satisfy the FOC: L 1 (c 1, c 2, λ ) = U 1 (c 1, c 2) λ p 1 = 0 and L 2 (c 1, c 2, λ ) = U 2 (c 1, c 2) λ p 2 = 0 Move the terms with minus signs to the other side, and divide the first of these FOC by the second to obtain U 1 (c 1, c 2) U 2 (c 1, c 2) = p 1, p 2 which is just the familiar condition that says that the optimizing consumer should set the slope of his or her indifference curve, the marginal rate of substitution, equal to the slope of his or her budget constraint, the ratio of prices. Now consider I as one of the model s parameters, and let the functions c 1(I), c 2(I), and λ (I) describe how the optimal choices c 1 and c 2 and the associated value λ of the multiplier depend on I. In addition, define the maimum value function as The Kuhn-Tucker theorem tells us that and hence V (I) = ma c 1,c 2 U(c 1, c 2 ) subject to I p 1 c 1 + p 2 c 2 λ (I)[I p 1 c 1(I) p 2 c 2(I)] = 0 V (I) = U[c 1(I), c 2(I)] = U[c 1(I), c 2(I)] + λ (I)[I p 1 c 1(I) p 2 c 2(I)]. 12

The envelope theorem tells us that we can ignore the dependence of c 1 and c 2 on I in calculating V (I) = λ (I), which gives us an interpretation of the multiplier λ as the marginal utility of income. 3.2 Cost Minimization The Kuhn-Tucker and envelope conditions can also be used to study constrained minimization problems. Consider a firm that produces output y using capital k and labor l, according to the technology described by f(k, l) y. r = rental rate for capital w = wage rate Suppose that the firm takes its output y as given, and chooses inputs k and l to minimize costs. Then the firm solves If we set up the Lagrangian as min rk + wl subject to f(k, l) y k,l L(k, l, λ) = rk + wl λ[f(k, l) y], where the term involving the multiplier λ is subtracted rather than added in the case of a minimization problem, the Kuhn-Tucker conditions (1)-(4) continue to apply, eactly as before. Thus, according to the Kuhn-Tucker theorem, the optimal choices k and l and the associated multiplier λ must satisfy the FOC: L 1 (k, l, λ ) = r λ f 1 (k, l ) = 0 (9) and L 2 (k, l, λ ) = w λ f 2 (k, l ) = 0 (10) Move the terms with minus signs over to the other side, and divide the first FOC by the second to obtain f 1 (k, l ) f 2 (k, l ) = r w, which is another familiar condition that says that the optimizing firm chooses factor inputs so that the marginal rate of substitution between inputs in production equals the ratio of factor prices. 13

Now suppose that the constraint binds, as it usually will: y = f(k, l ) (11) Then (9)-(11) represent 3 equations that determine the three unknowns k, l, and λ as functions of the model s parameters r, w, and y. In particular, we can think of the functions k = k (r, w, y) and l = l (r, w, y) as demand curves for capital and labor: strictly speaking, they are conditional (on y) factor demand functions. Now define the minimum cost function as C(r, w, y) = min rk + wl subject to f(k, l) y k,l = rk (r, w, y) + wl (r, w, y) = rk (r, w, y) + wl (r, w, y) λ (r, w, y){f[k (r, w, y), l (r, w, y)] y} The envelope theorem tells us that in calculating the derivatives of the cost function, we can ignore the dependence of k, l, and λ on r, w, and y. Hence: C 1 (r, w, y) = k (r, w, y), C 2 (r, w, y) = l (r, w, y), and C 3 (r, w, y) = λ (r, w, y). The first two of these equations are statements of Shephard s lemma; they tell us that the derivatives of the cost function with respect to factor prices coincide with the conditional factor demand curves. The third equation gives us an interpretation of the multiplier λ as a measure of the marginal cost of increasing output. Thus, our two eamples illustrate how we can apply the Kuhn-Tucker and envelope theorems in specific economic problems. The two eamples also show how, in the contet of specific economic problems, it is often possible to attach an economic interpretation to the multiplier λ. 14