The Kuhn-Tucker and Envelope Theorems

Similar documents
The Kuhn-Tucker and Envelope Theorems

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

CHAPTER 1-2: SHADOW PRICES

The Necessity of the Transversality Condition at Infinity: A (Very) Special Case

Optimization. A first course on mathematics for economists

CHAPTER 3: OPTIMIZATION

Problem Set 2 Solutions

INTRODUCTORY MATHEMATICS FOR ECONOMICS MSCS. LECTURE 3: MULTIVARIABLE FUNCTIONS AND CONSTRAINED OPTIMIZATION. HUW DAVID DIXON CARDIFF BUSINESS SCHOOL.

The Kuhn-Tucker Problem

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013

Nonlinear Programming and the Kuhn-Tucker Conditions

Mathematical Foundations -1- Constrained Optimization. Constrained Optimization. An intuitive approach 2. First Order Conditions (FOC) 7

The Envelope Theorem

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag.

How to Characterize Solutions to Constrained Optimization Problems

September Math Course: First Order Derivative

Midterm Exam - Solutions

Optimization Paul Schrimpf September 12, 2018

1.4 FOUNDATIONS OF CONSTRAINED OPTIMIZATION

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

Properties of Walrasian Demand

Chiang/Wainwright: Fundamental Methods of Mathematical Economics

Constrained Optimization

Econ Slides from Lecture 14

Convex Optimization Overview (cnt d)

Finite Dimensional Optimization Part I: The KKT Theorem 1

It is convenient to introduce some notation for this type of problems. I will write this as. max u (x 1 ; x 2 ) subj. to. p 1 x 1 + p 2 x 2 m ;

Sometimes the domains X and Z will be the same, so this might be written:

First Variation of a Functional

Review of Optimization Methods

Monetary Economics: Solutions Problem Set 1

In economics, the amount of a good x demanded is a function of the price of that good. In other words,

z = f (x; y) f (x ; y ) f (x; y) f (x; y )

Review of Optimization Basics

Maximum Value Functions and the Envelope Theorem

E 600 Chapter 4: Optimization

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION

Microeconomic Theory -1- Introduction

ECONOMIC OPTIMALITY. Date: October 10, 2005.

Topic 2. Consumption/Saving and Productivity shocks

Gi en Demand for Several Goods

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

1 Objective. 2 Constrained optimization. 2.1 Utility maximization. Dieter Balkenborg Department of Economics

Optimization Methods: Optimization using Calculus - Equality constraints 1. Module 2 Lecture Notes 4

ECON FINANCIAL ECONOMICS

Consider this problem. A person s utility function depends on consumption and leisure. Of his market time (the complement to leisure), h t

Test code: ME I/ME II, 2004 Syllabus for ME I. Matrix Algebra: Matrices and Vectors, Matrix Operations, Determinants,

Introduction to Machine Learning Spring 2018 Note Duality. 1.1 Primal and Dual Problem

Mathematical Appendix

Economics 205 Exercises

Econ 508-A FINITE DIMENSIONAL OPTIMIZATION - NECESSARY CONDITIONS. Carmen Astorne-Figari Washington University in St. Louis.

Chapter 2 THE MATHEMATICS OF OPTIMIZATION. Copyright 2005 by South-Western, a division of Thomson Learning. All rights reserved.

The Fundamental Welfare Theorems

Macroeconomic Theory and Analysis Suggested Solution for Midterm 1

Solow Growth Model. Michael Bar. February 28, Introduction Some facts about modern growth Questions... 4

where u is the decision-maker s payoff function over her actions and S is the set of her feasible actions.

Econ Slides from Lecture 10

MATH 1325 Business Calculus Guided Notes

EC487 Advanced Microeconomics, Part I: Lecture 2

Lecture 6: Discrete-Time Dynamic Optimization

Urban Economics. Yves Zenou Research Institute of Industrial Economics. July 17, 2006

Reexamination of the A J effect. Abstract

Lecture 7: Weak Duality

Hicksian Demand and Expenditure Function Duality, Slutsky Equation

Equilibrium in a Production Economy

Computational Optimization. Constrained Optimization Part 2

Constrained maxima and Lagrangean saddlepoints

ECON 186 Class Notes: Optimization Part 2

Chapter 4. Maximum Theorem, Implicit Function Theorem and Envelope Theorem

12.1 The Extrema of a Function

c 2007 Je rey A. Miron

Mathematical Preliminaries

Mathematics Review For GSB 420. Instructor: Tim Opiela

Universidad Carlos III de Madrid

Constrained optimization.

Adding Production to the Theory

Economics 101. Lecture 2 - The Walrasian Model and Consumer Choice

Basic mathematics of economic models. 3. Maximization

Homework 3 Suggested Answers

Intro to Nonlinear Optimization

Final Exam - Math Camp August 27, 2014

Math 119 Main Points of Discussion

18.303: Introduction to Green s functions and operator inverses

Dynamic Macroeconomic Theory Notes. David L. Kelly. Department of Economics University of Miami Box Coral Gables, FL

Microeconomic Theory. Microeconomic Theory. Everyday Economics. The Course:

Nonlinear Programming (NLP)

Getting to page 31 in Galí (2008)

Hidden information. Principal s payoff: π (e) w,

A Basic Course in Real Analysis Prof. P. D. Srivastava Department of Mathematics Indian Institute of Technology, Kharagpur

Dynamic Problem Set 1 Solutions

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Week 7: The Consumer (Malinvaud, Chapter 2 and 4) / Consumer November Theory 1, 2015 (Jehle and 1 / Reny, 32

Mathematical Foundations II

Lecture Notes on Monotone Comparative Statics and Producer Theory

EconS 301. Math Review. Math Concepts

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko

Part I Analysis in Economics

1 Recursive Competitive Equilibrium

Transcription:

The Kuhn-Tucker and Envelope Theorems Peter Ireland ECON 77200 - Math for Economists Boston College, Department of Economics Fall 207 The Kuhn-Tucker and envelope theorems can be used to characterize the solution to a wide range of constrained optimization problems: static or dynamic, and under perfect foresight or featuring randomness and uncertainty. In addition, these same two results provide foundations for the work on the maimum principle and dynamic programming that we will do later on. For both of these reasons, the Kuhn-Tucker and envelope theorems provide the starting point for our analysis. Let s consider each in turn, first in fairly general or abstract settings and then applied to some economic eamples. The Kuhn-Tucker Theorem References: Diit, Chapters 2 and 3. Simon-Blume, Chapters 8 and 9. Acemoglu, Appendi A. Consider a simple constrained optimization problem: R choice variable F : R R objective function, continuously differentiable c G() constraint, with c R and G : R R, also continuously differentiable. The problem can be stated as: ma F () subject to c G() These lecture notes are licensed under a Creative Commons Attribution-NonCommerical-ShareAlike 4.0 International (CC BY-NC-SA 4.0) License. http://creativecommons.org/licenses/by-nc-sa/4.0/

This problem is simple because it is static and contains no random or stochastic elements that would force decisions to be made under uncertainty. This problem is also simple because it has a single choice variable and a single constraint. All these simplifications will make our statement and proof of the Kuhn-Tucker theorem as clean and intuitive as possible. But the results can be generalized along all of these dimensions and, throughout the semester, we will work through eamples that do so. Probably the easiest way to solve this problem is via the method of Lagrange multipliers. The mathematical foundations that allow for the application of this method are given to us by Lagrange s Theorem or, in its most general form, the Kuhn-Tucker Theorem. To prove this theorem, begin by defining the Lagrangian: for any R and λ R. L(, λ) = F () + λ[c G()] Theorem (Kuhn-Tucker) Suppose that maimizes F () subject to c G(), where F and G are both continuously differentiable, and suppose that G ( ) 0. Then there eists a value λ of λ such that and λ satisfy the following four conditions: L (, λ ) = F ( ) λ G ( ) = 0, () and L 2 (, λ ) = c G( ) 0, (2) λ 0, (3) λ [c G( )] = 0. (4) Proof Consider two possible cases, depending on whether or not the constraint is binding at. Case : Nonbinding Constraint. If c > G( ), then let λ = 0. Clearly, (2)-(4) are satisfied, so it only remains to show that () must hold. With λ = 0, () holds if and only if F ( ) = 0. (5) We can show that (5) must hold using a proof by contradiction. Suppose that instead of (5), it turns out that F ( ) < 0. Then, by the continuity of F and G, there must eist an ε > 0 such that F ( ε) > F ( ) and c > G( ε). 2

But this result contradicts the assumption that maimizes F () subject to c G(). Similarly, if it turns out that F ( ) > 0, then by the continuity of F and G there must eist an ε > 0 such that F ( + ε) > F ( ) and c > G( + ε), But, again, this result contradicts the assumption that maimizes F () subject to c G(). This establishes that (5) must hold, completing the proof for case. Case 2: Binding Constraint. Notes: If c = G( ), then let λ = F ( )/G ( ). This is possible, given the assumption that G ( ) 0. Clearly, (), (2), and (4) are satisfied, so it only remains to show that (3) must hold. With λ = F ( )/G ( ), (3) holds if and only if F ( )/G ( ) 0. (6) We can show that (6) must hold using a proof by contradiction. Suppose that instead of (6), it turns out that F ( )/G ( ) < 0. One way that this can happen is if F ( ) > 0 and G ( ) < 0. But if these conditions hold, then the continuity of F and G implies the eistence of an ε > 0 such that F ( + ε) > F ( ) and c = G( ) > G( + ε), which contradicts the assumption that maimizes F () subject to c G(). And if, instead, F ( )/G ( ) < 0 because F ( ) < 0 and G ( ) > 0, then the continuity of F and G implies the eistence of an ε > 0 such that F ( ε) > F ( ) and c = G( ) > G( ε), which again contradicts the assumption that maimizes F () subject to c G(). This establishes that (6) must hold, completing the proof for case 2. a) The theorem can be etended to handle cases with more than one choice variable and more than one constraint: see Diit, Simon-Blume, Acemoglu, or section 4. of the notes below. b) Equations ()-(4) are necessary conditions: If is a solution to the optimization problem, then there eists a λ such that ()-(4) must hold. But ()-(4) are not sufficient conditions: if and λ satisfy ()-(4), it does not follow automatically that is a solution to the optimization problem. 3

As an eample that illustrates point (b), consider the problem: With the Lagrangian defined as the Kuhn-Tucker conditions are and ma e 2 subject to. + 2 L(, λ) = e 2 + λ( ), + 2 L (, λ ) = e + 4 [ + ( ) 2 ] 2 λ = 0, L 2 (, λ ) = 0, λ 0, λ ( ) = 0. These conditions are satisfied when = and λ = e +, which corresponds to the solution to the problem, but they are also satisfied when = 0.2205 and λ = 0, which corresponds instead to the solution to the minimization problem min e 2 subject to. + 2 But despite point (b) listed above, the Kuhn-Tucker theorem is etremely useful in practice. Suppose that we are looking for the solution to the constrained optimization problem ma F () subject to c G(). The theorem tells us that if we form the Lagrangian L(, λ) = F () + λ[c G()], then and the associated λ must satisfy the first-order condition (FOC) obtained by differentiating L by and setting the result equal to zero: In addition, we know that must satisfy the constraint: L (, λ ) = F ( ) λ G ( ) = 0, () We know that the Lagrange multiplier λ must be nonnegative: c G( ). (2) λ 0. (3) 4

And finally, we know that the complementary slackness condition λ [c G( )] = 0, (4) must hold: If λ > 0, then the constraint must bind; if the constraint does not bind, then λ = 0. In searching for the value of that solves the constrained optimization problem, we only need to consider values of that satisfy ()-(4). In the eample from above, for instance, it is straightforward to compare the values of the objective function when = and = 0.2205 to conclude that the solution to the optimization problem is =. Two pieces of terminology: a) The etra assumption that G ( ) 0 is needed to guarantee the eistence of a multiplier λ satisfying ()-(4). This etra assumption is called the constraint qualification, and almost always holds in practice. b) Note that () is a FOC for, while (2) is like a FOC for λ. In many economic applications, where F () is concave and G() is conve, maimizes L(, λ), while λ minimizes L(, λ). For this reason, (, λ ) is typically a saddle-point of L(, λ). To see that the saddle-point property characterizes the solution to concave programming problems, consider ma F () subject to c G(), where F is concave, G is conve, and the constraint qualification holds. Under these assumptions, the Kuhn-Tucker conditions ()-(4) are both necessary and sufficient for the value of that solves the problem and the associated value λ of the Lagrange multiplier (see Simon and Blume s Theorem 2.22, pp.532-533). Consider first the Lagrangian, viewed as a function of with λ fied at λ : L(, λ ) = F () + λ [c G()]. Since F is concave, G is conve, and λ > 0, L(, λ ) is a concave function of, and is therefore maimized when L(, λ ) = F ( ) λ G ( ) holds. But this first-order condition for maimizing L with respect to is the same as the first-order condition () for maimizing F () with respect to subject to the constraint c G(). Now consider the Lagrangian, viewed as a function of λ with fied at : L(, λ) = F ( ) + λ[c G( )]. When c > G( ), L(, λ) is minimized subject to λ > 0 with λ = 0. And when c = G( ), L(, λ) is minimized subject to λ > 0 with λ = F ( )/G ( ) so long as the constraint qualification holds. 5

Thus, point (b) gives us some intuition about how the Kuhn-Tucker theorem works. It uses the Lagrangian to turn a constrained optimization problem into an unconstrained problem, where the solution for is a critical point of L(, λ ) rather than a critical point of F (). And when the etra assumptions made in concave programming are adopted, is not only a critical point of L(, λ ), it also maimizes L(, λ ). As an eample illustrating the saddle-point property, consider the problem ma ln() subject to. Since the logarithmic objective function is strictly increasing, we know that = is the solution. Forming the Lagrangian and taking the first-order condition L(, λ) = ln() + λ( ) L (, λ ) = λ = 0 reveals that the associated value of λ is λ =. Substituting this value for λ back into the Lagrangian L(, λ ) = ln() +, we can see that L(, λ ) is maimized with =. Note, however, that in the general case where F () need not be concave, will always be a critical point of the Lagrangian that is, it will satisfy the first-order condition () but need not maimize L(, λ). An eample of how this can happen is given by Diit s eample 7.2 (p.03). Consider the problem ma e subject to. Since the eponential objective function is strictly increasing, we know again that this problem is solved with =. Forming the Lagrangian and taking the first-order condition L (, λ ) = e λ = 0, shows that the associated value of λ is e. But λ = e implies that L(, λ ) = e + λ ( ) = e + e( ) = e e + e, and since e e grows without bound as either or, = does not maimize the Lagrangian given λ. Hence, (, λ ) is not a saddle-point of L, since the objective function from the original problem is conve, not concave. One final note: Our general constraint, c G(), nests as a special case the nonnegativity constraint 0, obtained by setting c = 0 and G() =. 6

So nonnegativity constraints can be introduced into the Lagrangian in the same way as all other constraints. If we consider, for eample, the etended problem ma F () subject to c G() and 0, then we can introduce a second multiplier µ, form the Lagrangian as L(, λ, µ) = F () + λ[c G()] + µ, and write the first order condition for the optimal as L (, λ, µ ) = F ( ) λ G ( ) + µ = 0. ( ) In addition, analogs to our earlier conditions (2)-(4) must also hold for the second constraint: 0, µ 0, and µ = 0. Kuhn and Tucker s original statement of the theorem, however, does not incorporate nonnegativity constraints into the Lagrangian. Instead, even with the additional nonnegativity constraint 0, they continue to define the Lagrangian as L(, λ) = F () + λ[c G()]. If this case, the first order condition for must be modified to read L (, λ ) = F ( ) λ G ( ) 0, with equality if > 0. ( ) Of course, in ( ), µ 0 in general and µ = 0 if > 0. So a close inspection reveals that these two approaches to handling nonnegativity constraints lead in the end to the same results. 2 The Envelope Theorem References: Diit, Chapter 5. Simon-Blume, Chapter 9. Acemoglu, Appendi A. In our discussion of the Kuhn-Tucker theorem, we considered an optimization problem of the form ma F () subject to c G() Now, let s generalize the problem by allowing the functions F and G to depend on a parameter θ R. The problem can now be stated as ma F (, θ) subject to c G(, θ) 7

For this problem, define the maimum value function V : R R as V (θ) = ma F (, θ) subject to c G(, θ) Note that evaluating V requires a two-step procedure: First, given θ, find the value of that solves the constrained optimization problem. Second, substitute this value of, together with the given value of θ, into the objective function to obtain V (θ) = F (, θ) Now suppose that we want to investigate the properties of this function V. Suppose, in particular, that we want to take the derivative of V with respect to its argument θ. As the first step in evaluating V (θ), consider solving the constrained optimization problem for any given value of θ by setting up the Lagrangian L(, λ) = F (, θ) + λ[c G(, θ)] We know from the Kuhn-Tucker theorem that the solution to the optimization problem and the associated value of the multiplier λ must satisfy the complementary slackness condition: λ [c G(, θ)] = 0 Use this last result to rewrite the epression for V as V (θ) = F (, θ) = F (, θ) + λ [c G(, θ)] So suppose that we tried to calculate V (θ) simply by differentiating both sides of this equation with respect to θ: V (θ) = F 2 (, θ) λ G 2 (, θ). But, in principle, this formula may not be correct. The reason is that and λ will themselves depend on the parameter θ, and we must take this dependence into account when differentiating V with respect to θ. However, the envelope theorem tells us that our formula for V (θ) is, in fact, correct. That is, the envelope theorem tells us that we can ignore the dependence of and λ on θ in calculating V (θ). To see why, for any θ, let (θ) denote the solution to the problem: ma F (, θ) subject to c G(, θ), and let λ (θ) be the associated Lagrange multiplier. 8

Theorem (Envelope) Let F and G be continuously differentiable functions of and θ. For any given θ, let (θ) maimize F (, θ) subject to c G(, θ), and let λ (θ) be the associated value of the Lagrange multiplier. Suppose, further, that (θ) and λ (θ) are also continuously differentiable functions, and that the constraint qualification G [ (θ), θ] 0 holds for all values of θ. Then the maimum value function defined by V (θ) = ma F (, θ) subject to c G(, θ) satisfies V (θ) = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ]. (7) Proof The Kuhn-Tucker theorem tells us that for any given value of θ, (θ) and λ (θ) must satisfy L [ (θ), λ (θ)] = F [ (θ), θ] λ (θ)g [ (θ), θ] = 0, () and In light of (4), λ (θ){c G[ (θ), θ]} = 0. (4) V (θ) = F [ (θ), θ] = F [ (θ), θ] + λ (θ){c G[ (θ), θ]} Differentiating both sides of this epression with respect to θ yields V (θ) = F [ (θ), θ] (θ) + F 2 [ (θ), θ] +λ (θ){c G[ (θ), θ]} λ (θ)g [ (θ), θ] (θ) λ (θ)g 2 [ (θ), θ] which shows that, in principle, we must take the dependence of and λ on θ into account when calculating V (θ). Note, however, that V (θ) = {F [ (θ), θ] λ (θ)g [ (θ), θ]} (θ) which by () reduces to +F 2 [ (θ), θ] + λ (θ){c G[ (θ), θ]} λ (θ)g 2 [ (θ), θ], V (θ) = F 2 [ (θ), θ] + λ (θ){c G[ (θ), θ]} λ (θ)g 2 [ (θ), θ] Thus, it only remains to show that λ (θ){c G[ (θ), θ]} = 0 (8) Clearly, (8) holds for any θ such that the constraint is binding. 9

For θ such that the constraint is not binding, (4) implies that λ (θ) must equal zero. Furthermore, by the continuity of G and, if the constraint does not bind at θ, there eists an ε > 0 such that the constraint does not bind for all θ + ε with ε > ε. Hence, (4) also implies that λ (θ + ε) = 0 for all ε > ε. Using the definition of the derivative Thus, λ λ (θ + ε) λ (θ) (θ) = lim ε 0 ε it once again becomes apparent that (8) must hold. as claimed in the theorem. 0 = lim ε 0 ε = 0, V (θ) = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ] Once again, this theorem is useful because it tells us that we can ignore the dependence of and λ on θ in calculating V (θ). And once again, the theorem can be etended to apply in more general settings: see Diit, Simon-Blume, Acemoglu, or section 4.2 of the notes below. It is worth noting that the assumptions required by the envelope theorem are more restrictive than those required by the Kuhn-Tucker theorem. For eample, the statement of the Kuhn-Tucker theorem makes reference to the value of that solves the constrained optimization problem, thereby assuming implicitly that a solution to the problem eists. But the envelope theorem requires that be depicted as a function of θ, implying that the solution must not only eist but be unique as well. Further, the solution must vary smoothly with θ, so that and λ can be written as continuously differentiable functions of the parameter. Although most statements of the envelope theorem assume directly that (θ) and λ (θ) are continuously differentiable, it is also possible to impose restrictions on F (, θ) and G(, θ) that guarantee this. Assume, for eample, that the constraint always binds at the optimum, so that by the Kuhn-Tucker theorem, (θ) and λ (θ) must satisfy and L [ (θ), λ (θ)] = F [ (θ), θ] λ (θ)g [ (θ), θ] = 0 L 2 [ (θ), λ (θ)] = λ (θ){c G[ (θ), θ]} = 0. If F and G are two times continuously differentiable in and θ, and if the matri of second derivatives of the Lagrangian [ ] L [ (θ), λ (θ)] L 2 [ (θ), λ (θ)] L 2 [ (θ), λ (θ)] L 22 [ (θ), λ (θ)] is nonsingular, then the implicit function theorem (Simon and Blume, Theorem 5.7, pp.355-356) will imply that (θ) and λ (θ) eist and are continuously differentiable. 0

But what is the intuition for why the envelope theorem holds? To obtain some intuition, begin by considering the simpler, unconstrained optimization problem: ma F (, θ), where is the choice variable and θ is the parameter. Associated with this unconstrained problem, define the maimum value function in the same way as before: V (θ) = ma F (, θ). To evaluate V for any given value of θ, use the same two-step procedure as before. First, find the value (θ) that solves the unconstrained maimization problem for that value of θ. Second, substitute that value of back into the objective function to obtain V (θ) = F [ (θ), θ]. Now differentiate both sides of this epression through by θ, carefully taking the dependence of on θ into account: V (θ) = F [ (θ), θ] (θ) + F 2 [ (θ), θ]. But, if (θ) is the value of that maimizes F given θ, we know that (θ) must be a critical point of F : F [ (θ), θ] = 0. Hence, for the unconstrained problem, the envelope theorem implies that V (θ) = F 2 [ (θ), θ], so that, again, we can ignore the dependence of on θ in differentiating the maimum value function. And this result holds not because fails to depend on θ: to the contrary, in fact, will typically depend on θ through the function (θ). Instead, the result holds because since is chosen optimally, (θ) is a critical point of F given θ. Now return to the constrained optimization problem ma F (, θ) subject to c G(, θ) and define the maimum value function as before: V (θ) = ma F (, θ) subject to c G(, θ). The envelope theorem for this constrained problem tells us that we can also ignore the dependence of on θ when differentiating V with respect to θ, but only if we start by adding the complementary slackness condition to the maimized objective function to first obtain V (θ) = F [ (θ), θ] + λ (θ){c G[ (θ), θ]}.

In taking this first step, we are actually evaluating the entire Lagrangian at the optimum, instead of just the objective function. We need to take this first step because for the constrained problem, the Kuhn-Tucker condition () tells us that (θ) is a critical point, not of the objective function by itself, but of the entire Lagrangian formed by adding the product of the multiplier and the constraint to the objective function. And what gives the envelope theorem its name? The envelope theorem refers to a geometrical presentation of the same result that we ve just worked through. To see where that geometrical interpretation comes from, consider again the simpler, unconstrained optimization problem: ma F (, θ), where is the choice variable and θ is a parameter. Following along with our previous notation, let (θ) denote the solution to this problem for any given value of θ, so that the function (θ) tells us how the optimal choice of depends on the parameter θ. Also, continue to define the maimum value function V in the same way as before: V (θ) = ma F (, θ). Now let θ denote a particular value of θ, and let denote the optimal value of associated with this particular value θ. That is, let = (θ ). After substituting this value of into the function F, we can think about how F (, θ) varies as θ varies that is, we can think about F (, θ) as a function of θ, holding fied. In the same way, let θ 2 denote another particular value of θ, with θ 2 > θ let s say. And following the same steps as above, let 2 denote the optimal value of associated with this particular value θ 2, so that 2 = (θ 2 ). Once again, we can hold 2 fied and consider F ( 2, θ) as a function of θ. The geometrical presentation of the envelope theorem can be derived by thinking about the properties of these three functions of θ: V (θ), F (, θ), and F ( 2, θ). One thing that we know about these three functions is that for θ = θ : V (θ ) = F (, θ ) > F ( 2, θ ), where the first equality and the second inequality both follow from the fact that, by definition, maimizes F (, θ ) by choice of. 2

Another thing that we know about these three functions is that for θ = θ 2 : V (θ 2 ) = F ( 2, θ 2 ) > F (, θ 2 ), because again, by definition, 2 maimizes F (, θ 2 ) by choice of. On a graph, these relationships imply that: At θ, V (θ) coincides with F (, θ), which lies above F ( 2, θ). At θ 2, V (θ) coincides with F ( 2, θ), which lies above F (, θ). And we could find more and more values of V by repeating this procedure for more and more specific values of θ i, i =, 2, 3,... In other words: V (θ) traces out the upper envelope of the collection of functions F ( i, θ), formed by holding i = (θ i ) fied and varying θ. Moreover, V (θ) is tangent to each individual function F ( i, θ) at the value θ i of θ for which i is optimal, or equivalently: If, for eample, then V (θ) = F 2 [ (θ), θ], which is the same analytical result that we derived earlier for the unconstrained optimization problem. F (, θ) = ( θ) 2 + θ 2 = 2 + 2θ, V (θ) = ma ( θ) 2 + θ 2 = θ 2, since, in this case, (θ) = θ for all values of θ. The figure below sets θ = 2 and θ 2 = 7; hence = 2 and 2 = 7, then plots and to show how and F (, θ) = 4 + 4θ, F ( 2, θ) = 49 + 4θ, V (θ) = θ 2 V (θ ) = F (, θ ) > F ( 2, θ ) at θ = 2, V (θ 2 ) = F ( 2, θ 2 ) > F (, θ 2 ) at θ 2 = 7, and how, more generally, V (θ) traces out the upper envelope of the family of functions F ( i, θ), where each i maimizes F (, θ) for some value θ i of θ. 3

To generalize these arguments so that they apply to the constrained optimization problem ma F (, θ) subject to c G(, θ), simply use the fact that in many cases (as when F is concave and G is conve) the value (θ) that solves the constrained optimization problem for any given value of θ also maimizes the Lagrangian function so that L(, λ, θ) = F (, θ) + λ[c G(, θ)], V (θ) = ma F (, θ) subject to c G(, θ) = ma L(, λ, θ) Now just replace the function F with the function L in working through the arguments from above to conclude that V (θ) = L 3 [ (θ), λ (θ), θ] = F 2 [ (θ), θ] λ (θ)g 2 [ (θ), θ], which is again the same result that we derived before for the constrained optimization problem. Note that in the figure, the maimum value function V (θ) is conve, with a strictly positive second derivative, whereas the functions F (, θ) and F ( 2, θ) are both linear. In general, it can be shown that there is a sense in which the maimum value function V (θ) will always be more conve or less concave than each member of the family of functions F ( i, θ). 4

To see this, consider the two functions V (θ) and F (, θ) as they were defined for the general unconstrained problem ma F (, θ). Second order Taylor approimations imply that for values of θ sufficiently close to θ : and However, V (θ) V (θ ) + V (θ )(θ θ ) + (/2)V (θ )(θ θ ) 2 F (, θ) F (, θ ) + F 2 (, θ )(θ θ ) + (/2)F 22 (, θ )(θ θ ) 2. since = (θ ), and V (θ ) = F (, θ ), V (θ ) = F 2 (, θ ), by the envelope theorem itself. Since the definition of V implies that V (θ) F (, θ) for all values of θ, the Taylor approimations imply V (θ ) F 22 (, θ ), so that, more specifically, the second derivative of the maimum value function will always be larger than the second derivatives of the functions F ( i, θ). Finally, to begin getting a feel for the usefulness of the envelope theorem, consider a preliminary economic eample. Suppose that a firm hires n workers, paying each the competitive real wage w (real wage, so that we don t have to consider separately the price of output), in order to produce y units of output according to the production function n α y, where α lies between zero and one: 0 < α <. For simplicity, let s depict the firm as solving the unconstrained optimization problem with first-order condition that leads to the labor demand curve ma n α wn, n α(n ) α w = 0 n (w) = ( ) α w α. α 5

Net, define the maimum value function so that V (w) = ma n α wn, n V (w) = [n (w)] α wn (w). The envelope theorem immediately implies that V (w) = n (w) = ( ) α w α. α Suppose instead we substitute our earlier solution for n (w) into the epression for V (w): V (w) = [n (w)] α wn (w) [ ( ) ] α ( ) α = w α α w w α α α ( ) α ( ) α α = w α α α w α α α [ ( ) α ( ) ] α α = w α α. α α Now differentiate with respect to w and simplify to get V (w) = = = ( α α ) [ ( α ( α ) ( α α ( ) α w α, α ) α ( ) α α α ) α ( α ) ] w α w α eactly as we found, much more quickly, using the envelope theorem. In terms of its economics, this eample shows that when the firm faces a higher wage, there will be two effects on its profits. The direct effect: it must pay each of its workers a higher wage, so profits fall by n (w). The indirect effect: it can try to mitigate the direct effect by hiring fewer workers. The envelope theorem says that since the firm has already chosen n optimally, the first-order effect on profits of adjusting this decision in response to an arbitrarily small change in the wage is zero. Only the direct effect remains: V (w) = n (w). 6

3 Three Eamples 3. Utility Maimization A consumer has a utility function defined over consumption of two goods: U(c, c 2 ) Prices: p and p 2 Income: I Budget constraint: I p c + p 2 c 2 = G(c, c 2 ) The consumer s problem is: ma c,c 2 U(c, c 2 ) subject to I p c + p 2 c 2 The Kuhn-Tucker theorem tells us that if we set up the Lagrangian: L(c, c 2, λ) = U(c, c 2 ) + λ(i p c p 2 c 2 ) Then the optimal consumptions c and c 2 and the associated multiplier λ must satisfy the FOC: L (c, c 2, λ ) = U (c, c 2) λ p = 0 and L 2 (c, c 2, λ ) = U 2 (c, c 2) λ p 2 = 0 Move the terms with minus signs to the other side, and divide the first of these FOC by the second to obtain U (c, c 2) U 2 (c, c 2) = p, p 2 which is just the familiar condition that says that the optimizing consumer should set the slope of his or her indifference curve, the marginal rate of substitution, equal to the slope of his or her budget constraint, the ratio of prices. Now consider I as one of the model s parameters, and let the functions c (I), c 2(I), and λ (I) describe how the optimal choices c and c 2 and the associated value λ of the multiplier depend on I. In addition, define the maimum value function as The Kuhn-Tucker theorem tells us that and hence V (I) = ma c,c 2 U(c, c 2 ) subject to I p c + p 2 c 2 λ (I)[I p c (I) p 2 c 2(I)] = 0 V (I) = U[c (I), c 2(I)] = U[c (I), c 2(I)] + λ (I)[I p c (I) p 2 c 2(I)]. 7

The envelope theorem tells us that we can ignore the dependence of c and c 2 on I in calculating V (I) = λ (I), which gives us an interpretation of the multiplier λ as the marginal utility of income. 3.2 Cost Minimization The Kuhn-Tucker and envelope conditions can also be used to study constrained minimization problems. Consider a firm that produces output y using capital k and labor l, according to the technology described by f(k, l) y. r = rental rate for capital w = wage rate Suppose that the firm takes its output y as given, and chooses inputs k and l to minimize costs. Then the firm solves If we set up the Lagrangian as min rk + wl subject to f(k, l) y k,l L(k, l, λ) = rk + wl λ[f(k, l) y], where the term involving the multiplier λ is subtracted rather than added in the case of a minimization problem, the Kuhn-Tucker conditions ()-(4) continue to apply, eactly as before. Thus, according to the Kuhn-Tucker theorem, the optimal choices k and l and the associated multiplier λ must satisfy the FOC: L (k, l, λ ) = r λ f (k, l ) = 0 (9) and L 2 (k, l, λ ) = w λ f 2 (k, l ) = 0 (0) Move the terms with minus signs over to the other side, and divide the first FOC by the second to obtain f (k, l ) f 2 (k, l ) = r w, which is another familiar condition that says that the optimizing firm chooses factor inputs so that the marginal rate of substitution between inputs in production equals the ratio of factor prices. 8

Now suppose that the constraint binds, as it usually will: y = f(k, l ) () Then (9)-() represent 3 equations that determine the three unknowns k, l, and λ as functions of the model s parameters r, w, and y. In particular, we can think of the functions k = k (r, w, y) and l = l (r, w, y) as demand curves for capital and labor: strictly speaking, they are conditional (on y) factor demand functions. Now define the minimum cost function as C(r, w, y) = min rk + wl subject to f(k, l) y k,l = rk (r, w, y) + wl (r, w, y) = rk (r, w, y) + wl (r, w, y) λ (r, w, y){f[k (r, w, y), l (r, w, y)] y} The envelope theorem tells us that in calculating the derivatives of the cost function, we can ignore the dependence of k, l, and λ on r, w, and y. Hence: C (r, w, y) = k (r, w, y), C 2 (r, w, y) = l (r, w, y), and C 3 (r, w, y) = λ (r, w, y). The first two of these equations are statements of Shephard s lemma; they tell us that the derivatives of the cost function with respect to factor prices coincide with the conditional factor demand curves. The third equation gives us an interpretation of the multiplier λ as a measure of the marginal cost of increasing output. Thus, our two eamples illustrate how we can apply the Kuhn-Tucker and envelope theorems in specific economic problems. The two eamples also show how, in the contet of specific economic problems, it is often possible to attach an economic interpretation to the multiplier λ. 9

3.3 Le Chatelier s Principle For the net eample, etend the previous cost minimization problem by introducing a third input: m = materials input q = price of materials Define the minimum cost function C(r, w, q, y) = min rk + wl + qm subject to f(k, l, m) y. k,l,m Then Shephard s lemma implies and C (r, w, q, y) = k (r, w, q, y), C 2 (r, w, q, y) = l (r, w, q, y), C 3 (r, w, q, y) = m (r, w, q, y), where k (r, w, q, y), l (r, w, q, y), and m (r, w, q, y) are the conditional factor demand curves and the envelope theorem also implies that C 4 (r, w, q, y) = λ (r, w, q, y), so that the Lagrange multiplier λ (r, w, q, y) is again a measure of marginal cost. Net, let s use Shephard s lemma to deduce another property of the conditional factor demand curves. Since C (r, w, q, y) = k (r, w, q, y), it follows that And since it follows that C 2 (r, w, q, y) = k 2(r, w, q, y). C 2 (r, w, q, y) = l (r, w, q, y), C 2 (r, w, q, y) = l (r, w, q, y). But, since the symmetry of the second partial derivatives of C implies C 2 (r, w, q, y) = C 2 (r, w, q, y), Shephard s lemma can be viewed as having, as a corollary, the reciprocity condition for conditional factor demand curves: k 2(r, w, q, y) = l (r, w, q, y), or, equivalently, k (r, w, q, y) w 20 = l (r, w, q, y). r

Now consider a short-run version of the firm s problem, treating the capital stock k as fied at some pre-determined level k: min l,m r k + wl + qm subject to f( k, l, m) y. With the Lagrangian for the short-run problem defined as the first-order conditions are L(l, m, µ) = r k + wl + qm µ[f( k, l, m) y], w µ s f 2 ( k, l s, m s ) = 0 and and the binding constraint is q µ s f 3 ( k, l s, m s ) = 0 f( k, l s, m s ) y = 0. Interestingly, none of these optimality conditions depends on the rental rate r for capital. Hence, we can solve for short-run conditional factor demand curves of the form and l s = l s (w, q, y, k) m s = m s (w, q, y, k), and use another application of the envelope theorem, this time to the short-run problem, to interpret µ s = µ s (w, q, y, k) as a measure of short-run marginal cost. Le Chatelier s principle says that labor demand should be more responsive to a change in the wage rate w in the long run than in the short run since, intuitively, in the long run the firm has the chance to substitute capital for labor in response to that change in the wage. To show the sense in which this conjecture proves true, suppose that the fied, short-run value of k just happens to equal the optimal value k that would have been chosen anyway in the long run: k = k (r, w, q, y). In this case, a comparison of the Kuhn-Tucker conditions for the short-run problem with those for the long-run problem confirms that the firm will choose a value of l s in the short run equal to the long-run value l : l (r, w, q, y) = l s [w, q, y, k (r, w, q, y)]. 2

Differentiate both sides of this epression with respect to w to obtain l 2(r, w, q, y) = l s [w, q, y, k (r, w, q, y)] + l s 4[w, q, y, k (r, w, y)]k 2(r, w, q, y). The term on the left-hand side of this epression measures the long-run responsiveness of labor demand to a change in the wage; the first term on the right-hand side of this epression measures the short-run responsiveness of labor demand to a change in the wage. The second-order conditions for the long-run and short-run problems will generally imply that under typical conditions, both of these terms are negative: when the wage goes up, labor demand falls. But, is the long-run response more negative? This depends on the sign of the second term on the right-hand side, which in turn depends on l s 4, measuring the response of short-run labor demand to a change in k, and k 2, measuring the response of long-run capital demand to a change in w. The signs of both terms would seem to be ambiguous, dependent in some loose sense on whether, in the long run, the capital and labor inputs are complements or substitutes. Yet, as it turns out, something more definite can be said. Return once more to l (r, w, q, y) = l s [w, q, y, k (r, w, y)], but this time differentiate both sides with respect to r to obtain l (r, w, q, y) = l s 4[w, q, y, k (r, w, y)]k (r, w, q, y). Rearrange this epression so that it reads l s 4[w, q, y, k (r, w, q, y)] = l (r, w, q, y) k (r, w, q, y), and substitute this result into the previous one to get l 2(r, w, q, y) = l s [w, q, y, k (r, w, q, y)] + l s 4[w, q, y, k (r, w, q, y)]k 2(r, w, q, y). l2(r, w, q, y) = l[w, s q, y, k (r, w, q, y)] + l (r, w, q, y)k2(r, w, q, y). k(r, w, q, y) Now it is clear that the second term must be negative. This is because the epression in the numerator is, by the reciprocity condition implied by Shephard s lemma, equal to the perfect square [l(r, w, q, y)] 2 ; meanwhile, the term in the denominator will be negative, as implied by the second-order conditions to the long-run problem: the optimal capital stock falls when the rental rate for capital rises. Thus, it must be that l 2(r, w, q, y) l s [w, q, y, k (r, w, y)] and, since both of these terms are negative, l (r, w, q, y) w l s (w, q, y, k ) w, which is the relationship we were after: labor demand is more responsive to a change in the wage in the long run than in the short run. Note, however, that this is a result that only holds locally, when k is sufficiently close to k (r, w, q, y). 22

4 Generalizing the Basic Results 4. The Kuhn-Tucker Theorem Our simple version of the Kuhn-Tucker theorem applies to a problem with only one choice variable and one constraint. Section 9.6 of Simon and Blume s book develops a proof for the more general case, with n choice variables and m constraints. Their proof makes repeated, clever use of the implicit function theorem, which makes the arguments surprisingly short but also works to obscure some of the intuition provided by the analysis of the simplest case. Nevertheless, having gained the intuition the intuition from working through the simple case, it is useful to see how the result etends. Simon and Blume (Chapter 5) and Acemoglu (Appendi A) both present fairly general statements of the implicit function theorem. The special case or application of their results that we will need works as follows. Consider a system of n equations, involving n endogenous variables y, y 2,..., y n and n eogenous variables c, c 2,... c n : H (y, y 2,..., y n ) = c, H 2 (y, y 2,..., y n ) = c 2, H m (y, y 2,..., y n ) = c n. Now suppose that for a specific set of values c, c 2,... c n for the eogenous variables, all the equations in the system are satisfied with the endogenous variables set equal to y, y 2,..., y n, so that H (y, y 2,..., y n) = c, H 2 (y, y 2,..., y n) = c 2, H n (y, y 2,..., y n) = c n. Assume that each function H i, i =,..., n, is continuously differentiable and that the n n matri of derivatives H / y H / y n H 2 / y H 2 / y n..... H n / y H n / y n is nonsingular at y, y 2,..., y n... 23

Then there eist continuously differentiable functions y (c, c 2,..., c n ), y 2 (c, c 2,..., c n ),. y n (c, c 2..., c n ), defined in an open subset C of R n containing (c, c 2,..., c n), such that H (y (c, c 2,..., c n ), y 2 (c, c 2,..., c n ),..., y n (c, c 2,..., c n )) = c, H 2 (y (c, c 2,..., c n ), y 2 (c, c 2,..., c n ),..., y n (c, c 2,..., c n )) = c 2, H n (y (c, c 2,..., c n ), y 2 (c, c 2,..., c n ),..., y n (c, c 2,..., c n )) = c n. for all (c, c 2,..., c n ) C. With this result in hand, consider the following generalized version of the Kuhn-Tucker theorem we proved earlier. Let there be n choice variables,, 2,..., n. The objective function F : R n R is continuously differentiable, as are the m functions G j : R n R, j =, 2,..., m that enter into the constraints where c j R for all j =, 2,..., m. The problem can be stated as: c j G j (, 2,..., n ), ma F (, 2,..., n ) subject to c j G j (, 2,..., n ) for all j =, 2,..., m., 2,..., n Note that, typically, m n will have to hold so that there is a set of values for the choice variables that satisfy all of the constraints. To define the Lagrangian, introduce the multipliers λ j, j =, 2,..., m, one for each constraint. Then m L(, 2,..., n, λ, λ 2,..., λ m ) = F (, 2,..., n ) + λ j [c j G j (, 2,..., n )]. Theorem (Kuhn-Tucker) Suppose that, 2,..., n maimize F (, 2,..., n ) subject to c j G j (, 2,..., n ) for all j =, 2,..., m, where F and the G j s are all continuously differentiable. Suppose (without loss of generality) that the first m m constraints bind at the optimum and that the remaining m m 0 constraints are nonbinding, and assume that the m n matri of derivatives G, (, 2,..., n)... G,n (, 2,..., n) G 2, (, 2,..., n)... G 2,n (, 2,..., n)....., (2) G m, (, 2,..., n)... G m,n (, 2,..., n) 24 j=.

where G j,i = G j / i, has rank m. Then there eist values λ, λ 2,..., λ m that, together with, 2,..., n, satisfy: L i (, 2,..., n, λ, λ 2,..., λ n) = F i (, 2,..., n) m λ jg j,i (, 2,..., n) = 0 j= (3) for i =, 2,..., n, L n+j (, 2,..., n, λ, λ 2,..., λ n) = c j G j (, 2,..., n) 0, (4) for j =, 2,..., m, for j =, 2,..., m, and λ j 0, (5) λ j[c j G j (, 2,..., n)] = 0, (6) for j =, 2,..., m. Proof To begin, set the multipliers λ m+, λ m+2,..., λ m associated with the nonbinding contraints equal to zero. Since each of the functions G j, j = m +, m + 2,..., m, is continuously differentiable, sufficiently small adjustments in the choice variables can be made without violating these m m constraints or causing any of them to become binding. Net, note that the m + n matri F (, 2,..., n)... F n (, 2,..., n) G, (, 2,..., n)... G,n (, 2,..., n) G 2, (, 2,..., n)... G 2,n (, 2,..., n). (7)..... G m, (, 2,..., n)... G m,n (, 2,..., n) must have rank m < m +. To see why, consider the system of equations F (, 2,..., n ) = y G (, 2,..., n ) = c G 2 (, 2,..., n ) = c 2 G m (, 2,..., n ) = c m. With y set equal to the maimized value of the objective function, y = F (, 2,..., n),. 25

each of these m + equations holds when the functions are evaluated at, 2,..., n. In this case, the implicit function theorem implies that it should be possible to adjust the values of m+ of the choice variables so to find a new set of values, 2,..., n such that F (, 2,..., n ) = y + ε G (, 2,..., n ) = c G 2 (, 2,..., n ) = c 2 G m (, 2,..., n ) = c m. for a strictly positive but sufficiently small value of ε. But this contradicts the assumption that, 2,..., n solves the constrained optimization problem. Since the matri in (7) has rank m < m +, its m + rows must be linearly dependent. Hence, there eist scalars α 0, α,... α m, at least one of which is nonzero, such that F (, 2,..., n) 0. = α 0. 0 F n (, 2,..., n) (8) G, (, 2,..., n) G m, (, 2,..., n) + α. +... + α m.. G,n (, 2,..., n) G m,n (, 2,..., n) Moreover, in (8), α 0 0, since otherwise, the matri in (2) would have rank less than m. Thus, for j =, 2,..., m, set λ j = α j /α 0. With these settings for λ, λ 2,..., λ m, plus the settings λ m+ = λ m+2 = λ m = 0 chosen earlier, (8) implies that (3) must hold for all i =, 2,..., n. Clearly, (4) and (6) are satisfied for all j =, 2,..., m, and (5) holds for all j = m +, m + 2,..., m. So it only remains to show that (5) holds for j =, 2,..., m. To see that these last conditions must hold, consider the system of equations G (, 2,..., n ) = c δ. G 2 (, 2,..., n ) = c 2 G m (, 2,..., n ) = c m,. (9) where δ 0. These equations hold, with δ = 0, at, 2,..., n. And since the matri in (2) has rank m, the implicit function theorem implies that there are functions (δ), 2 (δ),..., n (δ) such that the same equations hold for all sufficiently small values of δ. 26

Since c δ c, the choices (δ), 2 (δ),..., n (δ) satisfy all of the constraints from the original optimization problem. And since, by assumption, (0) =, 2 (0) = 2,..., n (0) = n maimizes the objective function subject to the constraints, it must be that df ( (δ), 2 (δ),..., n (δ)) dδ = δ=0 n F i (, 2,..., n) i(0) 0. (20) In addition, the equations in (9) implicitly defining (δ), 2 (δ),..., n (δ) imply dg ( (δ), 2 (δ),..., n (δ)) n dδ = G,i (, 2,..., n) i(0) = (2) δ=0 and for j = 2, 3,..., m. dg j ( (δ), 2 (δ),..., n (δ)) dδ = δ=0 Putting all these results together, (3) implies 0 = F i (, 2,..., n) i= i= n G j,i (, 2,..., n) i(0) = 0 (22) i= m λ jg j,i (, 2,..., n). j= for all i =, 2,..., n. Multiplying each of these equations by i(0) and summing over all i yields n n m 0 = F i (, 2,..., n) i(0) λ jg j,i (, 2,..., n) i(0), or 0 = i= n F i (, 2,..., n) i(0) i= i= m j= or, since λ j = 0 for j = m +, m + 2,..., m, 0 = n F i (, 2,..., n) i(0) i= λ j m λ j In light of (2) and (22), this last equation simplifies to n 0 = F i (, 2,..., n) i(0) + λ. And hence, in light of (20), i= Analogous arguments show that j= λ 0. λ j 0 for j = 2, 3,..., m as well, completing the proof. 27 j= [ n ] G j,i (, 2,..., n) i(0), i= [ n ] G j,i (, 2,..., n) i(0). i=

4.2 The Envelope Theorem Proving a generalized version of the envelope theorem requires no new ideas, just repeated applications of the previous ones. Consider, again, the constrained optimization problem with n choice variables and m constraints: ma F (, 2,..., n ) subject to c j G j (, 2,..., n ) for all j =, 2,..., m., 2,..., n Now etend this problem by allowing the functions F and G j, j =, 2,..., m, to depend on a parameter θ R: ma, 2,..., n F (, 2,..., n, θ) subject to c j G j (, 2,..., n, θ) for all j =, 2,..., m. Just as before, define the maimum value function V : R R as V (θ) = ma, 2,..., n F (, 2,..., n, θ) subject to c j G j (, 2,..., n, θ) for all j =, 2,..., m. Note that V is still a function of the single parameter θ, since the n choice variables are optimized out. Put another way, evaluating V requires the same two-step procedure as before: First, given θ, find the values (θ), 2(θ),..., n(θ) that solve the constrained optimization problem. Second, substitute these values (θ), 2(θ),..., n(θ), together with the given value of θ, into the objective function to obtain V (θ) = F ( (θ), 2(θ),..., n(θ), θ). And just as before, the envelope theorem tells us that we can calculate the derivative V (θ) of the maimum value function while ignoring the dependence of, 2,..., n and λ, λ 2,..., λ m on θ, provided we invoke the complementary slackness conditions (6) to add the sum of all of the multipliers times all of the constraints to the objective function before differentiating through by θ. Theorem (Envelope) Let F and G j, j =, 2,..., m, be continuously differentiable functions of, 2,..., n and θ. For any value of θ, let (θ), 2(θ),..., n(θ) maimize F (, 2,..., n, θ) subject to c j G j (, 2,..., n, θ) for all j =, 2,..., m, and let λ (θ), λ 2(θ),..., λ m(θ) be the associated values of the Lagrange multipliers. Suppose, further, that (θ), 2(θ),..., n(θ) and λ (θ), λ 2(θ),..., λ m(θ) are all continuously differentiable functions, and that the m(θ) m matri of derivatives G, ( (θ), 2(θ),..., n(θ), θ)... G,n ( (θ), 2(θ),..., n(θ), θ) G 2, ( (θ), 2(θ),..., n(θ), θ)... G 2,n ( (θ), 2(θ),..., n(θ), θ)..... G m(θ), ( (θ), 2(θ),..., n(θ), θ)... G m(θ),n ( (θ), 2(θ),..., n(θ), θ) 28

associated with the m(θ) m binding constraints has rank m(θ) for each value of θ. Then the maimum value function defined by satisfies V (θ) = ma, 2,..., n F (, 2,..., n, θ) subject to c j G j (, 2,..., n, θ) for all j =, 2,..., m V (θ) = F n+ ( (θ), 2(θ),..., n(θ), θ) m λ j(θ)g j,n+ ( (θ), 2(θ),..., n(θ), θ). j= (23) Proof The Kuhn-Tucker theorem implies that for any given value of θ, F i ( (θ), 2(θ),..., n(θ), θ) for i =, 2,..., n, and m λ j(θ)g j,i ( (θ), 2(θ),..., n(θ), θ) = 0 (3) j= λ j(θ)[c j G j ( (θ), 2(θ),..., n(θ), θ)] = 0, (6) for j =, 2,..., m must hold. In light of (6), m V (θ) = F ( (θ), 2(θ),..., n(θ), θ) + λ j(θ)[c j G j ( (θ), 2(θ),..., n(θ), θ)]. j= Differentiating both sides of this epression by θ yields V (θ) = n F i ( (θ), 2(θ),..., n(θ), θ) (θ) i= +F n+ ( (θ), 2(θ),..., n(θ), θ) m + λ j (θ)[c j G j ( (θ), 2(θ),..., n(θ), θ)] j= n i= m λ j(θ)g j,i ( (θ), 2(θ),..., n(θ), θ) (θ) j= m λ j(θ)g j,n+ ( (θ), 2(θ),..., n(θ), θ). j= which shows that, in principle, we must take the dependence of (θ), 2(θ),..., n(θ) and λ (θ), λ 2(θ),..., λ m(θ) on θ into account when calculating V (θ). 29

Note, however, that (3) implies that the sums in the first and fourth lines of this last epression together equal zero. Hence, to show that (23) holds, it only remains to show that m λ j (θ)[c j G j ( (θ), 2(θ),..., n(θ), θ)] = 0 and this is true if j= for all j =, 2,..., m. λ j (θ)[c j G j ( (θ), 2(θ),..., n(θ), θ)] = 0 (24) Clearly, (24) holds for any θ such that constraint j is binding. For θ such that constraint j is not binding, (6) implies that λ j(θ) = 0. Furthermore, by the continuity of G j and i (θ), i =, 2,..., n, if constraint j does not bind at θ, there eists an ε > 0 such that constraint j does not bind for all θ + ε with ε > ε. Hence, λ j (θ) = lim ε 0 λ j(θ + ε) λ j(θ) ε 0 = lim ε 0 ε = 0, and once again it becomes apparent that (24) must hold. Hence, (23) must hold as well. 30