University of California, Davis Department of Agricultural and Resource Economics ARE 252 Lecture Notes 2 Quirino Paris

Similar documents
Nonlinear Programming and the Kuhn-Tucker Conditions

Mathematical Foundations -1- Constrained Optimization. Constrained Optimization. An intuitive approach 2. First Order Conditions (FOC) 7

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

Constrained Optimization

Convex Optimization & Lagrange Duality

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Summary Notes on Maximization

The Monopolist. The Pure Monopolist with symmetric D matrix

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

More on Lagrange multipliers

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

Generalization to inequality constrained problem. Maximize

Optimization. A first course on mathematics for economists

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

Numerical Optimization

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Outline. Roadmap for the NPP segment: 1 Preliminaries: role of convexity. 2 Existence of a solution

CONSTRAINED OPTIMALITY CRITERIA

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Lecture 18: Optimization Programming

+ 5x 2. = x x. + x 2. Transform the original system into a system x 2 = x x 1. = x 1

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Constrained maxima and Lagrangean saddlepoints

Symmetric Lagrange function

Convex Optimization M2

Economic Foundations of Symmetric Programming

Constrained Optimization and Lagrangian Duality

ICS-E4030 Kernel Methods in Machine Learning

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Introduction to Machine Learning Prof. Sudeshna Sarkar Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

CHAPTER 1-2: SHADOW PRICES

1. f(β) 0 (that is, β is a feasible point for the constraints)

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

FIN 550 Practice Exam Answers. A. Linear programs typically have interior solutions.

Finite Dimensional Optimization Part I: The KKT Theorem 1

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

The Dual of the Maximum Likelihood Method

2.098/6.255/ Optimization Methods Practice True/False Questions

Lecture 4: Optimization. Maximizing a function of a single variable

Microeconomics I. September, c Leopold Sögner

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

Stochastic Programming Math Review and MultiPeriod Models

Sharpening the Karush-John optimality conditions

OPTIMISATION /09 EXAM PREPARATION GUIDELINES

Duality Theory of Constrained Optimization

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Mathematical Economics. Lecture Notes (in extracts)

5. Duality. Lagrangian

LECTURE 7 Support vector machines

4TE3/6TE3. Algorithms for. Continuous Optimization

Econ 508-A FINITE DIMENSIONAL OPTIMIZATION - NECESSARY CONDITIONS. Carmen Astorne-Figari Washington University in St. Louis.

Lagrange Relaxation and Duality

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

Roles of Convexity in Optimization Theory. Efor, T. E and Nshi C. E

Optimality Conditions for Constrained Optimization

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control

The Karush-Kuhn-Tucker (KKT) conditions

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES

CONVEX FUNCTIONS AND OPTIMIZATION TECHINIQUES A THESIS SUBMITTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF

Jørgen Tind, Department of Statistics and Operations Research, University of Copenhagen, Universitetsparken 5, 2100 Copenhagen O, Denmark.

The Fundamental Welfare Theorems

Linear and Combinatorial Optimization

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Constrained Optimization

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Chap 2. Optimality conditions

Lecture 7: Convex Optimizations

Optimization Theory. Lectures 4-6

KKT Examples. Stanley B. Gershwin Massachusetts Institute of Technology

Linear programming: Theory

Lecture Notes on Support Vector Machine

Convex Optimization Lecture 6: KKT Conditions, and applications

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

1.3 The Indirect Utility Function

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION

Convex Optimization Overview (cnt d)

Tutorial on Convex Optimization: Part II

EE/AA 578, Univ of Washington, Fall Duality

Support Vector Machines

DEPARTMENT OF STATISTICS AND OPERATIONS RESEARCH OPERATIONS RESEARCH DETERMINISTIC QUALIFYING EXAMINATION. Part I: Short Questions

The Kuhn-Tucker Problem

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

Convex Programs. Carlo Tomasi. December 4, 2018

Finite Dimensional Optimization Part III: Convex Optimization 1

Linear & nonlinear classifiers

Mathematical Economics: Lecture 16

Lecture: Duality of LP, SOCP and SDP

CS-E4830 Kernel Methods in Machine Learning

Lecture: Duality.

Econ Slides from Lecture 14

CO 250 Final Exam Guide

Introduction to Support Vector Machines

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

Optimality, Duality, Complementarity for Constrained Optimization

Transcription:

University of California, Davis Department of Agricultural and Resource Economics ARE 5 Lecture Notes Quirino Paris Karush-Kuhn-Tucker conditions................................................. page Specification of a nonlinear programming problem NPP................................. Review of the traditional Lagrange method............................................. Conversion of the NPP into a traditional Lagrange problem................................ KKT conditions in vector notation.................................................... 5 Geometric interpretation of KKT conditions............................................ 6 Constraint qualification............................................................ 7 Matching the KKT conditions with the structure of the Economic Equilibrium................. 8 Karush-Kuhn-Tucker conditions The formulation of mathematical programming problems began in 939 with the publication of the Master of Science thesis by William Karush (W. Karush (939. "Minima of Functions of Several Variables with Inequalities as Side Constraints". M.Sc. Dissertation. Dept. of Mathematics, Univ. of Chicago, Chicago, Illinois.. This work was ignored for a long time to the point that H.W. Kuhn and A.W. Tucker published their 95 article without citing the seminal work of Karush (H. W. Kuhn and A. W. Tucker, Nonlinear Programming, Proc. Second Berkeley Symp. on Math. Statist. and Prob. (Univ. of Calif. Press, 95, 48-49. Many authors still cite only Kuhn and Tucker when discussing Kuhn-Tucker conditions. The omission of Karush s name is unacceptable. His name belongs to the Karush-Kuhn-Tucker theory of nonlinear programming KKT conditions, for short. Specification of a nonlinear programming problem Most of the time, in this course, we will be dealing with a primal nonlinear programming problem (NPP like NPP max f (x x,x,...,x n subject to g (x b g (x b... g m (x b m x j, j =,...,n f (x is a differentiable concave (quasiconcave function called the objective function (to be optimized. The other relations are differentiable convex (quasiconvex functions called constraints (including the non-negativities of the endogenous variables x j, j =,...,n. As we will see soon, to solve the nonlinear problem stated above we will have to state also a constraint qualification to rule out strange behavior of the constraints.

Question: How should the first-order necessary conditions (FONC of the NPP be stated? The answer constitutes the KKT conditions. To understand the FONC of the NPP we will transform the NPP into a problem subject to equations and unrestricted variables in order to be able to apply the traditional Lagrange method. Review of the traditional Lagrange method (See also chapter, Symmetric Programming textbook Only one constraint is sufficient. For differentiable functions, max f (x x,x,...,x n subject to g(x = b. The Lagrange function is stated as L = f (x + λ[b g(x ] f g with FONC = λ, j =,...,n x j x j x j = b g(x,..., x n. λ The traditional Lagrange problem involves only equations and has no restricted variables. The solution of the FONC will provide the equilibrium values of the endogenous variables, x * j, j =,...,n and λ *. *,λ* f ( x * Recall that, at equilibrium, L(x = = λ * and λ * b b may be either a positive or a negative value f / x j depending on the derivatives λ * = g/ x j. Lagrange multipliers are interpreted (always as shadow prices of the corresponding constraint. Figure illustrates the shape of a Lagrange function. Figure. The graph of Lagrange function. y is the Lagrange multiplier Conversion of the NPP into a traditional Lagrange problem (See also chapter 3, Symmetric Programming textbook To make it simple, we will deal with only two constraints and two endogenous variables. NPP, max f (x x,x subject to

g (x b g (x b x x Let us hypothesize that the above NPP, represents the problem of maximizing total revenue, TR = f (x, subject to the technology, g (x b, g (x b, of transforming the supply of two inputs, b, b, into two outputs, x. To convert the NPP, into a traditional Lagrange problem (with only equations and unrestricted variables, add slack variables, s h, h =,...,4, to the inequalities to transform them into equations and unrestricted variables: NPP, max f (x x,x subject to g (x + s = b λ g (x + s = b λ x s 3 b 3 λ 3 b 4 λ 4 x s 4 In principle, in this NPP,, there are no inequalities and restricted variables. Therefore, we can apply the traditional Lagrange method in a straightforward way: the Lagrange function is L = f (x + λ [b s g (x ]+ λ [b s g (x ] + λ 3 (b 3 + s 3 x + λ 4 (b 4 + s 4 x with FONC f g g. = λ λ λ 3 g g. = f λ λ λ 4 x x x x 3. = λ s s 4. = λ s s 5. = λ 3 s 3 s 3 6. = λ 4 s 4 s 4 3

7. = b s g (x λ 8. = b s g (x λ 9. = b 3 + s 3 x λ 3 0. = b 4 + s 4 x λ 4 In the programming literature, the KKT conditions are presented as a set of inequalities and a set complementary slackness relations. To achieve that specification it is sufficient to eliminate the slack variables introduced above and discuss the remaining relations using the structure of the Economic Equilibrium problem discussed in Lecture Notes. Beginning with FONC 7 (and similarly for 8 l = b g (x = s λ Note that FONC 7 is interpreted as S D using the Economic Equilibrium structure. From FONC 9 (and similarly for 0 l = s3 = x + (0 b 3 x λ 3 Note that FONC 9 is interpreted as Q using the Economic Equilibrium structure. Furthermore, multiplying FONC 7 by λ (and similarly for FONC 8 λ = λ [b s g (x ] = λ [b g (x ] λ because of FONC 3 ( λ s (and similarly FONC 4 for FONC 8. Note that this result is interpreted as the complementary slackness condition P(S D of the Economic Equilibrium structure. Consider now FONC and rewrite it as (similarly for FONC l f g g λ = = λ 3 " x λ #%%%$%%% & MR 0 MR MC = λ 3 Therefore, the Lagrange multiplier λ 3 is interpreted as the (negative Opportunity Cost of producing (or not producing output x. Since λ 3 is non-positive according to the Economic Equilibrium structure the elimination of ( λ 3 from FONC (and similarly for FONC results in the inequality f g g λ = 0 x λ Furthermore, note that from FONC 9 ( b 3 0 and FONC 5 (and similarly from FONC 0 and 6 MC x = s 3 λ 3 x = λ 3 s 3 4

and, therefore, by multiplying FONC by x (and similarly FONC by x and using FONC 5 f g g f g g x = x λ λ λ 3 = x λ λ which corresponds to the complementary slackness condition Q(MR MC of the Economic Equilibrium structure. There remains to sign the Lagrange multipliers λ, λ. Since, by hypothesis, f ( is the total revenue function and b, b are supplies of inputs, an increase of input supply will not decrease the total revenue and, therefore, f f = λ, = λ. b b We have derived all the six (6 sets of KKT conditions which match the six (6 sets of relations of the Economic Equilibrium structure. Except that it is a little difficult to list them all clearly as they were developed in this simplified problem. For this reason, we restate now the KKT conditions using vector notation. KKT condition in vector notation Let the problem to solve be stated as the maximization of a differentiable and concave (quasi-concave function f (x subject to a differentiable and convex (quasi-convex vector function g(x such as max f (x = f (x,..., x n x subject to g (x,..., x n b g(x b!!! g m (x,..., x n b m x The dimension of the vector x is (n while the dimension of the vector b is (m with m < n simply as an arbitrary choice. The Lagrange function is and KKT conditions are L = f (x + λ [b g(x] f g. = λ 0 x x x f g Dual relations. x = x λ x x x 3. x 5

4. = b g(x λ Primal relations 5. λ = λ [b g(x] λ 6. λ Geometric interpretation of KKT conditions In order to be able to graph the relevant KKT conditions we return to the example with constraints and endogenous variables. The relevant KKT conditions are assumed to be binding f g g λ = x λ f g g λ = x x x x λ It is convenient to rewrite these KKT conditions in gradient notation. Let x * be the optimal point. Then * λ L x (x * = f x (x * g x (x * g x (x * * λ which corresponds to * f x (x * = g x (x * * λ + g x (x * λ * * with λ and λ. This means that the gradient of the objective function evaluated at the optimal point x * represents the main diagonal of the parallelogram constructed using the gradients of the constraints as a basis (mathematical ruler. Figure illustrates this fact. Figure. NPP and KKT conditions 6

In figure, D. The parallelogram expresses the KKT conditions with the gradient of the objective function being the main diagonal of the parallelogram constructed on the gradients of the constraints multiplied by positive scalars λ and λ. If the gradient of the objective function, f x (x *, were to fall outside the cone generated by the two gradients of the binding constraints, g x (x *, g x (x *, KKT conditions would be violated. (In a space of two dimensions, a cone is an arrangement of two vectors that form an angle less than 80 degrees. Constraint Qualification In solving a nonlinear programming problem by KKT theory there is the necessity of ruling out a peculiar arrangement of the constraints graph in the neighborhood of the optimal point. This assumption is called constraint qualification. It says that, if the optimal point happens to be in the neighborhood of a cusp, the KKT conditions break down. The Webster dictionary says that a cusp is a fixed point on a mathematical curve at which a point tracing the curve would exactly reverse its direction of motion. Hence, it is wise to stay away from this event by invoking the constraint qualification. A simple numerical example is sufficient to illustrate this unfavorable mathematical event. max f (x = x subject to x ( x 3 0, x Figure 3 illustrates the problem s graph. Figure 3. Violation of KKT conditions This problem fails to satisfy the associated KKT conditions. In other words, it is not possible to express the gradient of the objective function as a linear combination (let alone a positive linear combination of the gradient of the constraint. To verify this assertion we specify the Lagrange function: L = x + λ[( x 3 x ] with KKT conditions for x = 3λ( x 0 x = x [ 3λ( x ] 7

Looking at figure 3, the objective function is maximized at x * =. But at this optimal value of the endogenous variable the KKT conditions result in a contradiction because 0. Matching the KKT conditions with the structure of the Economic Equilibrium f g. = λ 0 x x x MR MC 0 f g. x = x λ x x x Q(MR MC 3. x Q 4. = b g(x S D λ 5. λ = λ [b g(x] P(S D λ 6. λ P As stated in Lecture Notes, the Economic Equilibrium structure is more general than the KKT conditions since it does not imply differentiability, integrability and optimization. This matching of the two structures is very important for all the material that we will discuss in this course. I suggest (recommend to learn all these relationships by heart for how they look like and for what they mean geometrically and economically. 8