Topology of Lagrange multipliers

Similar documents
AN ALTERNATIVE TO THE THEORY OF EXTREMA

Nonlinear Programming

Lagrange Relaxation and Duality

Palais-Smale condition for multi-time actions that produce Poisson-gradient PDEs

Multitime sine-gordon solitons via geometric characteristics

MATH2070 Optimisation

Equivalence of multitime optimal control problems

GEOMETRIC PROGRAMMING APPROACHES OF RELIABILITY ALLOCATION

H-convex Riemannian submanifolds

Generalization to inequality constrained problem. Maximize

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Symmetric and Asymmetric Duality

1 Lagrange Multiplier Method

Constrained Optimization and Lagrangian Duality

Lagrange Multipliers

Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers

Linearized geometric dynamics of Tobin-Benhabib-Miyao economic flow

Convex Optimization M2

Linear and Combinatorial Optimization

Mechanical Systems II. Method of Lagrange Multipliers

Additional Homework Problems

EE 227A: Convex Optimization and Applications October 14, 2008

Applications of Linear Programming

Lagrangian Duality Theory

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lagrangian Duality. Evelien van der Hurk. DTU Management Engineering

Lagrangian Duality. Richard Lusby. Department of Management Engineering Technical University of Denmark

4TE3/6TE3. Algorithms for. Continuous Optimization

Convex Functions and Optimization

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

PARALLEL TRANSPORT AND MULTI-TEMPORAL CONTROLLABILITY

Lecture: Duality of LP, SOCP and SDP

8. Constrained Optimization

Semi-infinite programming, duality, discretization and optimality conditions

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Examination paper for TMA4180 Optimization I

1 Eigenvalues and eigenvectors of the Laplacian on a parallelepiped

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Convex Optimization and Modeling

Numerical study for two point boundary problems in nonlinear fluid mechanics

Optimality, Duality, Complementarity for Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Optimality Conditions for Constrained Optimization

3.7 Constrained Optimization and Lagrange Multipliers

SUBTANGENT-LIKE STATISTICAL MANIFOLDS. 1. Introduction

Convex Optimization Boyd & Vandenberghe. 5. Duality

ICS-E4030 Kernel Methods in Machine Learning

Multitime Samuelson-Hicks diagonal recurrence

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Lecture 4: Optimization. Maximizing a function of a single variable

Publications: Journal Articles

5. Duality. Lagrangian

IE 5531: Engineering Optimization I

Cylindrical Tzitzeica Curves Implies Forced Harmonic Oscillators

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Convex Optimization & Lagrange Duality

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

g(x,y) = c. For instance (see Figure 1 on the right), consider the optimization problem maximize subject to

Module 2: First-Order Partial Differential Equations

Fractal Magnetic Dynamics around a Koch Fractal Electric Circuit

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

MATH529 Fundamentals of Optimization Constrained Optimization I

Part IB Optimisation

Chap 2. Optimality conditions

Algorithms for constrained local optimization

EE/AA 578, Univ of Washington, Fall Duality

MATH20132 Calculus of Several Variables. 2018

CS-E4830 Kernel Methods in Machine Learning

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture: Duality.

On some special vector fields

Introduction to Nonlinear Stochastic Programming

Доклади на Българската академия на науките Comptes rendus de l Académie bulgare des Sciences Tome 69, No 9, 2016 GOLDEN-STATISTICAL STRUCTURES

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Non-Classical Lagrangian Dynamics and Potential Maps

Math 273a: Optimization Subgradients of convex functions

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Implicit Functions, Curves and Surfaces

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Preliminary draft only: please check for final version

Constrained maxima and Lagrangean saddlepoints

3. Minimization with constraints Problem III. Minimize f(x) in R n given that x satisfies the equality constraints. g j (x) = c j, j = 1,...

Support Vector Machines

Lecture 8. Strong Duality Results. September 22, 2008

Optimization for Machine Learning

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

LECTURE 10 LECTURE OUTLINE

Constrained Optimization Theory

A DUALITY ALGORITHM FOR THE OBSTACLE PROBLEM

Chapter 3: Constrained Extrema

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

subject to (x 2)(x 4) u,

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Transcription:

Topology of Lagrange multipliers C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru Abstract. We re-discuss the well-known programs with holonomic constraints insisting on the following issues: (i) the Lagrange-dual problem with weak respectively strong duality; (ii) the Wolfe-dual problem; (iii) the topology of Lagrange multipliers; (iv) the interpretation of Lagrange multipliers; (v) pertinent examples. M.S.C. 2010: 49K35, 58A17. Key words: optimization; dual problems; topology of Lagrange multipliers; meaning of Lagrange multipliers. 1 Classical Lagrange and Wolfe dual programs In this Section, we re-discuss the well-known programs with holonomic constraints insisting on the following issues [1], [3], [4]: (i) the Lagrange-dual problem with weak respectively strong duality; (ii) the Wolfe-dual problem. 1.1 The Lagrange dual problem Let D be a domain in R n, let x = (x 1,..., x n ) be a point in D and f : D R and g α : D R be convex functions. Denote g = (g α ) and we introduce the set Ω = {x D g α (x) 0, α = 1,..., m} = {x D g(x) 0}. A complete notation for this set is (Ω, g, ), but for short the sign or the pair (g, ) are suppressed in the notation. Let us consider the convex program (P ) min{f(x) x Ω}. x The Lagrange function (or Lagrangian) of (P) L(x, λ) = f(x) + λ α g α (x) = f(x)+ < λ, g >, x D, λ 0 BSG Proceedings 22. The International Conference Differential Geometry - Dynamical Systems DGDS-2014, September 1-4, 2014, Mangalia-Romania, pp. 84-97. c Balkan Society of Geometers, Geometry Balkan Press 2015.

Topology of Lagrange multipliers 85 is convex in x and linear in λ. Remark We can create an original Riemannian geometry on the set of critical points, using similar ideas we shall develop in a further 3.2.1. For all λ 0, the inequality holds. Consequently λ α g α (x) 0, x Ω (1) L(x, λ) f(x), x Ω, λ 0. The equality holds iff (complementarity conditions) λ α g α (x) = 0 (for each α = 1,..., m). Let us introduce the Lagrange dual function ψ(λ) = inf {f(x) + m λ α g α (x), x D, λ 0}. x D This function ψ(λ) is concave, because it is a point-wise infimum of affine functions. Indeed, using the linearity of L(x, λ) with respect to λ, and introducing λ 1 0, λ 2 0, and 0 t 1, we have ψ(tλ 1 + (1 t)λ 2 ) = inf x D L(x, tλ1 + (1 t)λ 2 ) = inf x D (tl(x, λ1 ) + (1 t)l(x, λ 2 )) inf x D (tl(x, λ1 )) + inf x D ((1 t)l(x, λ2 )) = t inf x D L(x, λ1 ) + (1 t) inf x D L(x, λ2 ) = tψ(λ 1 ) + (1 t)ψ(λ 2 ). Definition 1.1. The problem sup {ψ(λ) λ 0} λ is the so-called Lagrange dual problem of (P). The Lagrange dual problem can be called convex because it is equivalent to the convex problem inf { ψ(λ) λ 0}. λ The Lagrange-dual problem is also defined in this way if (P) is not convex. The following theorem holds also in that case. Theorem 1.1. (weak duality) The dual function yields lower bounds of the initial optimal value f, i.e., for any λ, we have ψ(λ) f. In other words, sup {ψ(λ) λ 0} inf {f(x)}. λ x Ω

86 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru Proof In the foregoing statements, we have the relation (1). Since Ω D, for each λ 0, we find ψ(λ) = inf L(x, λ) inf L(x, λ) inf f(x). x D x Ω x Ω Thus the statement in the theorem is true. The problem of finding the best lower bound on f obtained from the Lagrange dual function is called the Lagrange dual problem for the original or primal problem. The optimal values may be different. However, they are equal if (P) satisfies the Slater condition and has finite optimal value. This is the next result. Theorem 1.2. (strong duality) If the program (P) satisfies the Slater condition and has finite optimal value, then sup {ψ(λ) λ 0} = inf {f(x) g(x) 0}. λ x D Moreover, then the dual optimal value is attained. Proof Denote by f the optimal value of (P). Taking a = f in the Convex Farkas Lemma, it follows that there exists a vector λ = (λ 1,..., λ m ) 0 such that L(x, λ ) = f(x) + λ α g α (x) f, x D. Using the definition of ψ(λ ) this implies ψ(λ ) f. By the weak duality theorem, it follows that ψ(λ ) = f. This not only proves that the optimal values are equal, but also that λ is an optimal solution of the dual problem. Remark Unlike in Linear Programming theory, the strong duality theorem cannot always be established for general optimization problems. 2 Topology of Lagrange multipliers The aim of this Section is to give some original results regarding the topology of Lagrange multipliers set. Let f : D R n R and g : D R n R p, p < n, of class C 2 with rank J g = p in D. Let L (x, λ) = f (x) + λ g (x), with λ R p. We recall that H (x, λ) = f (x) + λ g (x) = 0 is the equation of critical points with respect to x of the Lagrange function. Let A = {x λ cu H (x, λ) = 0} and B = {λ x cu H (x, λ) = 0}. Introduce h : A B such that H (x, h (x)) = 0. The function h is well defined since the equation H (x, λ) = 0 is linear in λ (system with unique solution). Hence, for any λ B, the set h 1 (λ) is non-void, and it consists of all critical points corresponding to λ (set in which the nondegenerate critical points are isolated). Proposition 2.1. Let λ 0 B such that there exists x 0 h 1 (λ 0 ) with the property that x 0 is nondegenerate, i.e., the Hessian d 2 f (x 0 ) + λ 0 d 2 g (x 0 ) is nondegenerate. Then h admits a differentiable section s λ0 : I λ0 A.

Topology of Lagrange multipliers 87 Proof. Since H x (x 0, λ 0 ) = d 2 f (x 0 ) + λ 0 d 2 g (x 0 ) is non-degenerate, by hypothesis, there exists a neighborhood I λ0 of λ 0 and a differentiable function s λ0 : I λ0 A such that H (s λ0 (λ), λ) = 0, λ I λ0 and s λ0 (λ 0 ) = λ 0. Moreover, the function s λ0 is unique, with these properties. For any λ B, let S λ be the set of all sections of h defined in a neighborhood of λ, set which is eventually void. Remark 2.1. (i) If h 1 (λ) contains at least one nondgenerate critical point, then S λ is non-void. If h 1 (λ) does not contain degenerate critical points, then the sets h 1 (λ) and S λ have{ the same cardinal and are discrete sets. (ii) The set C = λ B S λ } is open. In the following, we suppose that the set S λ is finite, for any λ B. We can define f : B R by f (λ) = max s Sλ f (s (λ)), if S λ and f (λ) =, if S λ =. Proposition 2.2. (i) For any λ B, we have f (λ) sup x h 1 (λ) f (x). (ii) If h 1 (λ) does not contain degenerate critical points, then f (λ) = sup x h 1 (λ) f (x) = max x h 1 (λ) f (x). Proof. (i) Let s 0 S λ cu f (s 0 (λ)) = max s Sλ f (s (λ)) = f (λ). Since s 0 (λ) h 1 (λ), it follows that f (λ) sup x h 1 (λ) f (x). (2) By hypothesis, the sets S λ şi h 1 (λ) have the same cardinal, hence h 1 (λ) is finite. Let y 0 h 1 (λ) with f (y 0 ) = max x h 1 (λ) f (x). Since (y 0, λ) is a nondegenerate critical point, there exists s 1 S λ with s 1 (λ) = y 0. Then, it follows that max f (s (λ)) f (y 0 ) = max (λ) f (x). s S 1 λ Proposition 2.3. Let λ 0 B such that h 1 (λ 0 ) does not contain degenerate critical points. Suppose, also, that f h 1 (λ 0) is injective. Then there exists s 0 S λ0, s 0 : I 0 A such that f (λ) = f (s 0 (λ)), for any λ I 0. Proof. Let s 0 S λ0, s 0 : I 0 A such that f (λ 0 ) = f (s 0 (λ 0 )). Then f (λ 0 ) = f (s 0 (λ 0 )) > f (s (λ 0 )), s S λ0, s : I s A. Since f is continuous and the set S λ0 is finite, it follows that we can restrict the neighborhood I 0 such that f (s 0 (λ)) > f (s (λ)), λ I 0, s S λ0, i.e., f (λ) = f (s 0 (λ)), λ S λ0. x h 3 The meaning of Lagrange multiplier In our mostly geometrical discussion, λ is just an artificial variable that lets us compare the directions of the gradients without worrying about their magnitudes. To express mathematically the meaning of the multiplier, we write the constraint in the form g(x) = c for some constant c. This is mathematically equivalent to our usual

88 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru g(x) = 0, but allows us to easily describe a whole family of constraints. For any given value of c, we can use Lagrange multipliers to find the optimal value of f(x) and the point where it occurs. Call that optimal value f, occurring at coordinates x 0 and with Lagrange multiplier λ 0. The answers we get will all depend on what value we used for c in the constraint, so we can think of these as functions of c : f (c), x 0 (c), λ 0 (c). Of course, f(x) only depends on c because the optimal coordinates x 0 depend on c: we could write it as f (c). To find how the optimal value changes when we change the constraint, just take the derivative df dc = f x i 0 dx i 0 dc = f dx 0 dc. Use the equation of critical points to substitute f = λ 0 g 0 and obtain df dc = λ 0 g 0 dx 0 dc = λ dg 0 0 dc. But the constraint function g 0 = g(x 0 (c)) is always equal to c, so dg 0 /dc = 1. Thus, df /dc = λ 0. That is, the Lagrange multiplier is the rate of change of the optimal value with respect to changes in the constraint. Of course, f depends on c through of λ, and then df dc = df dλ. We can define dλ dc c(λ) by Cauchy problem (EC) dc dλ = 1 df λ dλ, c(λ 0) = 0. Then another Lagrange dual function may be (LDF ) ϕ(λ) = f (x 0 (λ)) + λ c(λ). Proposition If optimum points are critical points, both Lagrange dual functions give the same solution. Hence strong duality holds. Proof Indeed, using (EC) we have ϕ (λ) = df dc + c(λ) + λ dλ dλ = c(λ) and ϕ (λ) = 0 implies c(λ) = 0, that is for λ 0. Often the Lagrange multiplier have an interpretation as some quantity of interest: (i) λ is the rate of change of the quantity being optimized as a function of the constraint variable since L c = λ; (ii) by the envelope theorem the optimal value of a Lagrange multiplier has an interpretation as the marginal effect of the corresponding constraint constant upon the optimal attainable value of the original objective function: if we denote values at the optimum with an asterisk, then it can be shown that d dc f = d dc f(x(c)) = λ. For details regarding classical theory of programs see [1], [3], [4].

Topology of Lagrange multipliers 89 If we have more constraints g α (x) = c α, α = 1,..., m, then the Lagrange function is L(x, λ) = f(x) + m λ α(g α (x) c α ) and the system of critical points is m f x i + λ α g α x i = 0. Because the optimal coordinates x 0 and the optimal value f depend on vector c, taking the derivatives we have f c α = f x i 0 x i 0 c α = β=1 λ 0 β g β x i 0 x i 0 c α = β=1 λ 0 β g β c α = λ 0 β δ αβ = λ 0 α. Then we can define c(λ) by the partial differential system, with initial condition, written in matrix language as (EC) [λ 1... λ m ] c 1 c 1 λ m λ 1............ c m λ 1... c m λ m Then another Lagrange dual function may be (LDF ) ϕ(λ) = f (x 0 (λ)) + = [ f λ 1... λ α c α (λ). β=1 ] f, c(λ 0 ) = 0. λ m Deriving the β-th equation with respect to λ α and the α-th equation with respect to λ β, in the previous system, we obtain the complete integrability conditions c α = c β (symmetric Jacobian matrix); consequently c(λ) is the gradient of a λ β λ α scalar function, namely the Lagrange dual function ϕ(λ). Moreover, the previous square matrix being a symmetrical one, we can write the equation (EC) as [λ 1... λ m ] c 1 c m λ 1... λ 1......... c 1 λ m... c m λ m = [ ] f λ 1... f λ m. Consequently, in a regular case, we have the following situation: Solving a constrained optimum problem we obtain the optimal value as f = f(c 1,..., c m ). For the dual problem we use a f = f(λ 1,..., λ m ). If the correspondence between (c 1,..., c m ) and (λ 1,..., λ m ) is like a change of variables there hold the relations: grad c f = λ ; c (λ) λ = grad λ f, c (λ) L(R m, R m ). 4 Examples and counter-examples We solve pertinent examples of constrained optimization problems.

90 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru Figure 1: Geometry of Lagrange duality (1) Let us consider the functions f(x, y) = x 2 + y 2 and g(x, y) = x 2 + y 2 2x and the problem min f(x, y) constrained by g(x, y) = c, c 1. The Lagrange function of this problem is L(x, y, λ) = x 2 + y 2 + λ(x 2 + y 2 2x c). The critical points of the partial function (x, y) L(x, y, λ) are the solutions of the system 1 L 2 x = x + λx λ = 0, 1 L = y + λy = 0. 2 y ( 2. It follows x = λ λ+1, y = 0 and hence f λ = λ+1) On the other hand, by restriction, in critical points, we have the relation ( ) 2 λ 2 λ λ + 1 λ + 1 = c. It follows df dc = df dλ 1 dc dλ and finally, we obtain the geometrical interpretation df dc The dual function is = λ. λ2 ψ(λ) = L(x(λ), y(λ), λ) = λ + 1 λc. The value ψ(λ) is a minimum for λ > 1 and a maximum for λ < 1, in the initial problem. The condition of extremum (critical point), ψ (λ) = 0, is equivalent to (c + 1)(λ + 1) 2 = 1 and the dual problem has the same solution as the primal one.

Topology of Lagrange multipliers 91 On the other hand the equation (EC) for this problem is dc dλ = 1 λ We find c(λ) = d dλ ( ) 2 λ, c(λ 0 ) = 0, where λ 0 + 1 = ± 1. λ + 1 c + 1 1 1 c and the dual Lagrange function (λ + 1) 2 ϕ(λ) = ( ) 2 ( ) λ 1 + λ λ + 1 (λ + 1) 2 1 c = ψ(λ) as the above one. The Geometry of Lagrange duality is suggested in Fig. 1. (2) A problem with two constraints Solve the following constrained optimum problem: f(x, y, z) = xyz = extremum constrained by The Lagrange function is g 1 (x, y, z) = x + y a = 0, g 2 (x, y, z) = xz + yz b = 0. L(x, y, z, λ, µ) = xyz + λ(x + y a) + µ(xz + yz b) and the feasible solution of the problem is, only, x = y = a 2 = 2µ, z = b a = λ µ, λ = b 4, µ = a 4, f = ab 4 = 4λµ. The Lagrange dual function is ψ(λ, µ) = 4λµ aλ bµ. The partial differential system which defines c 1 (λ, µ) and c 2 (λ, µ) becomes, in this case, (EC) [λ µ] c 1 λ c 2 λ c 1 µ c 2 µ ( = [4µ 4λ], c 1 b 4, a 4 ) ( = c 2 b ) 4, a = 0. 4 Taking into account that c 1 µ = c 2, we obtain two quasilinear PDEs λ with solutions, respectively λ c 1 λ + µ c 1 µ = 4µ, λ c 2 λ + µ c 2 µ = 4λ, c 1 (λ, µ) = 4µ + α ( ) λ, c 2 (λ, µ) = 4λ + β µ ( ) λ, µ

92 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru α, β arbitrary functions. The condition c 1 µ = c 2 is verified, for instance, if α and λ β are constant functions. Using the initial conditions, we find finally c 1 (λ, µ) = 4µ a, c 2 (λ, µ) = 4λ b, ϕ(λ, µ) = 4λµ + λ( 4µ a) + µ( 4λ b) = ψ(λ, µ). (3) A strange problem Solve the following constrained optimum problem: constrained by So the Lagrange function is f(x, y, z) = xyz = extremum g 1 (x, y, z) = x + y + z a = 0, g 2 (x, y, z) = xy + xz + yz b = 0. L(x, y, z, λ, µ) = xyz + λ(x + y + z a) + µ(xy + xz + yz b) and one from the solutions of the problem is, for instance, µ = a a 2 3b 3 x = y = a + a 2 3b 3 with the extremum value f = 1 27, λ = 2a2 3b + 2a a 2 3b 9 = µ, z = a 2 a 2 3b 3 = µ 2, = a + 2µ, ( 2a 3 + 9ab 2(a 2 3b) 3/2) = λ(a + 2µ), only if a 2 3b 0. Remark Another solution of the problem is with the extremum value f = 1 27 µ = a + a 2 3b 3,... and so on ( 2a 3 + 9ab + 2(a 2 3b) 3/2) = λ(a + 2µ). The interval [f, f ] solves the following algebraic problem: Find the real numbers m such that the equation t 3 at 2 + bt m = 0, a, b R, has tree real roots. It is easily to verify that a f (a, b) = λ and b f (a, b) = µ. (λ, µ) On the other hand, (a, b) = 0 and f cannot be expressed as function of λ and µ only. Then we have to consider f = f (a, b, λ(a, b), µ(a, b)) and the following relations λ = D a f = f a + f λ λ a + f µ µ a

Topology of Lagrange multipliers 93 µ = D b f = f b + f λ λ b + f µ µ b which is easily to verify also (here D. is an operator of total derivative.) Question Which is the dual Lagrange function ψ(λ, µ) in this case? Solving the system of the critical points with respect to x, y and z we find, for instance, x = y = µ, z undeterminate and λ = µ 2. With these, one obtains the dual Lagrange function ψ(λ, µ) = χ(µ) = µ 3 aµ 2 bµ. Remark Although z is undeterminate, the dual Lagrange function does not depend upon z, because with the above solutions the derivative L vanishes identically. z The critical points condition for the dual Lagrange function dχ dµ = (3µ2 + 2aµ + b) = 0 gives us the same solutions as in primal problem. Open problem How it means and how we find the functions c 1 and c 2 in the (λ, µ) situation, like this, when (a, b) = 0? (4) Let us consider the functions f(x, y) = x 2 + y 2 and g(x, y) = x + y, with (x, y) R 2, and the problem The Lagrange function is min f(x, y) constrained by g(x, y) 1. L(x, y, λ) = x 2 + y 2 + λ(1 x y), (x, y) R 2, λ 0. The function (x, y) L(x, y, λ) is convex. Consequently, it is minimal iff L L = 0, x y = 0. This holds if x = λ 2, y = λ 2. Substitution gives ψ(λ) = λ λ2 2, λ 0. The dual problem max ψ(λ) has the optimal point λ = 1. Consequently, x = y = 1 2 is the optimal solution of the original (primal) problem. In both cases the optimal value equals 1 2, i.e., at optimality the duality gap is zero! (5) Let us solve the program min x subject to x 2 0, x R. This program is not Slater regular. On the other hand, we have { 1 ψ(λ) = inf (x + x R λx2 ) = 2λ for λ > 0 for λ = 0.

94 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru Obviously, ψ(λ) < 0 for all λ 0. Consequently, sup{ψ(λ) λ 0} = 0. So the Lagrange-dual has the same optimal value as the primal problem. In spite of the lack of Slater regularity there is no duality gap. (6) (Example with positive duality gap) We consider the program min e y subject to x 2 + y 2 x 0, (x, y) R 2. Here the feasible region is Ω = {(x, y) R 2 x 0, y = 0}. Consequently this program is not Slater regular. The optimal value of the objective function is 1. The Lagrange function is L(x, y, λ) = e y + λ( x 2 + y 2 x). The Lagrange dual program can be written in the form sup ψ(λ) subject to λ 0. Note that L(x, y, λ) > 0 implies ψ(λ) 0. Now let ɛ > 0. Fixing y = ln ɛ and x = y2 ɛ 2 2ɛ, we find x 2 + y 2 x = ɛ and In this way, ψ(λ) = inf (x,y) R L(x, y, λ) = (1 + λ)ɛ. L(x, y, λ) inf (1 + λ)ɛ = 0. 2 On the other hand, we also have ψ(λ) 0, and consequently the optimal value of the Lagrange dual program is 0, and hence the minimal duality gap equals 1. Of course, here we have no strong duality here. 5 The Wolfe-dual problem ɛ>0 The Lagrange dual program can be written in the form sup { inf {f(x) + m λ 0 x D λ α g α (x)}}. Assume that D = R n and the functions f, g 1,..., g m are continuously differentiable and convex. For a given λ 0 the inner minimization problem is convex, and we can use the fact that the infimum is attained if and only if the gradient with respect to x is zero. Definition 5.1. The problem (W P ) sup x,λ subject to {f(x) + f m x i (x) + λ α g α (x)} is called the Wolfe dual of the program (P). g α λ α (x) = 0, λ 0 xi

Topology of Lagrange multipliers 95 Obviously, the constraints in Wolfe dual are usually nonlinear. In such cases the Wolfe-dual is not a convex program. The Wolfe dual has the weak duality property. Theorem 5.1. (weak duality property) Suppose that D = R n and the functions f, g 1,..., g m are continuously differentiable and convex. If ˆx is a feasible solution of (P) and ( x, λ) is a feasible solution for (WP), then L( x, λ) f(ˆx). In other words, weak duality holds for (P) and (WP). 5.1 Example (1) Let us consider the convex program min x,y x + ey subject to 3x 2e y 10, y 0, (x, y) R 2. Then the optimal value is 5 with x = 4, y = 0. The Wolfe dual of this program is subject to sup {x + e y + λ 1 (10 3x + 2e y ) λ 2 y} x,y,λ 1 3λ 1 = 0, e y + 2e y λ 1 λ 2 = 0, (x, y) R 2, λ 0. Obviously, the Wolfe dual program is not convex. It follows λ 1 = 1 3 and the second constraint becomes 5 3 ey λ 2 = 0. Eliminating λ 1, λ 2 from the objective function, we find g(y) = 5 3 ey 5 3 y ey + 10 3. This function has a maximum when g (y) = 0, i.e., y = 0 and f(0) = 5. Hence the optimal value of (WP) is 5 and then (x, y, λ 1, λ 2 ) = (4, 0, 1 3, 5 3 ). Remark The substitution z = e y 1 makes the problem linear. 6 Minimax inequality For any function φ of two vector variables x X, y Y, the minimax inequality is true. Indeed, start from max y Y min φ(x, y) min max φ(x, y) x X x X y Y x, y : min x X φ(x, y) max y Y φ(x, y ) and take the minimum over x X on the right-hand side, then the maximum over y Y on the left-hand side. Weak duality is a direct consequence of the minimax inequality. To see this, start from the unconstrained formulation of Lagrange, and apply the above inequality, with φ = L the Lagrangian of the original problem, and y = λ the Lagrange vector multiplier. Acknowledgements Partially supported by University Politehnica of Bucharest and by Academy of Romanian Scientists.

96 C. Udrişte, M. Constantinescu, I. Ţevy, O. Dogaru References [1] H. Bonnel, Analyse Fonctionnelle, Maitrise de Mathématiques, Ingénerie Mathéematiques, Université de La Reunion, Faculté de Sciences et Technologies, 2015. [2] G. Darboux, Sur le problème de Pfaff, Bull. Sci. Math. 6 (1882), 14-36, 49-68. [3] R. B. Holmes, Geometric Functional Analysis and Its Applications, Springer- Verlag, New York, 1975. [4] K. Roos, Nonlinear Programming, LNMB Course, De Uithof, Utrecht, TUDelft, February 6 - May 8, A.D. 2006. [5] S. Sternberg, Lectures on Differential Geometry, Prentice Hall, 1964. [6] O. Dogaru, V. Dogaru, Extrema Constrained by C k Curves, Balkan Journal of Geometry and Its Applications, 4, 1 (1999), 45-42. [7] O. Dogaru, I. Ţevy, Extrema Constrained by a Family of Curves, Proc. Workshop ob Global Analysis, Diff. Geom. and Lie Algebras, 1996, Ed. Gr. Tsagas, Geometry Balkan Press, 1999, 185-195. [8] I. Ekeland, Exterior Differential Calculus and Applications to Economic Theory, Quaderni Scuola Normale Superiore di Pisa, 1998, Italy. [9] R. Montgomery, A Tour of Subriemanniene Geometries, Their Geodesics and Applications, Mathematical Surveys and Monographs, 91, American mathematical Society, 2002. [10] V. Radcenco, C. Udrişte, D. Udrişte, Thermodynamic Systems and Their Interaction, Sci. Bull. P.I.B., Electrical Engineering, vol. 53, no. 3-4 (1991), 285-294. [11] Gr. Tsagas, C. Udrişte, Vector Fields and Their Applications, Geometry Balkan Press, Bucharest, 2002. [12] C. Udrişte, O. Dogaru, Mathematical Programming Problems with Nonholonomic Constraints, Seminarul de Mecanică, Univ. of Timişoara, Facultatea de Ştiinţe ale Naturii, vol. 14, 1988. [13] C. Udrişte, O. Dogaru, Extrema with Nonholonomic Constraints, Sci. Bull., Polytechnic Institute of Bucharest, Seria Energetică, Tomul L, 1988, 3-8. [14] C. Udrişte, O. Dogaru, Extreme condiţionate pe orbite, Sci. Bull., 51 (1991), 3-9. [15] C. Udrişte, O. Dogaru, Convex Nonholonomic Hypersurfaces, Math. Heritage of C.F. Gauss, 769-784, Ed. G. Rassias, World Scientific, 1991. [16] C. Udrişte, O. Dogaru, I. Ţevy, Sufficient Conditions for Extremum on Differentiable Manifolds, Sci. Bull., Polytechnic Institute of Bucharest, Electrical Engineering, vol. 53, no. 3-4 (1991), 341-344. [17] C. Udrişte, O. Dogaru, I. Ţevy, Extremum Points Associated with Pfaff Forms, Presented at the 90th Anniversary Conference of Akitsugu Kawaguchi s Birth, Bucharest, Aug. 24-29, 1992; Tensor, N.S., Vol. 54 (1993), 115-121. [18] C. Udrişte, O. Dogaru, I. Ţevy, Open Problem in Extrema Theory, Sci. Bull. P.U.B., Series A, Vol. 55, no.3-4 (1993), 273-277. [19] O. Dogaru, I. Ţevy, C. Udrişte, Extrema Constrained by a Family of Curves and Local Extrema, JOTA, vol. 97, no.3, June 1998, 605-621. [20] C. Udrişte, O. Dogaru, I. Ţevy, Extrema Constrained by a Pfaff System, Hadronic J. Supplement, USA, 1991-Proc. Int. Workshop on Fundam. Open Problems in Math., Phys. and Other Sciences, Beijing, August 28, 1997.

Topology of Lagrange multipliers 97 [21] C. Udrişte, I. Ţevy, M. Ferrara, Nonholonomic Economic Systems, see [28], 139-150. [22] C. Udrişte, I. Ţevy, Geometry of test Functions and Pfaff Equations, see [28], 151-165. [23] C. Udrişte, O. Dogaru, I. Ţevy, Extrema with Nonholonomic Constraints, Geometry Balkan Press, Bucharest, 2002. [24] C. Udrişte, O. Dogaru, M. Ferrara, I. Ţevy, Pfaff Inequalities and Semi-curves in Optimum Problems, Recent Advances in Optimization, pp. 191-202, Proceedings of the Workshop held in Varese, Italy, June 13/14th 2002, Ed. G.P. Crespi, A. Guerraggio, E. Miglierina, M. Rocca, DATANOVA, 2003. [25] C. Udrişte, O. Dogaru, M. Ferrara, I. Ţevy, Pfaff inequalities and semi-curves in optimum problems, in Edt. G. P. Crespi, A. Guerraggio, E. Miglierina, M. Rocca, Recent Advances in Optimization, Proceedings of the Workshop held in Varese, Italy, June 13-14, 2002, pp. 191-202. [26] C. Udrişte, O. Dogaru, M. Ferrara, I. Ţevy, Extrema with constraints on points and/or velocities, Balkan Journal of Geometry and Its Applications, 8, 1(2003), 115-123. [27] C. Udrişte, O. Dogaru, M. Ferrara, I. Ţevy, Nonholonomic optimization theory, see [28], 177-192, Geometry Balkan Press, Bucharest, 2004. [28] C. Udriste, M. Ferrara, D. Opris, Economic Geometric Dynamics, Geometry Balkan Press, Bucharest, 2004. [29] Gh. Vrănceanu, Leçons de Geometry Differentielle, Editions de l Academie Roumaine, Bucarest, 1957-1975. Authors address: Constantin Udrişte, Mădălina Constantinescu, Ionel Ţevy and Oltin Dogaru University Politehnica of Bucharest, Faculty of Applied Sciences Department of Mathematics and Informatics, Splaiul Independentei 313 Bucharest 060042, Romania. Email: udriste@mathem.pub.ro, madalina.const@yahoo.com vascatevy@yahoo.fr, oltin.horia@yahoo.com