PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS WITH A RANDOM TECHNOLOGY MATRIX

Similar documents
ON PROBABILISTIC CONSTRAINED PROGRAMMING

A CONVEXITY THEOREM IN PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS

Maximization of a Strongly Unimodal Multivariate Discrete Distribution

CONTRIBUTIONS TO THE THEORY OF STOCHASTIC PROGRAMMING. Andras Prekopa. Technological University and Hungarian Academy of Sciences, Budapest, Hungary

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

On Strong Unimodality of Multivariate Discrete Distributions

ON SUM OF SQUARES DECOMPOSITION FOR A BIQUADRATIC MATRIX FUNCTION

Assignment 1: From the Definition of Convexity to Helley Theorem

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

CS 246 Review of Linear Algebra 01/17/19

2 Chance constrained programming

Applications of Linear Programming

INVEX FUNCTIONS AND CONSTRAINED LOCAL MINIMA

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

1 Strict local optimality in unconstrained optimization

LINEAR INTERVAL INEQUALITIES

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

LECTURE 13 LECTURE OUTLINE

Linear Algebra. Session 12

Constrained Optimization

1 The linear algebra of linear programs (March 15 and 22, 2015)

Semidefinite Programming Basics and Applications


CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Comparing Convex Relaxations for Quadratically Constrained Quadratic Programming

Summary Notes on Maximization

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Optimization Problems with Probabilistic Constraints

Research Division. Computer and Automation Institute, Hungarian Academy of Sciences. H-1518 Budapest, P.O.Box 63. Ujvári, M. WP August, 2007

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Duality Theory of Constrained Optimization

R u t c o r Research R e p o r t. Application of the Solution of the Univariate Discrete Moment Problem for the Multivariate Case. Gergely Mádi-Nagy a

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

4. Convex optimization problems

Integer programming: an introduction. Alessandro Astolfi

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

Symmetric Matrices and Eigendecomposition

CONSTRAINED NONLINEAR PROGRAMMING

Written Examination

Chapter 2: Preliminaries and elements of convex analysis

GENERALIZED MATRIX MULTIPLICATION AND ITS SOME APPLICATION. 1. Introduction

Interior-Point Methods for Linear Optimization

Linear programs, convex polyhedra, extreme points

Chapter 1 Vector Spaces

Convex optimization problems. Optimization problem in standard form

Nonlinear Coordinate Transformations for Unconstrained Optimization I. Basic Transformations

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

Lecture 3. Optimization Problems and Iterative Algorithms

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

Mathematical Economics: Lecture 16

Linear Programming Duality

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

TECHNICAL NOTE. A Finite Algorithm for Finding the Projection of a Point onto the Canonical Simplex of R"

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)

Functions of Several Variables

Intrinsic products and factorizations of matrices

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

1 Positive definiteness and semidefiniteness

The Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment

Linear programming. Starch Proteins Vitamins Cost ($/kg) G G Nutrient content and cost per kg of food.

4. Convex optimization problems

Mathematics 206 Solutions for HWK 13b Section 5.2

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Lecture: Convex Optimization Problems

Lecture 13: Constrained optimization

The Singular Value Decomposition

Convexity of level sets for solutions to nonlinear elliptic problems in convex rings. Paola Cuoghi and Paolo Salani

II. Analysis of Linear Programming Solutions

Math (P)refresher Lecture 8: Unconstrained Optimization

The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1

1 Definition and Basic Properties of Compa Operator

1. Which one of the following points is a singular point of. f(x) = (x 1) 2/3? f(x) = 3x 3 4x 2 5x + 6? (C)

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

The extreme rays of the 5 5 copositive cone

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

A Tropical Extremal Problem with Nonlinear Objective Function and Linear Inequality Constraints

Chapter 1: Linear Programming

Lecture Unconstrained optimization. In this lecture we will study the unconstrained problem. minimize f(x), (2.1)

Notes on Linear Algebra and Matrix Theory

1 Convexity, concavity and quasi-concavity. (SB )

Planning and Optimization

MCE693/793: Analysis and Control of Nonlinear Systems

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Four new upper bounds for the stability number of a graph

Robust linear optimization under general norms

MATH 445/545 Test 1 Spring 2016

Lecture: Examples of LP, SOCP and SDP

MSc Quantitative Techniques Mathematics Weeks 1 and 2

MSc Quantitative Techniques Mathematics Weeks 1 and 2

ORIGINS OF STOCHASTIC PROGRAMMING

An LP-based inconsistency monitoring of pairwise comparison matrices

Abstract. We find Rodrigues type formula for multi-variate orthogonal. orthogonal polynomial solutions of a second order partial differential

Solving Dual Problems

On well definedness of the Central Path

SOME INDEFINITE METRICS AND COVARIANT DERIVATIVES OF THEIR CURVATURE TENSORS

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

On pairwise comparison matrices that can be made consistent by the modification of a few elements

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Linear Programming. (Com S 477/577 Notes) Yan-Bin Jia. Nov 28, 2017

Transcription:

Math. Operationsforsch. u. Statist. 5 974, Heft 2. pp. 09 6. PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS WITH A RANDOM TECHNOLOGY MATRIX András Prékopa Technological University of Budapest and Computer and Automation Institute of the Hungarian Academy of Sciences, H 502 Budapest XI. Kende utca 3 7, Hungary Eingereicht bei der Redaktion: 26. 2. 973 Abstract Probabilistic constraint of the type P Ax β p is considered and it is proved that under some conditions the constraining function is quasi-concave. The probabilistic constraint is embedded into a mathematical programming problem of which the algorithmic solution is also discussed. Introduction We consider the following function of the variable x: Gx =PAx β,. where A is an m n matrix, x is an n-component, β is an m-component vector and it is supposed that the entries of A as well as the components of β are random variables. The function. will be used as a constraining function in a nonlinear programming problem which we formulate below: Gx p, G i x 0, i =,...,N, a i x b i, i =,...,M, min fx..2 Here we suppose that a,...,a M are constant n-component vectors, b,...,b M are constants, the functions G,...,G M are quasi-concave in R n and fx isconvexinr n.many practical problems lead to the stochastic programming decision problem.2. Among

them we mention the nutrition problem in which case the nutrient contents of the foods are random and the ore buying problem in which case the ferrum contents of the ores are random and in both cases.2 is a reasonable decision principle. Regarding the joint distribution of the entries of A and the components of β, two special cases will be considered. In the first one we suppose that this distribution is normal satisfying further restrictive conditions while in the second one lognormal or some more general distribution is involved. The reason why we deal with these conditions is of mathematical nature: we can prove in these cases the quasi-concavity of the constraining function in the probabilistic constraint. In Section 2 we prove three theorems and in Section 3 the algorithmic solution of Problem.2 is discussed. The theorems of Section 2 are based on the following theorems of the author proved in [2] resp. [3]. Theorem Let fx be a probability density function in the n-dimensional Euclidean space R n and suppose that it is logarithmic concave i.e. it has the form fx =e Qx, x R n,.3 where Qx is a convex function in the entire space. Let further A, B be two convex subsets of R n and 0 <λ<. Denoting by PC the integral of fx over the measurable subset C of R n, we have PλA + λb [PA] λ [PB] λ..4 The constant multiple of a set and the Minkowski sum of two sets are defined by kc = {kc c C} resp. C + D = {c + d c C, d D}, wherek is a real number. Theorem implies that if A is a convex subset of R n then PA + x is a logarithmic concave function in the entire space. If we take in particular A = {t t 0}, thenwesee that the probability distribution function F x belonging to the density function fx is logarithmic concave in the entire space, since PA + x =F x. A probability measure P in R n is said to be logarithmic concave if for every pair A, B of convex subsets of R n and for every 0 <λ<, the inequality.4 is satisfied see [2]. Theorem 2 Let g i x, y, i =,...,r be convave functions in R m+n where x is an n-component and y is an m-component vector. Let ξ be an m-component random vector having a logarithmic concave probability distribution. Then the function of the variable x: is logarithmic concave in the space R n. Pg i x, ξ 0, i =,...,r.5 The formulation of Theorem 2 is slightly different from that given in [3]. Its proof is the same as the proof of the original version only a trivial modification is necessary. A function F x defined on a convex set K is said to be quasi-concave if for every pair x, x 2 K we have F λx + λx 2 minf x,fx 2. It is easy to see that a necessary and sufficient condition for this is that the set {x x K, F x b} is convex for every real b. 2

2 Quasi-concavity theorems concerning the function. Let ξ,...,ξ n denote the columns of the matrix A and introduce the notation ξ = β. Theorem 3 Suppose that the altogether mn + components of the random vectors ξ,...,ξ have a joint normal distribution where the cross-covariance matrices of ξ i and ξ j are constant multiples of a fixed covariance matrix C i.e. where Then the set of x vectors satisfying is convex provided p /2. E[ξ i μ i ξ j μ j ]=s ij C, i, j =,...,, 2. μ i = Eξ i, i =,...,. 2.2 PAx β p 2.3 Proof. Consider the covariance matrix of Ax βx which is equal to the following E ξ i x i μ i x i ξ i x i μ i x i = s ij x i x j C = z SzC, 2.4 where z =x,x ands is the matrix with entries s ij. If all elements of C are zero, then the vectors ξ,...ξ are constants with probability and the statement of the theorem holds trivially. If there is at least one nonzero element in C, then there is at least one positive element in its main diagonal. This implies that z Sz 0 for every z since 2.4 is a covariance matrix hence all elements in its main diagonal are nonnegative. Thus S is a positive semidefinite matrix. We may suppose that both C and S are positive definite matrices because otherwise we can use a perturbation of the type C + εi m, S + εi, where I m and I are unit matrices of the size m m and n + n +, respectively, prove the theorem for these positive definite matrices and then take the limit ε 0. Let x = everywhere in the sequel. We have ξ i μ i x i μ i x i PAx β =P. 2.5 z Sz 2 z Sz 2 i,j= Introduce the notation Lz = μ i x i 3

and denote L i z theith component of Lz fori =,...,m. If c ik denote the elements of C then 2.5 can be rewritten as L z L m z PAx β =Φ,..., ; R, 2.6 c z Sz 2 c mm z Sz 2 where R is the correlation matrix belonging to the covariance matrix C. For every i i =,...,mwehave Φ L i z L z L m z Φ,..., ; R c ii z Sz 2 c z Sz 2 c mm z Sz 2. 2.7 Thus if the right hand side of 2.7 is greater than or equal to p /2 then we conclude L i z 0, i =,..., m. It is well-known that z Sz 2 is a convex function in the entire space R. Hence it follows that for every z, z 2, L 2 z + 2 z 2 [ 2 z + 2 z 2 S 2 z + ] λ Lz Lz 2 2 z 2 z Sz + λ 2 z 2 Sz, 2.8 2 2 2 where z λ = Sz 2 z Sz 2 +z 2 Sz. 2 2 The probability density function of any nondegenerated normal distribution is logarithmic concave in the entire space. Hence it follows that log Φ v c 2,...,v mc 2 mm; R is concave in the variables v,...,v m. Thus if the value of the function 2.6 is greater than or equal to p /2 for z = z and also for z = z 2, then the value of 2.6 will be greater than or equal to p /2 at the point determined by the right hand side of 2.8. Since Φ is monotonically increasing in all variables it follows from 2.8 that 2.6 is greater than or equal to p /2 at the point determined by the left hand side of 2.8. This proves the theorem. Theorem 4 Let A i denote the i-th row of A and β i denote the i-th component of β, i =,...,m. Suppose that the random row vectors of n + components A i β i, i =,...m 2.9 are independent, normally distributed and their covariance matrices are constant multiples of a fixed covariance matrix C. Then the set of x vectors on which the function. is greater than or equal to p, is convex provided p /2. 4

Proof. We may suppose that C is positive definite. We also suppose that the covariance matrix of 2.9 equals s 2 i C with s i > 0, i =,...,m.leta 0 i, b i be the expectation of A i and β i, respectively, i =,...,m and introduce the notation: L i z = A 0 i,b i z, i =,...,m where z =x,...,x n,. s i Then we have m m A i A 0 i, β i + b i z PAx β = PA i, β i z 0 = P s i z Cz 2 m L i z L z L m z = Φ =Φ,..., ; I z Cz 2 z Cz 2 z Cz m p, 2 A0 i,b i z s i z Cz 2 where I m is the m m unit matrix. From here the proof goes in the same way as the proof of Theorem 3. Let a ij, i =,...,m; j =,...,n denote the elements of the matrix A and for the sake of simplicity suppose that the columns of A are numbered so that those come first which contain random variables. Let r be the number of these columns and introduce the following notations: J denotes the set of those ordered pairs i, j for which a ij i m, j r; is random variable, J i denotes the set of those subscripts j, forwhichi, j J, where i m; K i denotes the set {,...,r} J i ; L denotes the set of those subscripts i, forwhichβ i is a random variable, where i m; T denotes the set {,...,m} L. Nowweprovethefollowing Theorem 5 Suppose that the random variables a ij, i, j J are positive with probability and the constants a ij, i m, j r, i, j / J are nonnegative. Suppose further that the joint distribution of the random variables α ij, i, j J, β i, i L is a logarithmic concave probability distribution, where α ij =loga ij, i, j J. Under these conditions the function hx,...,x r,x r+,...,x n =Ge x,...,e xr,x r+,...,x n 2.0 is logarithmic concave in the entire space R n,wheregx is the function defined by.. 5

Proof. Consider the following functions g i = e u ij+x j n a ij e xj a ij x j + v i, i L, j J i j K i j=r+ g i = e u ij+x j n a ij e xj a ij x j + b i, i T. j J i j K i j=r+ These are all supposed to be functions of the variables 2. u ij, i, j J; x j, j n; v i, i L, though not every variable appears explicitely in every function. The functions g,...,g m are clearly concave in the entire space. Substituting u ij by α ij for every i, j J and v i by β i for every i L, instead of the functions g,...,g m we obtain random variables whichwedenotebyγ,...,γ m, respectively. By Theorem 2, the function Pγ 0,...,γ m 0 is a logarithmic concave function of the variables x,...,x n. On the other hand we have obviously Pγ 0,...,γ m 0 = Ge x,...,e xr,x r+,...,x n, thus the proof of the theorem is complete. 3 Algorithmic solution of Problem.2 In this section we deal with the algorithmic solution of Problem.2 under the conditions of Theorem 3, Theorem 4 and Theorem 5, respectively. Consider the case of Theorem 3. We suppose that C and S are positive definite matrices. The other cases can be treated in a similar way. We suppose further that C is a correlation matrix i.e. its diagonal elements are equal to. The random vectors ξ,...,ξ can always be defined so that the constraint Ax β remains equivalent and this condition holds. Using the notation Lz introduced in Section 2, we can write Gx =PAx β =P ξ i μ i x i Lz 3. = P z Sz 2 ξ i μ i x i Lz =Φ Lz; C, z Sz 2 z Sz 2 where Φy; C denotes the probability distribution of the m-dimensional normal distribution with N0, marginal distributions and correlation matrix C. It is well-known that Φy, C y i c ji y i =Φ y i c 2, j =,...,i,i+,...,m; C i ϕy i, 3.2 ji 6

where ϕy = 2π e y2 2, <y< 3.3 and C i is an m m correlation matrix with the entries c jk c ji c ki, j,k =,...,i,i+,...,m. 3.4 c 2 ji c 2 ki With the aid of 3. 3.3 the gradient of the function gx can be obtained by an elementary calculation. If a nonlinear programming procedure solves the problem with quasi-concave constraints and convex objective function is to be minimized, it can be applied to solve our problem. Such a method is the method of feasible directions see [6] the convergence proof of which for the case mentioned above is given in [4] and in a more detailed form in [5]. The case of Theorem 4 does not present new difficulty. In fact we see in the proof of Theorem 4 that the function PAx β has the same form as the function standing on the right hand side in 3.. In the case of Theorem 5 we substitute x i by e x i everywhere in Problem. for i =,...,r. Then the first constraining function becomes logarithmic concave in all variables. It may happen that the other constraining functions are concave while the objective function is convex after this substitution. If this is the case and the original problem contained constraints of the form x i >D i, i =,...,r, 3.5 where D i, i =,...,r are positive constants, then we can solve the problem by convex programming procedures e.g. by the SUMT method see [] which seems to be very suitable due to the logarithmic concavity of the first constraining function. In fact using the logarithmic penalty function, the unconstrained problems are convex problems. The values of the probabilities can be obtained by simulation. If instead of 3.5 we had nonnegativity constraints for x,...,x r in the original problem then using the constraints 3.5 instead of the constraints x 0,...,x r 0 in the original problem, by a suitable selection for the constants D i, i =,...,r we can come arbitrarily near the original optimum value. References [] Fiacco, A. V. and G. P. McCormick 968. Nonlinear Programming: Sequential Unconstrained Minimization Techniques. Wiley, New York. [2] Prékopa, A. 97. Logarithmic concave measures with application to stochastic programming. Acta Sci. Math., Szeged, 32, 30 36. 7

[3] Prékopa, A. 972. A class of stochastic programming decision problems. Math. Operationsforsch. Statist., 3, 349 354. [4] Prékopa, A. 970. On probabilistic constrained programming. Proc. Princeton Sympos. Math. Programming, Princeton Univ. Press, Princeton, New Jersey, 3 38. [5] Prékopa, A. 974. Eine Erweiterung der sogenannten Methode der zulässigen Richtungen der nichtlinearen Optimierung auf den Fall quasikonkaver Restriktionsfunktionen. Math. Operationsforsch. und Statistik, 5, 28 293. [6] Zoutendijk, G. 960. Methods of Feasible Directions. Elsevier, Amsterdam. 8