A CONVEXITY THEOREM IN PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS

Similar documents
On Strong Unimodality of Multivariate Discrete Distributions

Maximization of a Strongly Unimodal Multivariate Discrete Distribution

SHARP BOUNDS FOR PROBABILITIES WITH GIVEN SHAPE INFORMATION

Discrete Moment Problem with the Given Shape of the Distribution

Method of Multivariate Lagrange Interpolation for Generating Bivariate Bonferroni-Type Inequalities

Solution of Probabilistic Constrained Stochastic Programming Problems with Poisson, Binomial and Geometric Random Variables

R u t c o r Research R e p o r t. Uniform partitions and Erdös-Ko-Rado Theorem a. Vladimir Gurvich b. RRR , August, 2009

A Stochastic Programming Based Analysis of the Field Use in a Farm

Bounding in Multi-Stage. Stochastic Programming. Problems. Olga Fiedler a Andras Prekopa b

R u t c o r Research R e p o r t. Application of the Solution of the Univariate Discrete Moment Problem for the Multivariate Case. Gergely Mádi-Nagy a

PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS WITH A RANDOM TECHNOLOGY MATRIX

R u t c o r Research R e p o r t. The Optimization of the Move of Robot Arm by Benders Decomposition. Zsolt Robotka a RRR , DECEMBER 2005

Single commodity stochastic network design under probabilistic constraint with discrete random variables

Maximum k-regular induced subgraphs

STRATEGIC EQUILIBRIUM VS. GLOBAL OPTIMUM FOR A PAIR OF COMPETING SERVERS

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

R u t c o r Research R e p o r t. Relations of Threshold and k-interval Boolean Functions. David Kronus a. RRR , April 2008

A lower bound for discounting algorithms solving two-person zero-sum limit average payoff stochastic games

On the Chvatál-Complexity of Knapsack Problems

Threshold Boolean Form for Joint Probabilistic Constraints with Random Technology Matrix

R u t c o r Research R e p o r t. The Branch and Bound Method

An Intersection Inequality for Discrete Distributions and Related Generation Problems

Slow k-nim. Vladimir Gurvich a

YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL

Remarks on multifunction-based. dynamical systems. Bela Vizvari c. RRR , July, 2001

Optimization Problems with Probabilistic Constraints

A Method of Disaggregation for. Bounding Probabilities of. Boolean Functions of Events

Convexity of chance constraints with dependent random variables: the use of copulae

Absolute Value Programming

ON PROBABILISTIC CONSTRAINED PROGRAMMING

ON DOMINATION IN CUBIC GRAPHS

Endre Boros b Vladimir Gurvich d ;

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

A four parametric generalization of the Wythoff NIM and its recursive solution

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

2. Matrix Algebra and Random Vectors

Assignment 1: From the Definition of Convexity to Helley Theorem

May 9, 2014 MATH 408 MIDTERM EXAM OUTLINE. Sample Questions

Chapter 1. Preliminaries

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?

R u t c o r Research R e p o r t. Empirical Analysis of Polynomial Bases on the Numerical Solution of the Multivariate Discrete Moment Problem

A strongly polynomial algorithm for linear systems having a binary solution

R u t c o r Research R e p o r t. Minimal and locally minimal games and game forms. 1. Endre Boros 2 Vladimir Gurvich 3 Kazuhisa Makino 4

An inverse Newton transform

Sharpening the Karush-John optimality conditions

2018 Fall 2210Q Section 013 Midterm Exam I Solution

R u t c o r Research R e p o r t. A New Imputation Method for Incomplete Binary Data. Mine Subasi a Martin Anthony c. Ersoy Subasi b P.L.

ORIGINS OF STOCHASTIC PROGRAMMING

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

On the Sprague-Grundy function of exact k-nim

A Criterion for the Stochasticity of Matrices with Specified Order Relations

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Static Problem Set 2 Solutions

Sample width for multi-category classifiers

Roots and Coefficients Polynomials Preliminary Maths Extension 1

Convex relaxations of chance constrained optimization problems

Some Results on the Multivariate Truncated Normal Distribution

Endre Boros b Khaled Elbassioni d

Relaxations of linear programming problems with first order stochastic dominance constraints

Convergence Concepts of Random Variables and Functions

Endre Boros a Ondřej Čepekb Alexander Kogan c Petr Kučera d

R u t c o r Research R e p o r t. A Method to Schedule Both Transportation and Production at the Same Time in a Special FMS.

Lipschitz and differentiability properties of quasi-concave and singular normal distribution functions

Improvements of the Giaccardi and the Petrović inequality and related Stolarsky type means

Interval solutions for interval algebraic equations

Summary Notes on Maximization

2 Chance constrained programming

The games SEKI and D-SEKI

R u t c o r Research R e p o r t. On Acyclicity of Games with Cycles. Daniel Andersson a Vladimir Gurvich b Thomas Dueholm Hansen c

BALANCING GAUSSIAN VECTORS. 1. Introduction

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

Dual-Bounded Generating Problems: Weighted Transversals of a Hypergraph

ECON 5111 Mathematical Economics

Bulletin of the. Iranian Mathematical Society

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

A Residual Existence Theorem for Linear Equations

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Math 290, Midterm II-key

The Newton Bracketing method for the minimization of convex functions subject to affine constraints

1 Strict local optimality in unconstrained optimization

Math 121 Homework 4: Notes on Selected Problems

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

An Introduction to Linear Matrix Inequalities. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Chance constrained optimization - applications, properties and numerical issues

A Tropical Extremal Problem with Nonlinear Objective Function and Linear Inequality Constraints

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

USA Mathematical Talent Search Solutions to Problem 4/4/16

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Complexity of Bilevel Coherent Risk Programming

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as

Tropical Polynomials

APPROXIMATE METHODS FOR CONVEX MINIMIZATION PROBLEMS WITH SERIES PARALLEL STRUCTURE

Optimization Problems with Linear Chance Constraints Structure, Numerics and Applications. R. Henrion

1 Last time: inverses

Chapter 2: Linear Programming Basics. (Bertsimas & Tsitsiklis, Chapter 1)

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Transcription:

R u t c o r Research R e p o r t A CONVEXITY THEOREM IN PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS András Prékopa a Mine Subasi b RRR 32-2007, December 2007 RUTCOR Rutgers Center for Operations Research Rutgers University 640 Bartholomew Road Piscataway, New Jersey 08854-8003 Telephone: 732-445-3804 Telefax: 732-445-5472 Email: rrr@rutcorrutgersedu http://rutcorrutgersedu/ rrr a RUTCOR, Rutgers Center for Operations Research, 640 Bartholomew Road Piscataway, NJ 08854-8003, USA Email: prekopa@rutcorrutgersedu b RUTCOR, Rutgers Center for Operations Research, 640 Bartholomew Road Piscataway, NJ 08854-8003, USA Email: msub@rutcorrutgersedu

Rutcor Research Report RRR 31-2007, DECEMBER, 2007 A CONVEXITY THEOREM IN PROGRAMMING UNDER PROBABILISTIC CONSTRAINTS András Prékopa Mine Subasi AbstractPrékopa (1973, 1995) has proved that if T is an r n random matrix with independent, normally distributed rows such that their covariance matrices are constant multiples of each other, then the function h(x) = P (T x b) is quasi-concave in R n, where b is a constant vector We prove that, under same condition, the converse is also true, a special quasi-concavity of h(x) implies the above-mentioned property of the covariance matrices

RRR 31-2007 Page 1 1 Introduction In the theory of programming under probabilistic constraints an important problem is the convexity of the set D = {x h(x) p}, where h(x) = P (T x ξ), p is a fixed probability (0 < p < 1), T a random matrix and ξ a random vector For constant T and continuously distributed random ξ with logconcave pdf Prékopa (1971, 1973) has proved that h(x) is also a logconcave function This fact clearly implies the convexity of the set D For the case of a random technology matrix a few convexity theorems are also known One of them is the following Theorem 1 Let ξ be constant and T = a random matrix with independent, normally distributed rows such that their covariance matrices are constant multiples of each other Then h(x) is a quasi-concave function in R n The quasi-concavity of h(x) implies the convexity of the set D Similar theorem holds if the columns of T satisfy the condition that their covariance matrices are constant multiples of each other The right hand side vector ξ may also be random In this case we assume that it is independent of T and its covariance matrix is a constant multiple of any of the other covariance matrices Recently Henrion (2006) proved the convexity of the set D for independent T 1,, T r under same condition The purpose of the paper is to show that under same conditions the converse of Theorem 1 is also true While the sum of concave functions is also concave, the same is not true, in general, for quasi-concave functions However, we can define a special class of quasi-concave functions such that the sums and products, within the class, are also quasi-concave Definition Let h 1 (x),, h r (x) be quasi-concave functions in a convex set E We say that they are uniformly quasi-concave if for any x, y E either or T 1 T r min(h i (x), h i (y)) = h i (x), i = 1,, r min(h i (x), h i (y)) = h i (y), i = 1,, r Obviously, the sum of uniformly quasi-concave functions, on the same set, is also quasiconcave and if the functions are also nonnegative, then the same holds for their product as well The latter property is used in the next section, where we prove our main result

Page 2 RRR 31-2007 2 The Main Theorem First let r = 1 and consider the function h(x) = P (T x b), where T is a random vector and b is a constant The following lemma was first proved by Kataoka and van de Panne and Popp (1963) See also Prékopa (1965) Lemma If T has normal distribution, then the function h(x) is quasi-concave on the set { x P (T x b) 1 } 2 Proof We prove the equivalent statement: for any p 1 2 the set {x P (T x b) p } (21) is convex Let Φ designate the cdf of the N(0, 1)-distribution and let µ = E(T ) For any x that satisfies x T Cx > 0, h(x) = P (T x b) ( (T µ)x = P xt Cx b µx ) xt Cx ( ) b µx = Φ p xt Cx is equivalent to µ T x + Φ 1 (p) x T Cx b (22) Since Φ 1 (p) 0 and x T Cx is a convex function in R n, it follows that the inequality (22) determines a convex set Let r be an arbitrary positive integer and introduce the function: h i (x) = P (T i x b i ), i = 1,, r If µ i = E(T i ) = 0, i = 1,, r and p 1/2, then the inequality P (T i x b i ) P (T x b) p shows that we have to assume b i 0, i = 1,, r, otherwise the set (21) is empty So, let b i 0, i = 1,, r We also see that, under the same condition, r E = {x P (T i x b i ) p} (23) i=1

RRR 31-2007 Page 3 { r = x x T Cx i=1 Each function h i is quasi-concave on the set (23) b } i Φ 1 (p) Theorem 2 Let b i > 0, i = 1,, r and E(T i ) = 0, i = 1,, r If the functions h 1, h r are uniformly quasi-concave on the set E, defined by (23), and C i 0, i = 1,, r, then each C i is a constant multiple of a covariance matrix C Proof We already know that the functions h 1,, h r are all quasi-concave on the set (23) and that ( ) b i h i (x) = Φ, i = 1,, r xt Cx It is enough to show that if we take two functions h i, h j, i j, then the corresponding covariance matrices C i, C j are constant multiples of each other Let h 1 and h 2 be the two functions Since h 1 and h 2 are uniformly quasi-concave on E, it follows that for any two vectors y, z E, the inequality Φ b ( 1 Φ y T C 1 y b 1 zt C 1 z Φ b ( ) 2 b Φ 2 y T C 2 y zt C 2 z An equivalent form of the statement is that the inequality ) z T C 1 z y T C 1 y (24) z T C 2 z y T C 2 y (25) From linear algebra we know that for any two quadratic forms, in the same variables, there exists a basis such that both quadratic forms are sums of squares if the variables are expressed in that basis In our case this means that there exist linearly independent vectors a 1,, a n such that the transformation x = a 1 u 1 + + a n u n = Au, applied to the quadratic forms x T C 1 x and x T C 2 x, takes them to the forms x T C 1 x = u T A T C 1 Au = λ 1 u 2 1 + + λ n u 2 n x T C 2 x = u T A T C 2 Au = γ 1 u 2 1 + + γ n u 2 n, where λ i, γ i 0, i = 1,, n and λ 1 + + λ n > 0, γ 1 + + γ n > 0

Page 4 RRR 31-2007 The transformation x = Au transforms the set E into a set H that is the intersection of ellipsoids with centers in the origin and main axes lying in the coordinate axes If u = A 1 y and v = A 1 z, then the statement that (24) implies (25) can be formulated in such a way that if u, v H, then λ 1 v 2 1 + + λ n v 2 n λ 1 u 2 1 + + λ n u 2 n This, in turn, is the same as the statement: γ 1 v 2 1 + + γ n v 2 n γ 1 u 2 1 + + γ n u 2 n λ 1 (v 2 1 u 2 1) + + λ n (v 2 n u 2 n) 0 γ 1 (v 2 1 u 2 1) + + γ n (v 2 n u 2 n) 0 Let us introduce the notation w i = v 2 i u 2 i, i = 1,, n Then a further form of the statement is: λ 1 w 1 + + λ n w n 0 (26) γ 1 w 1 + + γ n w n 0 (27) The above implication is true for any w = (w 1,, w n ) in an open convex set around the origin in R n It follows that it is also true without any limitation for the variables w 1,, w n By Farkas theorem there exists a nonnegative number α such that γ = (γ 1,, γ n ) = α(λ 1,, λ n ) = αλ (28) Since γ 0, λ 0, the number α must be positive The relation can be written in matrix form: γ 1 0 λ 1 0 = α (29) 0 γ n 0 λ n If we take into account that λ 1 0 0 λ n γ 1 0 0 γ n = AT C 1 A and combine it with (29), we can derive the equation = AT C 2 A, A T C 2 A = α A T C 1 A

RRR 31-2007 Page 5 Since A is a nonsingular matrix, we conclude that C 2 = α C 1 This proves the theorem References [1] R Henrion and C Strugarek, Convexity of Chance Constraints with Independent Random Variables, Stochastic Programming E-Print Series (SPEPS) 9 (2006), to appear in: Computational Optimization and Applications [2] A Prékopa, On logarithmic concave measures with application to stochastic programming, Acta Sci Math (Szeged) 32 (1971), 301-316 [3] A Prékopa, On logarithmic concave measures and functions, Acta Sci Math (Szeged) 34 (1973), 335 343 [4] A Prékopa, Programming under probabilistic constraints with random technology matrix, Mathematische Operationsforschung und Statistik 5 (1974), 109-116 [5] A Prékopa, Stochastic Programming, Kluwer Academic Publishers, Dordtecht, Boston, 1995 [6] S Kataoka, A stochastic programming model, Econometrica 31, (1963), 181 196 [7] C van de Panne and W Popp, Minimum-cost cattle feed under probabilistic protein constraints, Management Sciences 9 (1963) 405 430