EE 227A: Convex Optimization and Applications April 24, 2008

Similar documents
Handout 8: Dealing with Data Uncertainty

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Robust Optimization for Risk Control in Enterprise-wide Optimization

Handout 6: Some Applications of Conic Linear Programming

Estimation and Optimization: Gaps and Bridges. MURI Meeting June 20, Laurent El Ghaoui. UC Berkeley EECS

4. Convex optimization problems

Convex optimization problems. Optimization problem in standard form

Robust Optimization: Applications in Portfolio Selection Problems

Chance constrained optimization

Lecture: Convex Optimization Problems

LIGHT ROBUSTNESS. Matteo Fischetti, Michele Monaci. DEI, University of Padova. 1st ARRIVAL Annual Workshop and Review Meeting, Utrecht, April 19, 2007

Optimization Tools in an Uncertain Environment

4. Convex optimization problems

Short Course Robust Optimization and Machine Learning. Lecture 6: Robust Optimization in Machine Learning

Reinforcement of gas transmission networks with MIP constraints and uncertain demands 1

Distributionally Robust Optimization with ROME (part 1)

Robust linear optimization under general norms

Introduction to Convex Optimization

Convex Optimization in Classification Problems

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

Chance Constrained Programming

Mathematical Optimization Models and Applications

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

Chapter 2: Linear Programming Basics. (Bertsimas & Tsitsiklis, Chapter 1)

Robust portfolio selection under norm uncertainty

1 Robust optimization

2 Chance constrained programming

Second Order Cone Programming, Missing or Uncertain Data, and Sparse SVMs

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

Robust Growth-Optimal Portfolios

Stochastic Subgradient Methods

IE 521 Convex Optimization

On deterministic reformulations of distributionally robust joint chance constrained optimization problems

Convex Optimization and Modeling

A Geometric Characterization of the Power of Finite Adaptability in Multistage Stochastic and Adaptive Optimization

Combinatorial Data Mining Method for Multi-Portfolio Stochastic Asset Allocation

Threshold Boolean Form for Joint Probabilistic Constraints with Random Technology Matrix

Applications of Robust Optimization in Signal Processing: Beamforming and Power Control Fall 2012

Lecture 6: Conic Optimization September 8

A Geometric Characterization of the Power of Finite Adaptability in Multi-stage Stochastic and Adaptive Optimization

Distributionally Robust Discrete Optimization with Entropic Value-at-Risk

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

Advances in Convex Optimization: Theory, Algorithms, and Applications

Scenario Grouping and Decomposition Algorithms for Chance-constrained Programs

A Minimax Theorem with Applications to Machine Learning, Signal Processing, and Finance

Distributionally Robust Convex Optimization

Modeling Uncertainty in Linear Programs: Stochastic and Robust Linear Programming

Practical Robust Optimization - an introduction -

Reformulation of chance constrained problems using penalty functions

Robust Portfolio Selection Based on a Multi-stage Scenario Tree

CS295: Convex Optimization. Xiaohui Xie Department of Computer Science University of California, Irvine

Lecture 7: Weak Duality

Lecture: Cone programming. Approximating the Lorentz cone.

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

Lecture: Examples of LP, SOCP and SDP

Handout 6: Some Applications of Conic Linear Programming

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Branch-and-cut Approaches for Chance-constrained Formulations of Reliable Network Design Problems

Optimization Problems with Probabilistic Constraints

Optimisation and Operations Research

An Adaptive Partition-based Approach for Solving Two-stage Stochastic Programs with Fixed Recourse

Canonical Problem Forms. Ryan Tibshirani Convex Optimization

A Robust Optimization Perspective on Stochastic Programming

Robust Combinatorial Optimization under Convex and Discrete Cost Uncertainty

Discrete Optimization

Interval Data Classification under Partial Information: A Chance-constraint Approach

Stochastic Optimization One-stage problem

Pareto Efficiency in Robust Optimization

Theory and applications of Robust Optimization

Robust optimization to deal with uncertainty

16.1 L.P. Duality Applied to the Minimax Theorem

1 Column Generation and the Cutting Stock Problem

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Robust Model Predictive Control through Adjustable Variables: an application to Path Planning

WORST-CASE VALUE-AT-RISK AND ROBUST PORTFOLIO OPTIMIZATION: A CONIC PROGRAMMING APPROACH

Chance constrained optimization - applications, properties and numerical issues

Optimization (168) Lecture 7-8-9

LP Relaxations of Mixed Integer Programs

Dealing with Uncertainty

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

On Distributionally Robust Chance-Constrained Linear Programs 1

The Value of Adaptability

EE364a Review Session 5

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

Robust combinatorial optimization with variable budgeted uncertainty

Lecture 1: Introduction

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

SOME HISTORY OF STOCHASTIC PROGRAMMING

1 Review Session. 1.1 Lecture 2

Robustness and Regularization: An Optimization Perspective

LECTURE NOTE #8 PROF. ALAN YUILLE. Can we find a linear classifier that separates the position and negative examples?

Assignment 1: From the Definition of Convexity to Helley Theorem

Farkas Lemma, Dual Simplex and Sensitivity Analysis

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

Branch-and-cut (and-price) for the chance constrained vehicle routing problem

Ambiguous Chance Constrained Programs: Algorithms and Applications

Data-driven Chance Constrained Stochastic Program

Probability Theory Overview and Analysis of Randomized Algorithms

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Transcription:

EE 227A: Convex Optimization and Applications April 24, 2008 Lecture 24: Robust Optimization: Chance Constraints Lecturer: Laurent El Ghaoui Reading assignment: Chapter 2 of the book on Robust Optimization 1 24.1 Stochastic Optimization Example 1: Stochastic LP minc T x (24.1) st : A(w i )x b(w i ), i where A(w i ),b(w i ) are realizations of the random variables A,b. Example 2: Stochastic portfolio optimization maxe(y 1 u 1 +y 2 u 2 1 u 1,u 2 2 2 1u 2 1 1 2 2 1u 2 1 c 1 u 1 u 0 c 2 u 2 u 1 ) (24.2) where y 1,y 2 are random variables. The problem is maximizing the expectation. The stochastic recourse problem, however, becomes complicated. (Taking the expectation of the objective function in (24.11), E(c 2 (u ˆ 2 +U 2 y 1 ) u 1 ) usually is not in a simple form.) 24.2 Chance Optimization Consider the LP minc T x (24.3) st : a T i x b i where a i N(ˆ i,σ i ). We can formulate a chance constrained LP minc T x (24.4) st : Pr{a T i x b i } 1 ǫ If we take into account the correlation between a i s, the problem is minc T x (24.5) st : Pr{Ax b} 1 ǫ 1 A. Ben Tal, L. El Ghaoui, A. Nemirovski, Robust Optimization. 24-1

where A is the random variable. If c is also a random variable, the problem can be formulated as min t (24.6) st : Pr{t c T x} 1 ǫ,pr{ax b} 1 ǫ 24.3 Robust Optimization topics In the following lectures, we will cover the following topics about Robust Optimization Form efficient convex approximation Chance constraints relaxation (including the case with partially known distribution) Evaluate quality of relaxation Recourse programming 24.4 Chance Constraints Chance constraints are a probabilistic way of handling probabilistic uncertainty. We would like to convert chance constraints into robustness constraints, which are easier to deal with. Consider the LP: Our basic problem with chance constraints is: min c T x (24.7) s.t. Ax b (24.8) min c T x (24.9) s.t. Prob{Ax b} 1 ǫ (24.10) The technique we use to simplify the the chance constraints is to assume no correlation between rows of the A-matrix, which allows us to write the chance constraint as: Prob{a T i x b i } 1 ǫ, i = 1,2,...m (24.11) This is a less general model, but it is often much easier to solve. The question now becomes: how do we handle such scalar chance constraints? We will consider a constraint of the form Prob{a T x b} 1 ǫ, where a,b are random, and say that x is ǫ-reliable if it satisfies the above constraint. 24-2

24.5 Example: Gaussian Case In this particular case, this scalar constraint is easy to handle. Suppose a N(â,Σ), then for fixed x: ξ = a T x b N(â T x b,x T Σx) (24.12) Scaling with ξ = ˆξ +u, where ˆξ = â T x b, 2 = x T Σx, and u N(0,1), gives that: Prob{ξ 0} = Prob Defining Erf(α) = α { u ˆξ } = ˆξ 1 2π exp( u2 /2)du 1 ǫ (24.13) 1 2π exp( u2 /2)du, we can rewrite the inequality above as Erf ǫ. Note that Erf( ) is a decreasing function and that Erf(0) = 1/2. Further simplifying our expression, we get that: Erf ( ˆξ ) ǫ ˆξ ( ˆξ ) Erfinv(ǫ) =: ψ(ǫ) (24.14) ˆξ +ψ(ǫ) 0 (24.15) â T x+ψ(ǫ) x T Σx b (24.16) Note that ψ(ǫ) > 0 if ǫ < 1/2, and ψ(ǫ) is like a coefficient that can be precomputed. Consequently, the inequality above is a SOCP constraint. Making ǫ smaller causes ψ(ǫ) to become larger, which makes the constraint harder to satisfy. Also, we can interpret ψ(ǫ) x T Σx as a risk term and â T x as an average term. Besides this case, there are very few cases where we can do symbolic calculations. In practice, assuming Gaussian uncertainty is fine for most cases, even when the randomness is not Gaussian. Also, notice that the SOCP constraint derived above is the same as the one obtained using a robust optimization approach with an uncertainty model. Indeed, suppose that all we know is that: a U := {â+σ 1/2 u u 2 ψ(ǫ)} (24.17) Thenn, we have that a T x b, a U if and only if â T x+ψ(ǫ) x T Σx. This uncertainty model seems very restrictive, but it can be interpreted as a probabilistic approach; it is fairly reasonable. 24-3

24.6 Case with Independent, Bounded Random Uncertainties Suppose that we have a vector uncertainty ξ R L, which is a random variable, such that the coefficients ξ l s are mutually independent, and E(ξ l ) = 0, and ξ l 1, l. The last constraint is a deterministic bound, which tells us about the support of the random variable ξ. We consider the chance constraint Prob{a(ξ) T x b(ξ)} 1 ǫ (24.18) where a,b depend on the random variable ξ in an affine fashion, precisely a(ξ) = a 0 +Σξ i a i and b(ξ) = b 0 +Σξ i b i, where vector a i and scalars b i, i = 0,...,L, are given. In the case that ξ i = 1 with probability 1/2 and ξ i = 1 with probability 1/2, computing theprobabilityin(24.29)forfixedxcanbeshowntobenp-hard, andsoitisquitereasonable to use an approximation, precisely, a sufficient condition for the chance constraint to hold. Our chance constraint can be rewritten as Prob{ξ T z > b 0 a T ox} ǫ. (24.19) We will show later the following lemma. Let z R L be a deterministic vector, and ξ 1,...,ξ L be independent random variables with zero mean taking values in [ 1,1]. Then Prob{ξ T z > Ω z 2 } exp( Ω 2 /2). Applying the lemma with z i = a T i x b i, we see that the constraint b 0 a T ox Ω L (a T i x b i) 2 (24.20) implies that i=1 Prob{ξ T z > b 0 a T ox} Prob{ξ T z > Ω z 2 } ǫ, provided Ω 2log(1/ǫ) Thus, the SOC condition (24.31) is a sufficient for our chance constraint (24.30) to hold. Again, the conic constraint (24.31) can be reinterpreted as a robust condition, against ellipsoidal uncertainties: a T (ξ)x b(ξ), ξ U (24.21) U = {ξ ξ 2 Ω} (24.22) Define B Ω = {ξ ξ 2 Ω} and Box 1 = {ξ ξ 1}. We can impose robustness against a Box uncertainty: a T (ξ)x b(ξ), ξ Box 1 b 0 a T 0x+Σ L i=1 a T i x b i 24-4

This condition gives us 100% reliability, while a ball uncertainty, which corresponds to the SOCP condition (24.31) gives a reliability of ǫ. If Ω 7.44, then exp( Ω 2 /2) 10 12 = ǫ, so we get huge reliability for small Ω. In high dimensions, the ball is much smaller in volume than the box, precisely: volb Ω volbox 1 = ( Ω eπ/2 m ) m m 0 (24.23) Hence, in high dimensions, it is better to use B Ω : although it does not give 100% reliability, it does provide a highly reliable solution. A solution robust against a set that is much smaller in volume than the box is still highly reliable over the entire box. We can even arrange that the set over which we enforce a robustness condition is actually entirely included in the original box (this is not the case with the previous setting). Indeed, consider ξ U := Box 1 B Ω. It turns out that an equivalent representation of the robustness constraint a(ξ) T x b(ξ), ξ U is: there exist z,w such that z 1 +Ω Σwi 2 b 0 a T 0, z i +w i = b i a T i x, i. (24.24) Now this robustness constraint still guarantees a high reliability: If x is feasible for the above robustness condition, and infeasible for the constraint a(ξ) T x b(ξ) for some realization ξ, then Σξ i (a T i x b i ) > b 0 a T 0x Σz i ξ i Σw i ξ i > b 0 a T 0x Σ z i Σw i ξ i > b 0 a T 0x Σw i ξ i > Ω Σw 2 i Consequently, we have that Prob{a(ξ) T x > b(ξ)} Prob{ w T ξ > Ω w 2 } exp( Ω 2 /2). When m > Ω 2, ξ takes only ±1 values, the set Box 1 B Ω does not even contain a single realization of ξ. Still, we can find a highly reliable solution by enforcing the robustness constraint (24.35). This illustrates that a solution that is robust with respect to an uncertainty set that not only is smaller, but is actually contained in the original box, is actually still highly reliable. 24.7 Example: Investment Problem We have several assets including cash (or, risk-free asset) and stocks, and denote by r the n + 1 vector of returns over the investment period, with r 0 being the return of cash (say, r 0 = 1.05 for a 5% CD). How should we distribute some sum (say, one dollar) over these assets, where we take the returns to be a random variable? We invest amounts y 0,y 1,...,y n, such that Σy i = 1 and y i 0. The total return of the portfolio is given by r 0 y 0 +Σr i y i t. We make the further assumption that the returns are random in the form r i = µ i + i ξ i, where µ i is the mean and i is the standard deviation. 24-5

We have two models, the first is the robust case where ξ Box 1 and the second is the ball-box case where ξ Box 1 B Ω. Take the tolerance of Ω = 3.255 or ǫ = 0.005 and use µ i = 1.05+0.3(200 i)/199 and i = 0.05+0.6(200 i)/199. Using the first model, we get y 0 = 1 and y i = 0, which means that a purely robust approach leads to the very conservative investment of putting everything in a risk-free asset. The second model gives a worst-case return of 10.1% with a 0.5% chance of not getting this return. We thus observe that chance constraints allow to make the robustness condition much less conservative, at the expense of a very slight increase in risk. 24-6