Convex optimization problems. Optimization problem in standard form

Similar documents
4. Convex optimization problems

Lecture: Convex Optimization Problems

4. Convex optimization problems

Convex Optimization. 4. Convex Optimization Problems. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University

4. Convex optimization problems (part 1: general)

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Convex Optimization and Modeling

A Brief Review on Convex Optimization

Tutorial on Convex Optimization for Engineers Part II

Convex Optimization Problems. Prof. Daniel P. Palomar

12. Interior-point methods

Lecture: Examples of LP, SOCP and SDP

Optimisation convexe: performance, complexité et applications.

Convex Optimization and l 1 -minimization

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

CS675: Convex and Combinatorial Optimization Fall 2016 Convex Optimization Problems. Instructor: Shaddin Dughmi

Exercises. Exercises. Basic terminology and optimality conditions. 4.2 Consider the optimization problem

8. Geometric problems

Semidefinite Programming Basics and Applications

8. Geometric problems

12. Interior-point methods

Convex Optimization in Communications and Signal Processing

Chapter 2: Linear Programming Basics. (Bertsimas & Tsitsiklis, Chapter 1)

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Lecture 4: Linear and quadratic problems

Convex Optimization M2

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

EE364 Review Session 4

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

5. Duality. Lagrangian

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

ELE539A: Optimization of Communication Systems Lecture 6: Quadratic Programming, Geometric Programming, and Applications

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

9. Geometric problems

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Convex Optimization in Classification Problems

Convex Optimization & Lagrange Duality

Lecture: Duality.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Constrained Optimization and Lagrangian Duality

Lecture Note 5: Semidefinite Programming for Stability Analysis

EE5138R: Problem Set 5 Assigned: 16/02/15 Due: 06/03/15

CS295: Convex Optimization. Xiaohui Xie Department of Computer Science University of California, Irvine

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

subject to (x 2)(x 4) u,

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

The Q-parametrization (Youla) Lecture 13: Synthesis by Convex Optimization. Lecture 13: Synthesis by Convex Optimization. Example: Spring-mass System

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18

MOSEK Modeling Cookbook Release 3.0. MOSEK ApS

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

Linear and non-linear programming

Lecture: Duality of LP, SOCP and SDP

CS-E4830 Kernel Methods in Machine Learning

minimize x subject to (x 2)(x 4) u,

Lecture 10. ( x domf. Remark 1 The domain of the conjugate function is given by

FRTN10 Multivariable Control, Lecture 13. Course outline. The Q-parametrization (Youla) Example: Spring-mass System

ICS-E4030 Kernel Methods in Machine Learning

Nonlinear Programming Models

Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications

Analytic Center Cutting-Plane Method

Chapter 2: Preliminaries and elements of convex analysis

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Preliminaries Overview OPF and Extensions. Convex Optimization. Lecture 8 - Applications in Smart Grids. Instructor: Yuanzhang Xiao

Agenda. 1 Cone programming. 2 Convex cones. 3 Generalized inequalities. 4 Linear programming (LP) 5 Second-order cone programming (SOCP)

Conic Linear Programming. Yinyu Ye

EE/AA 578, Univ of Washington, Fall Duality

Lecture 9 Sequential unconstrained minimization

Advances in Convex Optimization: Theory, Algorithms, and Applications

Nonlinear Optimization: The art of modeling

Convex functions. Definition. f : R n! R is convex if dom f is a convex set and. f ( x +(1 )y) < f (x)+(1 )f (y) f ( x +(1 )y) apple f (x)+(1 )f (y)

CHAPTER 3: OPTIMIZATION

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

What can be expressed via Conic Quadratic and Semidefinite Programming?

Lecture 2: Convex functions

Problems in VLSI design

Lecture 6: Conic Optimization September 8

Lecture 7: Convex Optimizations

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Homework Set #6 - Solutions

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Part I Formulation of convex optimization problems

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

Conic Linear Programming. Yinyu Ye

Additional Homework Problems

Introduction and Math Preliminaries

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

Course Outline. FRTN10 Multivariable Control, Lecture 13. General idea for Lectures Lecture 13 Outline. Example 1 (Doyle Stein, 1979)

Convex Optimization M2

Convex Functions. Pontus Giselsson

Transcription:

Convex optimization problems optimization problem in standard form convex optimization problems linear optimization quadratic optimization geometric programming quasiconvex optimization generalized inequality constraints semidefinite programming vector optimization IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 1 Optimization problem in standard form minimize f 0 (x) subject to f i (x) 0, i = 1,..., m h i (x) = 0, i = 1,..., p x R n is the optimization variable f 0 : R n R is the objective or cost function f i : R n R, i = 1,..., m, are the inequality constraint functions h i : R n R are the equality constraint functions optimal value: p = inf{f 0 (x) f i (x) 0, i = 1,..., m, h i (x) = 0, i = 1,..., p} p = if problem is infeasible (no x satisfies the constraints) p = if problem is unbounded below IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 2

Optimal and locally optimal points x is feasible if x dom f 0 and it satisfies the constraints a feasible x is optimal if f 0 (x) = p ; X opt is the set of optimal points x is locally optimal if there is an R > 0 such that x is optimal for minimize (over z) f 0 (z) subject to f i (z) 0, i = 1,..., m, h i (z) = 0, i = 1,..., p z x 2 R examples (with n = 1, m = p = 0) f 0 (x) = 1/x, dom f 0 = R ++ : p = 0, no optimal point f 0 (x) = log x, dom f 0 = R ++ : p = f 0 (x) = x log x, dom f 0 = R ++ : p = 1/e, x = 1/e is optimal f 0 (x) = x 3 3x, p =, local optimum at x = 1 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 3 Implicit constraints The standard form optimization problem has an implicit constraint m p x D = dom f i dom h i, i=0 i=1 we call D the domain of the problem the constraints f i (x) 0, h i (x) = 0 are the explicit constraints a problem is unconstrained if it has no explicit constraints (m = p = 0) example: minimize f 0 (x) = k i=1 log(b i a T i x) is an unconstrained problem with implicit constraints a T i x < b i, i = 1,..., k IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 4

Feasibility problem find x subject to f i (x) 0, i = 1,..., m h i (x) = 0, i = 1,..., p can be considered a special case of the general problem with f 0 (x) = 0: minimize 0 subject to f i (x) 0, i = 1,..., m h i (x) = 0, i = 1,..., p p = 0 if constraints are feasible; any feasible x is optimal p = if constraints are infeasible IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 5 Convex optimization problem standard form convex optimization problem minimize f 0 (x) subject to f i (x) 0, i = 1,..., m ai T x = b i, i = 1,..., p f 0, f 1,..., f m are convex; equality constraints are affine problem is quasiconvex if f 0 is quasiconvex (and f 1,..., f m convex) often written as minimize f 0 (x) subject to f i (x) 0, i = 1,..., m Ax = b important property: feasible set of a convex optimization problem is convex IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 6

example minimize f 0 (x) = x1 2 + x 2 2 subject to f 1 (x) = x 1 /(1 + x2 2) 0 h 1 (x) = (x 1 + x 2 ) 2 = 0 f 0 is convex; feasible set {(x 1, x 2 ) x 1 = x 2 0} is convex not a convex problem (according to our definition): f 1 is not convex, h 1 is not affine equivalent (but not identical) to the convex problem minimize x 2 1 + x 2 2 subject to x 1 0 x 1 + x 2 = 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 7 Local and global optima Any locally optimal point of a convex problem is (globally) optimal proof: suppose x is locally optimal and y is optimal with f 0 (y) < f 0 (x). x locally optimal means there is an R > 0 such that z feasible, z x 2 R = f 0 (z) f 0 (x). Consider z = θy + (1 θ)x with θ = R/(2 y x 2 ) y x 2 > R, so 0 < θ < 1/2 z is a convex combination of two feasible points, hence also feasible z x 2 = R/2 and f 0 (z) θf 0 (x) + (1 θ)f 0 (y) < f 0 (x), which contradicts our assumption that x is locally optimal IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 8

Optimality criterion for differentiable f 0 x is optimal if and only if it is feasible and f 0 (x) T (y x) 0 for all feasible y X x f 0 (x) if nonzero, f 0 (x) defines a supporting hyperplane to feasible set X at x IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 9 Optimality criterion: special cases unconstrained problem: x is optimal if and only if x dom f 0, f 0 (x) = 0 equality constrained problem minimize f 0 (x) subject to Ax = b x is optimal if and only if there exists a ν such that x dom f 0, Ax = b, f 0 (x) + A T ν = 0 minimization over nonnegative orthant x is optimal if and only if minimize f 0 (x) subject to x 0 x dom f 0, x 0, { f0 (x) i 0 x i = 0 f 0 (x) i = 0 x i > 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 10

Equivalent convex problems two problems are (informally) equivalent if the solution of one is readily obtained from the solution of the other, and vice-versa. Some common transformations that preserve convexity: eliminating equality constraints is equivalent to minimize f 0 (x) subject to f i (x) 0, i = 1,..., m Ax = b minimize (over z) f 0 (Fz + x 0 ) subject to f i (Fz + x 0 ) 0, i = 1,..., m where F and x 0 are such that Ax = b x = Fz + x 0 for some z IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 11 introducing equality constraints is equivalent to minimize f 0 (A 0 x + b 0 ) subject to f i (A i x + b i ) 0, i = 1,..., m minimize (over x, y i ) f 0 (y 0 ) subject to f i (y i ) 0, i = 1,..., m y i = A i x + b i, i = 0, 1,..., m introducing slack variables for linear inequalities is equivalent to minimize f 0 (x) subject to ai T x b i, i = 1,..., m minimize (over x, s) f 0 (x) subject to ai T x + s i = b i, i = 1,..., m s i 0, i = 1,... m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 12

epigraph form: standard form convex problem is equivalent to minimize (over x, t) t subject to f 0 (x) t 0 f i (x) 0, i = 1,..., m Ax = b minimizing over some variables is equivalent to minimize f 0 (x 1, x 2 ) subject to f i (x 1 ) 0, i = 1,..., m minimize f0 (x 1 ) subject to f i (x 1 ) 0, i = 1,..., m where f 0 (x 1 ) = inf x2 f 0 (x 1, x 2 ) IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 13 Linear program (LP) minimize subject to c T x + d Gx h Ax = b convex problem with affine objective and constraint functions feasible set is a polyhedron c s P x IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 14

Examples diet problem: choose quantities x 1,..., x n of n foods one unit of food j costs c j, contains amount a ij of nutrient i healthy diet requires nutrient i in quantity at least b i to find cheapest healthy diet, minimize c T x subject to Ax b, x 0 piecewise-linear minimization minimize max i=1,...,m (ai T x + b i ) equivalent to an LP minimize t subject to ai T x + b i t, i = 1,..., m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 15 Chebyshev center of a polyhedron Chebyshev center of P = {x a T i x b i, i = 1,..., m} is center of largest inscribed ball x cheb B = {x c + u u 2 r} s a T i x b i for all x B if and only if sup{a T i (x c + u) u 2 r} = a T i x c + r a i 2 b i hence, x c, r can be determined by solving the LP maximize xc,r r subject to ai T x c + r a i 2 b i, i = 1,..., m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 16

Quadratic program (QP) minimize subject to (1/2)x T Px + q T x + r Gx h Ax = b P S n +, so objective is convex quadratic minimize a convex quadratic function over a polyhedron f 0 (x ) x P IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 17 Examples least-squares minimize Ax b 2 2 analytical solution x = A b (A is pseudo-inverse) can add linear constraints, e.g., l x u linear program with random cost minimize c T x + γx T Σx = E c T x + γ var(c T x) subject to Gx h, Ax = b c is random vector with mean c and covariance Σ hence, c T x is random variable with mean c T x and variance x T Σx γ > 0 is risk aversion parameter; controls the trade-off between expected cost and variance (risk) IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 18

Quadratically constrained quadratic program (QCQP) minimize (1/2)x T P 0 x + q0 T x + r 0 subject to (1/2)x T P i x + qi T x + r i 0, i = 1,..., m Ax = b P i S n +; objective and constraints are convex quadratic if P 1,..., P m S n ++, feasible region is intersection of m ellipsoids and an affine set IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 19 Second-order cone programming minimize f T x subject to A i x + b i 2 ci T x + d i, i = 1,..., m Fx = g (A i R n i n, F R p n ) inequalities are called second-order cone (SOC) constraints: (A i x + b i, c T i x + d i ) second-order cone in R n i +1 for n i = 0, reduces to an LP; if c i = 0, reduces to a QCQP more general than QCQP and LP IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 20

Robust linear programming the parameters in optimization problems are often uncertain, e.g., in an LP minimize c T x subject to ai T x b i, i = 1,..., m, there can be uncertainty in c, a i, b i two common approaches to handling uncertainty (in a i, for simplicity) deterministic model: constraints must hold for all a i E i minimize c T x subject to ai T x b i for all a i E i, i = 1,..., m, stochastic model: a i is random variable; constraints must hold with probability η minimize c T x subject to prob(ai T x b i ) η, i = 1,..., m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 21 Deterministic approach via SOCP choose an ellipsoid as E i : E i = {ā i + P i u u 2 1} (ā i R n, P i R n n ) center is ā i, semi-axes determined by singular values/vectors of P i robust LP minimize c T x subject to ai T x b i a i E i, i = 1,..., m is equivalent to the SOCP minimize c T x subject to āi T x + Pi T x 2 b i, i = 1,..., m (follows from sup u 2 1(ā i + P i u) T x = ā T i x + P T i x 2 ) IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 22

Stochastic approach via SOCP assume a i is Gaussian with mean ā i, covariance Σ i (a i N (ā i, Σ i )) ai T x is Gaussian r.v. with mean āi T x, variance x T Σ i x; hence ( ) prob(ai T b i āi T x x b i ) = Φ Σ 1/2 i x 2 where Φ(x) = (1/ 2π) x e t2 /2 dt is CDF of N (0, 1) robust LP minimize c T x subject to prob(ai T x b i ) η, i = 1,..., m, with η 1/2, is equivalent to the SOCP minimize c T x subject to āi T x + Φ 1 (η) Σ 1/2 i x 2 b i, i = 1,..., m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 23 Geometric programming monomial function f (x) = cx a 1 1 x a 2 2 x a n n, dom f = R n ++ with c > 0; exponent α i can be any real number posynomial function: sum of monomials f (x) = K k=1 geometric program (GP) c k x a 1k 1 x a 2k 2 x a nk n, dom f = R n ++ minimize f 0 (x) subject to f i (x) 1, i = 1,..., m h i (x) = 1, i = 1,..., p with f i posynomial, h i monomial IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 24

Geometric program in convex form change variables to y i = log x i, and take logarithm of cost, constraints monomial f (x) = cx a 1 1 x a n n transforms to log f (e y 1,..., e y n ) = a T y + b (b = log c) posynomial f (x) = K k=1 c kx a 1k 1 x a 2k 2 x a nk n ( K ) log f (e y 1,..., e y n ) = log e at k y+b k k=1 transforms to (b k = log c k ) geometric program transforms to convex problem ( K ) minimize log k=1 exp(at 0k y + b 0k) ( K ) subject to log k=1 exp(at ik y + b ik) 0, i = 1,..., m Gy + d = 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 25 Design of cantilever beam segment 4 segment 3 segment 2 segment 1 N segments with unit lengths, rectangular cross-sections of size w i h i given vertical force F applied at the right end design problem minimize subject to total weight upper & lower bounds on w i, h i upper bound & lower bounds on aspect ratios h i /w i upper bound on stress in each segment upper bound on vertical deflection at the end of the beam variables: w i, h i for i = 1,..., N IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 26 F

Objective and constraint functions total weight w 1 h 1 + + w N h N is posynomial aspect ratio h i /w i and inverse aspect ratio w i /h i are monomials maximum stress in segment i is given by 6iF /(w i h 2 i ), a monomial the vertical deflection y i and slope v i of central axis at the right end of segment i are defined recursively as v i = F 12(i 1/2) Ew i hi 3 + v i+1 y i = F 6(i 1/3) Ew i hi 3 + v i+1 + y i+1 for i = N, N 1,..., 1, with v N+1 = y N+1 = 0 (E is Young s modulus) v i and y i are posynomial functions of w, h IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 27 Formulation as a GP note minimize subject to w 1 h 1 + + w N h N w 1 maxw i 1, h 1 maxh i 1, w min w 1 i 1, i = 1,..., N h min h 1 i 1, i = 1,..., N S 1 maxw 1 i h i 1, S min w i h 1 i 1, i = 1,..., N 6iF σ 1 maxw 1 i y 1 maxy 1 1 h 2 i 1, i = 1,..., N we write w min w i w max and h min h i h max w min /w i 1, w i /w max 1, h min /h i 1, h i /h max 1 we write S min h i /w i S max as S min w i /h i 1, h i /(w i S max ) 1 The number of monomials appearing in y 1 grows approximately as N 2. IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 28

Minimizing spectral radius of nonnegative matrix Perron-Frobenius eigenvalue λ pf (A) Consider (elementwise) positive A R n n that is irreducible: (I + A) n 1 > 0. P-F Theorem: there is a real, positive eigenvalue of A, λ pf, equal to spectral radius max i λ i (A) determines asymptotic growth (decay) rate of A k : A k λ k pf as k alternative characterization: λ pf (A) = inf{λ Av λv for some v 0} minimizing spectral radius of matrix of posynomials minimize λ pf (A(x)), where the elements A(x) ij are posynomials of x equivalent geometric program: minimize subject to variables λ, v, x λ n j=1 A(x) ijv j /(λv i ) 1, i = 1,..., n IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 29 Quasiconvex optimization minimize f 0 (x) subject to f i (x) 0, i = 1,..., m Ax = b with f 0 : R n R quasiconvex, f 1,..., f m convex can have locally optimal points that are not (globally) optimal (x, f 0 (x)) IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 30

convex representation of sublevel sets of f 0 if f 0 is quasiconvex, there exists a family of functions φ t such that: φ t (x) is convex in x for fixed t t-sublevel set of f 0 is 0-sublevel set of φ t, i.e., f 0 (x) t φ t (x) 0 example f 0 (x) = p(x) q(x) with p convex, q concave, and p(x) 0, q(x) > 0 on dom f 0 can take φ t (x) = p(x) tq(x): for t 0, φ t convex in x p(x)/q(x) t if and only if φ t (x) 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 31 quasiconvex optimization via convex feasibility problems φ t (x) 0, f i (x) 0, i = 1,..., m, Ax = b (1) for fixed t, a convex feasibility problem in x if feasible, we can conclude that t p ; if infeasible, t p Bisection method for quasiconvex optimization given l p, u p, tolerance ɛ > 0. repeat 1. t := (l + u)/2. 2. Solve the convex feasibility problem (1). 3. if (1) is feasible, u := t; else l := t. until u l ɛ. requires exactly log 2 ((u l)/ɛ) iterations (where u, l are initial values) IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 32

(Generalized) linear-fractional program linear-fractional program minimize subject to f 0 (x) Gx h Ax = b f 0 (x) = ct x + d e T x + f, dom f 0(x) = {x e T x + f > 0} a quasiconvex optimization problem; can be solved by bisection also equivalent to the LP (variables y, z) minimize subject to c T y + dz Gy hz Ay = bz e T y + fz = 1 z 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 33 Generalized linear-fractional program f 0 (x) = max i=1,...,r ci T x + d i ei T, dom f 0 (x) = {x ei T x+f i > 0, i = 1,..., r} x + f i a quasiconvex optimization problem; can be solved by bisection example: Von Neumann model of a growing economy maximize (over x, x + ) min i=1,...,n x + i /x i subject to x + 0, Bx + Ax x, x + R n : activity levels of n sectors, in current and next period (Ax) i, (Bx + ) i : produced, resp. consumed, amounts of good i x + i /x i : growth rate of sector i allocate activity to maximize growth rate of slowest growing sector IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 34

Generalized inequality constraints convex problem with generalized inequality constraints minimize f 0 (x) subject to f i (x) Ki 0, i = 1,..., m Ax = b f 0 : R n R convex; f i : R n R k i cone K i K i -convex w.r.t. proper same properties as standard convex problem (convex feasible set, local optimum is global, etc.) conic form problem: special case with affine objective and constraints minimize c T x subject to Fx + g K 0 Ax = b extends linear programming (K = R m +) to nonpolyhedral cones IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 35 Semidefinite program (SDP) with F i, G S k minimize c T x subject to x 1 F 1 + x 2 F 2 + + x n F n + G 0 Ax = b inequality constraint is called linear matrix inequality (LMI) includes problems with multiple LMI constraints: for example, x 1 ˆF 1 + + x n ˆF n + Ĝ 0, x 1 F 1 + + x n F n + G 0 is equivalent to single LMI x 1 [ ˆF 1 0 0 F 1 ] +x 2 [ ˆF 2 0 0 F 2 ] + +x n [ ˆF n 0 0 F n ] [ Ĝ 0 + 0 G ] 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 36

LP and SOCP as SDP LP and equivalent SDP LP: minimize c T x subject to Ax b SDP: minimize c T x subject to diag(ax b) 0 (note different interpretation of generalized inequality ) SOCP and equivalent SDP SOCP: minimize f T x subject to A i x + b i 2 c T i x + d i, i = 1,..., m SDP: minimize f [ T x (c T subject to i x + d i )I A i x + b i (A i x + b i ) T ci T x + d i ] 0, i = 1,..., m IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 37 Eigenvalue minimization minimize λ max (A(x)) where A(x) = A 0 + x 1 A 1 + + x n A n (with given A i S k ) equivalent SDP minimize subject to t A(x) ti variables x R n, t R follows from λ max (A) t A ti IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 38

Matrix norm minimization minimize A(x) 2 = ( λ max (A(x) T A(x)) ) 1/2 where A(x) = A 0 + x 1 A 1 + + x n A n (with given A i S p q ) equivalent SDP minimize subject to t [ ti A(x) T A(x) ti ] 0 variables x R n, t R constraint follows from A 2 t A T A t 2 I, t 0 [ ] ti A A T 0 ti IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 39 Convexity of vector-valued functions f : R n R m is K-convex if dom f is convex and f (θx + (1 θ)y) K θf (x) + (1 θ)f (y) for x, y dom f, 0 θ 1 example f : S m S m, f (X ) = X 2 is S m +-convex proof: for fixed z R m, z T X 2 z = Xz 2 2 is convex in X, i.e., z T (θx + (1 θ)y ) 2 z θz T X 2 z + (1 θ)z T Y 2 z for X, Y S m, 0 θ 1 therefore (θx + (1 θ)y ) 2 θx 2 + (1 θ)y 2 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 40

Vector optimization general vector optimization problem minimize (w.r.t. K) f 0 (x) subject to f i (x) 0, i = 1,..., m h i (x) 0, i = 1,..., p vector objective f 0 : R n R q, minimized w.r.t. proper cone K R q convex vector optimization problem minimize (w.r.t. K) f 0 (x) subject to f i (x) 0, i = 1,..., m Ax = b with f 0 K-convex, f 1,..., f m convex IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 41 Optimal and Pareto optimal points set of achievable objective values O = {f 0 (x) x feasible} feasible x is optimal if f 0 (x) is a minimum value of O feasible x is Pareto optimal if f 0 (x) is a minimal value of O nts O P f 0 (x po ) O f 0 (x ) x is optimal x po is Pareto optimal IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 42

Multicriterion optimization vector optimization problem with K = R q + f 0 (x) = (F 1 (x),..., F q (x)) q different objectives F i ; roughly speaking we want all F i s to be small feasible x is optimal if y feasible = f 0 (x ) f 0 (y) if there exists an optimal point, the objectives are noncompeting feasible x po is Pareto optimal if y feasible, f 0 (y) f 0 (x po ) = f 0 (x po ) = f 0 (y) if there are multiple Pareto optimal values, there is a trade-off between the objectives IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 43 Regularized least-squares multicriterion problem with two objectives F 1 (x) = Ax b 2 2, F 2 (x) = x 2 2 15 example with A R 100 10 shaded region is O heavy line is formed by Pareto optimal points F2(x) = x 2 2 10 5 0 0 5 10 15 F 1 (x) = Ax b 2 2 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 44

Risk return trade-off in portfolio optimization minimize (w.r.t. R 2 +) ( p T x, x T Σx) subject to 1 T x = 1, x 0 x R n is investment portfolio; x i is fraction invested in asset i p R n is vector of relative asset price changes; modeled as a random variable with mean p, covariance Σ p T x = E r is expected return; x T Σx = var r is return variance example mean return 15% 10% 5% allocation x 1 0.5 x(4) x(3) x(2) x(1) 0% 0% 10% 20% standard deviation of return 0 0% 10% 20% standard deviation of return IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 45 Scalarization to find Pareto optimal points: choose λ K 0 and solve scalar problem minimize λ T f 0 (x) subject to f i (x) 0, i = 1,..., m h i (x) = 0, i = 1,..., p if x is optimal for scalar problem, then it is Pareto-optimal for vector optimization problem s f 0 (x 1 ) O f λ 0 (x 3 ) 1 f 0 (x 2 ) λ 2 for convex vector optimization problems, can find (almost) all Pareto optimal points by varying λ K 0 IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 46

examples for multicriterion problem, find Pareto optimal points by minimizing positive weighted sum λ T f 0 (x) = λ 1 F 1 (x) + + λ q F q (x) regularized least-squares of page 111 (with λ = (1, γ)) minimize Ax b 2 2 + γ x 2 2 for fixed γ > 0, a least-squares problem risk-return trade-off of page 112 (with λ = (1, γ)) minimize p T x + γx T Σx subject to 1 T x = 1, x 0 for fixed γ > 0, a QP IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4 47