HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Similar documents
Convex Optimization M2

5. Duality. Lagrangian

Linear and non-linear programming

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Lecture: Duality.

A Brief Review on Convex Optimization

Solution to EE 617 Mid-Term Exam, Fall November 2, 2017

CSCI : Optimization and Control of Networks. Review on Convex Optimization

4. Convex optimization problems

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Lecture: Duality of LP, SOCP and SDP

Constrained Optimization and Lagrangian Duality

subject to (x 2)(x 4) u,

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Additional Homework Problems

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Lecture: Convex Optimization Problems

EE/AA 578, Univ of Washington, Fall Duality

Tutorial on Convex Optimization for Engineers Part II

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Convex Optimization & Lagrange Duality

Convex optimization problems. Optimization problem in standard form

Convex Optimization and Modeling

Lecture Note 5: Semidefinite Programming for Stability Analysis

EE364a Review Session 5

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

The proximal mapping

Semidefinite Programming Basics and Applications

Lecture 8. Strong Duality Results. September 22, 2008

4. Convex optimization problems

Part IB Optimisation

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Applications of Linear Programming

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Summer School: Semidefinite Optimization

Lecture 2: Convex Sets and Functions

Assignment 1: From the Definition of Convexity to Helley Theorem

12. Interior-point methods

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Exercises. Exercises. Basic terminology and optimality conditions. 4.2 Consider the optimization problem

Lecture 2: Convex functions

4. Algebra and Duality

Lecture 6: Conic Optimization September 8

10725/36725 Optimization Homework 2 Solutions

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

Lecture 1 Introduction

Lecture 18: Optimization Programming

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 3. Optimization Problems and Iterative Algorithms

17.1 Hyperplanes and Linear Forms

Chapter 2: Preliminaries and elements of convex analysis

Convex analysis and profit/cost/support functions

Lecture 7: Convex Optimizations

Mathematics 530. Practice Problems. n + 1 }

Convex Optimization & Parsimony of L p-balls representation

Lagrangian Duality Theory

Convex Optimization in Communications and Signal Processing

8. Conjugate functions

10-725/ Optimization Midterm Exam

Nonlinear Programming Models

CS-E4830 Kernel Methods in Machine Learning

minimize x subject to (x 2)(x 4) u,

CS675: Convex and Combinatorial Optimization Fall 2016 Convex Optimization Problems. Instructor: Shaddin Dughmi

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

ICS-E4030 Kernel Methods in Machine Learning

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

COM S 578X: Optimization for Machine Learning

Mathematical Analysis Outline. William G. Faris

CO 250 Final Exam Guide

Constrained optimization

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

8. Geometric problems

Math 164-1: Optimization Instructor: Alpár R. Mészáros

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

12. Interior-point methods

Optimization for Machine Learning

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Zero sum games Proving the vn theorem. Zero sum games. Roberto Lucchetti. Politecnico di Milano

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Convex Optimization M2

Lecture: Examples of LP, SOCP and SDP

4. Convex optimization problems (part 1: general)

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Mathematical Preliminaries for Microeconomics: Exercises

Lecture 1: Background on Convex Analysis

Finite Dimensional Optimization Part III: Convex Optimization 1

Characterizing Robust Solution Sets of Convex Programs under Data Uncertainty

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

Transcription:

HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard probability simplex P = {p 1 T p = 1, p 0}. Which of the following conditions are convex in p? (That is, for which of the following conditions is the set of p P that satisfy the condition convex?) 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n p if(a i ). (The function f : R R is given.) 2. Prob(x > α) β. 3. E x 3 αe x. 4. Ex 2 α. 5. Ex 2 α. 6. var(x) α, where var(x) = E(x Ex) 2 is the variance of x. 7. var(x) α. 8. quartile(x) α, where quartile(x) = inf{β Prob(x β) 0.25}. 9. quartile(x) α. Solution 1 We first note that the constraints p i 0, i = 1,..., n, define halfspaces, and n p i = 1 defines a hyperplane, so P is a polyhedron. The first five constraints are, in fact, linear inequalities in the probabilities p i. 1. Ef(x) = n p if(a i ), so the constraint is equivalent to two linear inequalities α p i f(a i ) β. 2. Prob(x α) = i: a i α p i, so the constraint is equivalent to a linear inequality p i β. i: a i α 3. The constraint is equivalent to a linear inequality p i ( a 3 i α a i ) 0. 1

4. The constraint is equivalent to a linear inequality p i a 2 i α. 5. The constraint is equivalent to a linear inequality p i a 2 i α. The first five constraints therefore define convex sets. 6. The constraint var(x) = Ex 2 (Ex) 2 = p i a 2 i ( p i a i ) 2 α is not convex in general. As a counterexample, we can take n = 2, a 1 = 0, a 2 = 1, and α = 1/5. p = (1, 0) and p = (0, 1) are two points that satisfy var(x) α, but the convex combination p = (1/2, 1/2) does not. 7. This constraint is equivalent to a 2 i p i + ( a i p i ) 2 = b T p + p T Ap α where b i = a 2 i and A = aa T. This defines a convex set, since the matrix aa T is positive semidefinite. Let us denote quartile(x) = f(p) to emphasize it is a function of p. 8. The constraint f(p) α is equivalent to Prob(x β) < 0.25 for all β < α. If α a 1, this is always true. Otherwise, define k = max{i a i < α}. This is a fixed integer, independent of p. The constraint f(p) α holds if and only if Prob(x a k ) = k p i < 0.25. This is a strict linear inequality in p, which defines an open halfspace. 9. The constraint f(p) α is equivalent to Prob(x β) 0.25 for all β α. This can be expressed as a linear inequality p i 0.25. (If α a 1, we define k = 0.) i=k+1 2

Exercise 2 (Euclidean distance matrices.) Let x 1,..., x n R k. The matrix D S n defined by D ij = x i x j 2 2 is called a Euclidean distance matrix. It satisfies some obvious properties such as D ij = D ji, D ii = 0, D ij 0, and (from the triangle inequality) D 1/2 ik D 1/2 ij + D 1/2 jk. We now pose the question: When is a matrix D Sn a Euclidean distance matrix (for some points in R k, for some k)? A famous result answers this question: D S n is a Euclidean distance matrix if and only if D ii = 0 and x T Dx 0 for all x with 1 T x = 0. Show that the set of Euclidean distance matrices is a convex cone. Find the dual cone. Solution 2 The set of Euclidean distance matrices in S n is a closed convex cone because it is the intersection of (infinitely many) halfspaces defined by the following homogeneous inequalities: e T i De i 0, e T i De i 0, x T Dx = x j x k D jk 0, j,k for all i = 1,..., n, and all x with 1 T x = 1. It follows that dual cone is given by K = Co({ xx T 1 T x = 1} {e 1 e T 1, e 1 e T 1,..., e n e T n, e n e T n}). This can be made more explicit as follows. Define V R n (n 1) as { 1 1/n i = j V ij = 1/n i j. The columns of V form a basis for the set of vectors orthogonal to 1, i.e., a vector x satisfies 1 T x = 0 if and only if x = V y for some y. The dual cone is K = {V W V T + diag(u) W 0, u R n }. Exercise 3 (Composition rules.) Show that the following functions are convex. 1. f(x) = log( log( m eat i x+b i )) on dom f = {x m eat i x+b i < 1}. You can use the fact that log( n ey i ) is convex. 2. f(x, u, v) = uv x T x on dom f = {(x, u, v) uv > x T x, u, v > 0}. Use the fact that x T x/u is convex in (x, u) for u > 0, and that x 1 x 2 is convex on R 2 ++. Solution 3 1. g(x) = log( m eat i x+b i ) is convex (composition of the log-sum-exp function and an affine mapping), so g is concave. The function h(y) = log y is convex and decreasing. Therefore f(x) = h( g(x)) is convex. 2. We can express f as f(x, u, v) = u(v x T x/u). The function h(x 1, x 2 ) = x 1 x 2 is convex on R 2 ++, and decreasing in each argument. The functions g 1 (u, v, x) = u and g 2 (u, v, x) = v x T x/u are concave. Therefore f(u, v, x) = h(g(u, v, x)) is convex. 3

Exercise 4 (Problems involving l 1 - and l -norms.) Formulate the following problems as LPs. Explain in detail the relation between the optimal solution of each problem and the solution of its equivalent LP. 1. Minimize Ax b (l -norm approximation). 2. Minimize Ax b 1 (l 1 -norm approximation). 3. Minimize Ax b 1 x 1. 4. Minimize x 1 Ax b 1. 5. Minimize Ax b 1 + x. In each problem, A R m n and b R m are given. (See?? for more problems involving approximation and constrained approximation.) Solution 4 Solution. 1. Equivalent to the LP t Ax b t1 Ax b t1. in the variables x, t. To see the equivalence, assume x is fixed in this problem, and we optimize only over t. The constraints say that for each k, i.e., t a T k x b k, i.e., t a T k x b k t t max a T k x b k = Ax b. k Clearly, if x is fixed, the optimal value of the LP is p (x) = Ax b. Therefore optimizing over t and x simultaneously is equivalent to the original problem. 2. Equivalent to the LP 1 T s Ax b s Ax b s. Assume x is fixed in this problem, and we optimize only over s. The constraints say that s k a T k x b k s k for each k, i.e., s k a T k x b k. The objective function of the LP is separable, so we achieve the optimum over s by choosing s k = a T k x b k, and obtain the optimal value p (x) = Ax b 1. Therefore optimizing over t and s simultaneously is equivalent to the original problem. 4

3. Equivalent to the LP with variables x R n and y R m. 1 T y y Ax b y 1 x 1, 4. Equivalent to the LP with variables x and y. 1 T y y x y 1 Ax b 1 Another good solution is to write x as the difference of two nonnegative vectors x = x + x, and to express the problem as with variables x + R n and x R n. 1 T x + + 1 T x 1 Ax + Ax b 1 x + 0, x 0, 5. Equivalent to with variables x, y, and t. 1 T y + t y Ax b y t1 x t1, Exercise 5 (Linear separation of two sets of ellipsoids.) Suppose we are given K + L ellipsoids E i = {P i u + q i u 2 1}, i = 1,..., K + L, where P i S n. We are interested in finding a hyperplane that strictly separates E 1,..., E K from E K+1,..., E K+L, i.e., we want to compute a R n, b R such that a T x + b > 0 for x E 1 E K, a T x + b < 0 for x E K+1 E K+L, or prove that no such hyperplane exists. Express this problem as an SOCP feasibility problem. Solution 5 Solution. We first note that the problem is homogeneous in a and b, so we can replace the strict inequalities a T x+b > 0 and a T x+b < 0 with a T x+b 1 and a T x+b 1, respectively. The variables a and b must satisfy inf u 2 1 (at P i u + a T q i ) 1, 1,..., L 5

and The lefthand sides can be expressed as sup (a T P i u + a T q i ) 1, i = K + 1,..., K + L. u 2 1 inf u 2 1 (at P i u + a T q i ) = Pi T a 2 + a T q i + b, sup (a T P i u + a T q i ) = Pi T a 2 + a T q i + b. u 2 1 We therefore obtain a set of second-order cone constraints in a, b: Pi T a 2 + a T q i + b 1, i = 1,..., L Pi T a 2 + a T q i + b 1, i = K + 1,..., K + L. Exercise 6 (Dual of general LP.) Find the dual function of the LP c T x Gx h Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution 6 Solution. 1. The Lagrangian is L(x, λ, ν) = c T x + λ T (Gx h) + ν T (Ax b) = (c T + λ T G + ν T A)x hλ T ν T b, which is an affine function of x. It follows that the dual function is given by { λ g(λ, ν) = inf L(x, λ, ν) = T h ν T b c + G T λ + A T ν = 0 x otherwise. 2. The dual problem is maximize g(λ, ν) λ 0. After making the implicit constraints explicit, we obtain maximize λ T h ν T b c + G T λ + A T ν = 0 λ 0. 6