Semidefinite Programming

Similar documents
Interior Point Methods for Mathematical Programming

Conic Linear Optimization and its Dual. yyye

Semidefinite Programming. Yinyu Ye

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

On well definedness of the Central Path

Conic Linear Programming. Yinyu Ye

Lagrangian Duality Theory

Summer School: Semidefinite Optimization

Conic Linear Programming. Yinyu Ye

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

Lecture 6: Conic Optimization September 8

Nonlinear Programming Models

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Linear and non-linear programming

Convex Optimization M2

More First-Order Optimization Algorithms

5. Duality. Lagrangian

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Chapter 1. Preliminaries

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Interior Point Methods in Mathematical Programming

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

Optimization: Then and Now

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Largest dual ellipsoids inscribed in dual cones

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

Introduction to Semidefinite Programming I: Basic properties a

LECTURE 10 LECTURE OUTLINE

Lecture 10. Primal-Dual Interior Point Method for LP

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

An E cient A ne-scaling Algorithm for Hyperbolic Programming

7. Lecture notes on the ellipsoid algorithm

Lecture 9 Sequential unconstrained minimization

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Lecture 5. The Dual Cone and Dual Problem

Interior-Point Methods

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lecture 17: Primal-dual interior-point methods part II

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

8. Geometric problems

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Lecture: Duality.

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

c 2000 Society for Industrial and Applied Mathematics

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Interior-Point Methods for Linear Optimization

Theory and Internet Protocols

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

8. Geometric problems

10 Numerical methods for constrained problems

Lecture: Algorithms for LP, SOCP and SDP

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

Limiting behavior of the central path in semidefinite optimization

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

The Simplest Semidefinite Programs are Trivial

Convex optimization. Javier Peña Carnegie Mellon University. Universidad de los Andes Bogotá, Colombia September 2014

Advances in Convex Optimization: Theory, Algorithms, and Applications

MAT-INF4110/MAT-INF9110 Mathematical optimization

Constrained Optimization and Lagrangian Duality

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Lecture: Introduction to LP, SDP and SOCP

A ten page introduction to conic optimization

Primal-dual IPM with Asymmetric Barrier

Enlarging neighborhoods of interior-point algorithms for linear programming via least values of proximity measure functions

Polynomiality of Linear Programming

CS711008Z Algorithm Design and Analysis

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

A Redundant Klee-Minty Construction with All the Redundant Constraints Touching the Feasible Region

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Introduction and Math Preliminaries

A QUADRATIC CONE RELAXATION-BASED ALGORITHM FOR LINEAR PROGRAMMING

Lecture notes on the ellipsoid algorithm

LP. Kap. 17: Interior-point methods

Inner approximation of convex cones via primal-dual ellipsoidal norms

Chapter 6 Interior-Point Approach to Linear Programming

Chapter 1 Linear Programming. Paragraph 5 Duality

PRIMAL-DUAL INTERIOR-POINT METHODS FOR SELF-SCALED CONES

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

12. Interior-point methods

3. Linear Programming and Polyhedral Combinatorics

A Brief Review on Convex Optimization

Lecture 15: October 15

SCALING DUALITIES AND SELF-CONCORDANT HOMOGENEOUS PROGRAMMING IN FINITE DIMENSIONAL SPACES

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

Lecture 8. Strong Duality Results. September 22, 2008

A new primal-dual path-following method for convex quadratic programming

Semidefinite Programming

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

Transcription:

Chapter 2 Semidefinite Programming 2.0.1 Semi-definite programming (SDP) Given C M n, A i M n, i = 1, 2,..., m, and b R m, the semi-definite programming problem is to find a matrix X M n for the optimization problem: (SDP ) inf C X subject to A i X = b i, i = 1, 2,..., m, X 0. Recall that the operation is the matrix inner product A B := tra T B. The notation X 0 means that X is a positive semi-definite matrix, and X 0 means that X is a positive definite matrix. If a point X 0 and satisfies all equations in (SDP), it is called a (primal) strictly or interior feasible solution.. The dual problem to (SDP) can be written as: (SDD) sup b T y m subject to i y i A i + S = C, S 0, which is analogous to the dual (LD) of LP. Here y R m and S M n. If a point (y, S 0) satisfies all equations in (SDD), it is called a dual interior feasible solution. 31

32 CHAPTER 2. SEMIDEFINITE PROGRAMMING Example 2.1 Let P (y R m ) = C + m i y i A i, where C and A i, i = 1,..., m, are given symmetric matrices. The problem of minimizing the max-eigenvalue of P (y) can be cast as a (SDD) problem. In semi-definite programming, we minimize a linear function of a matrix in the positive semi-definite matrix cone subject to affine constraints. In contrast to the positive orthant cone of linear programming, the positive semi-definite matrix cone is non-polyhedral (or non-linear ), but convex. So positive semi-definite programs are convex optimization problems. Semi-definite programming unifies several standard problems, such as linear programming, quadratic programming, and convex quadratic minimization with convex quadratic constraints, and finds many applications in engineering, control, and combinatorial optimization. We have several theorems analogous to Farkas lemma. Theorem 2.1 (Farkas lemma in SDP) Let A i M n, i = 1,..., m, have rank m (i.e., m i y i A i = 0 implies y = 0) and b R m. Then, there exists a symmetric matrix X 0 with A i X = b i, i = 1,..., m, if and only if m i y i A i 0 and m i y i A i 0 implies b T y < 0. Note the difference between the LP and SDP. Theorem 2.2 (Weak duality theorem in SDP) Let F p and F d, the feasible sets for the primal and dual, be non-empty. Then, C X b T y where X F p, (y, S) F d. The weak duality theorem is identical to that of (LP) and (LD). Corollary 2.3 (Strong duality theorem in SDP) Let F p and F d be nonempty and have an interior. Then, X is optimal for (PS) if and only if the following conditions hold: i) X F p ; ii) there is (y, S) F d ; iii) C X = b T y or X S = 0. Again note the difference between the above theorem and the strong duality theorem for LP.

2.1. ANALYTIC CENTER 33 Example 2.2 Let C = 0 1 0 1 0 0, A 1 == 0 0 0 0 1 0, A 1 == 0 1 0 1 0 0 0 0 0 0 0 0 0 0 2 and ( 0 b = 10 ). Two positive semi-definite matrices are complementary to each other, X S = 0, if and only if XS = 0 (Exercise 1.22). From the optimality conditions, the solution set for certain (SDP) and (SDD) is or S p = {X F p, (y, S) F d : C X b T y = 0}, S p = {X F p, (y, S) F d : XS = 0}, which is a system of linear matrix inequalities and equations. In general, we have Theorem 2.4 (SDP duality theorem) If one of (SDP) or (SDD) has a strictly or interior feasible solution and its optimal value is finite, then the other is feasible and has the same optimal value. If one of (SDP) or (SDD) is unbounded then the other has no feasible solution. Note that a duality gap may exists if neither (SDP) nor (SDD) has a strictly feasible point. This is in contrast to (LP) and (LD) where no duality gap exists if both are feasible. Although semi-definite programs are much more general than linear programs, they are not much harder to solve. It has turned out that most interior-point methods for LP have been generalized to semi-definite programs. As in LP, these algorithms possess polynomial worst-case complexity under certain computation models. They also perform well in practice. We will describe such extensions later in this book. 2.1 Analytic Center 2.1.1 AC for polytope Let Ω be a bounded polytope in R m represented by n (> m) linear inequalities, i.e., Ω = {y R m : c A T y 0},

34 CHAPTER 2. SEMIDEFINITE PROGRAMMING where A R m n and c R n are given and A has rank m. Denote the interior of Ω by Ω= {y R m : c A T y > 0}. Define d(y) = n (c j a T j y), y Ω, where a.j is the jth column of A. Traditionally, we let s := c A T y and call it a slack vector. Thus, the function is the product of all slack variables. Its logarithm is called the (dual) potential function, B(y) := log d(y) = log(c j a Ṭ jy) = log s j, (2.1) and B(y) is the classical logarithmic barrier function. For convenience, in what follows we may write B(s) to replace B(y) where s is always equal to c A T y. Example 2.3 Let A = (1, 1) and c = (1; 1). Then the set of Ω is the interval [ 1, 1]. Let A = (1, 1, 1) and c = (1; 1; 1). Then the set of Ω is also the interval [ 1, 1]. Note that and d( 1/2) = (3/2)(1/2) = 3/4 and B( 1/2) = log(3/4), d ( 1/2) = (3/)(1/2)(1/2) = 3/8 and B ( 1/2) = log(3/8). The interior point, denoted by y a and s a = c A T y a, in Ω that maximizes the potential function is called the analytic center of Ω, i.e., B(Ω) := B(y a, Ω) = max log d(y, Ω). y Ω (y a, s a ) is uniquely defined, since the potential function is strictly concave in a bounded convex Ω. Setting B(y, Ω) = 0 and letting x a = (S a ) 1 e, the analytic center (y a, s a ) together with x a satisfy the following optimality conditions: Xs = e Ax = 0 (2.2) A T y s = c. Note that adding or deleting a redundant inequality changes the location of the analytic center.

2.1. ANALYTIC CENTER 35 Example 2.4 Consider Ω = {y R : y 0, y 1}, which is interval [0, 1]. The analytic center is y a = 1/2 with x a = (2, 2) T. Consider Ω = {y R : n times {}}{ y 0,, y 0, y 1}, which is, again, interval [0, 1] but y 0 is copied n times. The analytic center for this system is y a = n/(n + 1) with x a = ((n + 1)/n,, (n + 1)/n, (n + 1)) T. The analytic center can be defined when the interior is empty or equalities are presented, such as Ω = {y R m : c A T y 0, By = b}. Then the analytic center is chosen on the hyperplane {y : By = b} to maximize the product of the slack variables s = c A T y. Thus, the interior of Ω is not used in the sense that the topological interior for a set is used. Rather, it refers to the interior of the positive orthant of slack variables: R n + := {s : s 0}. When say Ω has an interior, we mean that Again R n + {s : s = c A T y for some y where By = b} =. R n +:= {s R n + : s > 0}, i.e., the interior of the orthant R n +. Thus, if Ω has only a single point y with s = c A T y > 0, we still say Ω is not empty. Example 2.5 Consider the system Ω = {x : Ax = 0, e T x = n, x 0}, which is called Karmarkar s canonical set. If x = e is in Ω then e is the analytic center of Ω, the intersection of the simplex {x : e T x = n, x 0} and the hyperplane {x : Ax = 0} (Figure 2.1). 2.1.2 AC for SDP Let Ω be a bounded convex set in R m represented by n (> m) a matrix inequality, i.e., m Ω = {y R m : C y i A i 0, }. Let S = C m i y i A i and B(y) := log det(s)) = log det(c i m y i A i ). (2.3) i

36 CHAPTER 2. SEMIDEFINITE PROGRAMMING x 3 (0,0,3) Ax=0. (1,1,1) (3,0,0) x 1 (0,3,0) x 2 Figure 2.1: Illustration of the Karmarkar (simplex) polytope and its analytic center. The interior point, denoted by y a and S a = C m i yi aa i, in Ω that maximizes the potential function is called the analytic center of Ω, i.e., max B(y). y Ω (y a, S a ) is uniquely defined, since the potential function is strictly concave in a bounded convex Ω. Setting B(y, Ω) = 0 and letting X a = (S a ) 1, the analytic center (y a, S a ) together with X a satisfy the following optimality conditions: XS = I AX = 0 (2.4) A T y S = C. 2.2 Potential Functions for LP and SDP We show how potential functions can be defined to solve linear programming problems and semi-definite programming. We assume that for a given LP data set (A, b, c), both the primal and dual have interior feasible point. We also let z be the optimal value of the standard form (LP) and (LD). Denote the feasible sets of (LP) and (LD) by F p and F d, respectively. Denote by F = F p F d, and the interior of F by F.

2.2. POTENTIAL FUNCTIONS FOR LP AND SDP 37 2.2.1 Primal potential function for LP Consider the level set Ω(z) = {y R m : c A T y 0, z + b T y 0}, (2.5) where z < z. Since both (LP) and (LD) have interior feasible point for given (A, b, c), Ω(z) is bounded and has an interior for any finite z, even Ω := F d is unbounded (Exercise 1.23). Clearly, Ω(z) Ω, and if z 2 z 1, Ω(z 2 ) Ω(z 1 ) and the inequality z + b T y is translated from z = z 1 to z = z 2. From the duality theorem again, finding a point in Ω(z) has a homogeneous primal problem For (x, x 0) satisfying let x := x /x 0 F p, i.e., minimize c T x zx 0 s.t. Ax bx 0 = 0, (x, x 0) 0. Ax bx 0 = 0, (x, x 0) > 0, Ax = b, x > 0. Then, the primal potential function for Ω(z) (Figure 2.2), as described in the preceding section, is P(x, Ω(z)) = (n + 1) log(c T x zx 0) = (n + 1) log(c T x z) log x j j=0 log x j =: P n+1 (x, z). The latter, P n+1 (x, z), is the Karmarkar potential function in the standard LP form with a lower bound z for z. One algorithm for solving (LD) is suggested in Figure 2.2. If the objective hyperplane is repeatedly translated to the analytic center, the sequence of new analytic centers will converge to an optimal solution and the potentials of the new polytopes will decrease to. As we illustrated before, one can represent Ω(z) differently: ρ times {}}{ Ω(z) = {y : c A T y 0, z + b T y 0,, z + b T y 0}, (2.6)

38 CHAPTER 2. SEMIDEFINITE PROGRAMMING y a b T y = b T y a y a b T y = z The objective hyperplane The updated objective hyperplane Figure 2.2: Intersections of a dual feasible region and the objective hyperplane; b T y z on the left and b T y b T y a on the right. i.e., z + b T y 0 is copied ρ times. Geometrically, this representation does not change Ω(z), but it changes the location of its analytic center. Since the last ρ inequalities in Ω(z) are identical, they must share the same slack value and the same corresponding primal variable. Let (x, x 0) be the primal variables. Then the primal problem can be written as ρ times {}}{ minimize c T x zx 0 zx 0 ρ times {}}{ s.t. Ax bx 0 bx 0 = 0, (x, x 0) 0. Let x = x /(ρx 0) F p. Then, the primal potential function for the new Ω(z) given by (2.6) is P(x, Ω(z)) = (n + ρ) log(c T x z(ρx 0)) = (n + ρ) log(c T x z) =: P n+ρ (x, z) + ρ log ρ. log x j ρ log x 0 log x j + ρ log ρ

2.2. POTENTIAL FUNCTIONS FOR LP AND SDP 39 The function P n+ρ (x, z) = (n + ρ) log(c T x z) log x j (2.7) is an extension of the Karmarkar potential function in the standard LP form with a lower bound z for z. It represents the volume of a coordinate-aligned ellipsoid whose intersection with A Ω(z) contains S Ω(z), where z + b T y 0 is duplicated ρ times. 2.2.2 Dual potential function for LP We can also develop a dual potential function, symmetric to the primal, for (y, s) F d B n+ρ (y, s, z) = (n + ρ) log(z b T y) log s j, (2.8) where z is a upper bound of z. One can show that it represents the volume of a coordinate-aligned ellipsoid whose intersection with the affine set {x : Ax = b} contains the primal level set {x F p : ρ times {}}{ c T x z 0,, c T x z 0}, where c T x z 0 is copied ρ times (Exercise 2.7). For symmetry, we may write B n+ρ (y, s, z) simply by B n+ρ (s, z), since we can always recover y from s using equation A T y = c s. 2.2.3 Primal-dual potential function for LP A primal-dual potential function for linear programming will be used later. For x F p and (y, s) F d it is defined by where ρ 0. We have the relation: ψ n+ρ (x, s) := (n + ρ) log(x T s) ψ n+ρ (x, s) = (n + ρ) log(c T x b T y) log(x j s j ), (2.9) log x j log s j

40 CHAPTER 2. SEMIDEFINITE PROGRAMMING Since = P n+ρ (x, b T y) = B n+ρ (s, c T x) log s j log x j. ψ n+ρ (x, s) = ρ log(x T s) + ψ n (x, s) ρ log(x T s) + n log n, then, for ρ > 0, ψ n+ρ (x, s) implies that x T s 0. More precisely, we have x T s exp( ψ n+ρ(x, s) n log n ). ρ We have the following theorem: Theorem 2.5 Define the level set i) Ψ(δ) := {(x, y, s) F: ψ n+ρ (x, s) δ}. Ψ(δ 1 ) Ψ(δ 2 ) if δ 1 δ 2. ii) Ψ (δ) = {(x, y, s) F : ψ n+ρ (x, s) < δ}. iii) For every δ, Ψ(δ) is bounded and its closure ˆΨ(δ) has non-empty intersection with the solution set. Later we will show that a potential reduction algorithm generates sequences {x k, y k, s k } F such that ψ n+ n (x k+1, y k+1, s k+1 ) ψ n+ n (x k, y k, s k ).05 for k = 0, 1, 2,... This indicates that the level sets shrink at least a constant rate independently of m or n.

2.2. POTENTIAL FUNCTIONS FOR LP AND SDP 41 2.2.4 Potential function for SDP The potential functions for SDP of Section 2.0.1 are analogous to those for LP. For given data, we assume that both (SDP) and (SDD) have interior feasible points. Then, for any X F p and (y, S) F d, the primal potential function is defined by P n+ρ (X, z) := (n + ρ) log(c X z) log det(x), z z ; the dual potential function is defined by B n+ρ (y, S, z) := (n + ρ) log(z b T y) log det(s), z z, where ρ 0 and z designates the optimal objective value. For X F p and (y, S) F d the primal-dual potential function for SDP is defined by ψ n+ρ (X, S) := (n + ρ) log(x S) log(det(x) det(s)) = (n + ρ) log(c X b T y) log det(x) log det(s) = P n+ρ (X, b T y) log det(s) = B n+ρ (S, C X) log det(x), where ρ 0. Note that if X and S are diagonal matrices, these definitions reduce to those for LP. Note that we still have (Exercise 2.8) ψ n+ρ (X, S) = ρ log(x S) + ψ n (X, S) ρ log(x S) + n log n. Then, for ρ > 0, ψ n+ρ (X, S) implies that X S 0. More precisely, we have X S exp( ψ n+ρ(x, S) n log n ). ρ We also have the following corollary: Corollary 2.6 Let (SDP) and (SDD) have non-empty interior and define the level set Ψ(δ) := {(X, y, S) F: ψ n+ρ (X, S) δ}. i) ii) Ψ(δ 1 ) Ψ(δ 2 ) if δ 1 δ 2. Ψ (δ) = {(X, y, S) F : ψ n+ρ (X, S) < δ}. iii) For every δ, Ψ(δ) is bounded and its closure ˆΨ(δ) has non-empty intersection with the solution set.

42 CHAPTER 2. SEMIDEFINITE PROGRAMMING 2.3 Central Paths of LP and SDP Many interior-point algorithms find a sequence of feasible points along a central path that connects the analytic center and the solution set. We now present this one of the most important foundations for the development of interior-point algorithms. 2.3.1 Central path for LP Consider a linear program in the standard form (LP) and (LD). Assume that F, i.e., both F p and F d, and denote z the optimal objective value. The central path can be expressed as { } C = (x, y, s) F: Xs = xt s n e in the primal-dual form. We also see { C = (x, y, s) F: } ψ n (x, s) = n log n. For any µ > 0 one can derive the central path simply by minimizing the primal LP with a logarithmic barrier function: (P ) minimize c T x µ n log x j s.t. Ax = b, x 0. Let x(µ) F p be the (unique) minimizer of (P). Then, for some y R m it satisfies the optimality conditions Xs = µe Ax = b A T y s = c. (2.10) Consider minimizing the dual LP with the barrier function: (D) maximize b T y + µ n log s j s.t. A T y + s = c, s 0. Let (y(µ), s(µ)) F d be the (unique) minimizer of (D). Then, for some x R n it satisfies the optimality conditions (2.10) as well. Thus, both minimizers x(µ) and (y(µ), s(µ)) are on the central path with x(µ) T s(µ) = nµ.

2.3. CENTRAL PATHS OF LP AND SDP 43 Another way to derive the central path is to consider again the dual level set Ω(z) of (2.5) for any z < z (Figure 2.3). Then, the analytic center (y(z), s(z)) of Ω(z) and a unique point (x (z), x 0(z)) satisfies Ax (z) bx 0(z) = 0, X (z)s = e, s = c A T y, and x 0(z)(b T y z) = 1. Let x(z) = x (z)/x 0(z), then we have Ax(z) = b, X(z)s(z) = e/x 0(z) = (b T y(z) z)e. Thus, the point (x(z), y(z), s(z)) is on the central path with µ = b T y(z) z and c T x(z) b T y(z) = x(z) T s(z) = n(b T y(z) z) = nµ. As we proved earlier in Section 2.2, (x(z), y(z), s(z)) exists and is uniquely defined, which imply the following theorem: Theorem 2.7 Let both (LP) and (LD) have interior feasible points for the given data set (A, b, c). Then for any 0 < µ <, the central path point (x(µ), y(µ), s(µ)) exists and is unique. y a The objective hyperplanes Figure 2.3: The central path of y(z) in a dual feasible region. The following theorem further characterizes the central path and utilizes it to solving linear programs.

44 CHAPTER 2. SEMIDEFINITE PROGRAMMING Theorem 2.8 Let (x(µ), y(µ), s(µ)) be on the central path. i) The central path point (x(µ), s(µ)) is bounded for 0 < µ µ 0 and any given 0 < µ 0 <. ii) For 0 < µ < µ, c T x(µ ) < c T x(µ) and b T y(µ ) > b T y(µ). iii) (x(µ), s(µ)) converges to an optimal solution pair for (LP) and (LD). Moreover, the limit point x(0) P is the analytic center on the primal optimal face, and the limit point s(0) Z is the analytic center on the dual optimal face, where (P, Z ) is the strict complementarity partition of the index set {1, 2,..., n}. Proof. Note that (x(µ 0 ) x(µ)) T (s(µ 0 ) s(µ)) = 0, since (x(µ 0 ) x(µ)) N (A) and (s(µ 0 ) s(µ)) R(A T ). This can be rewritten as or ( s(µ 0 ) j x(µ) j + x(µ 0 ) ) j s(µ) j = n(µ 0 + µ) 2nµ 0, j j ( x(µ)j x(µ 0 + s(µ) ) j ) j s(µ 0 2n. ) j Thus, x(µ) and s(µ) are bounded, which proves (i). We leave the proof of (ii) as an exercise. Since x(µ) and s(µ) are both bounded, they have at least one limit point which we denote by x(0) and s(0). Let x P (x Z = 0) and s Z (s P = 0), respectively, be the unique analytic centers on the primal and dual optimal faces: {x P : A P x P = b, x P 0} and {s Z : s Z = c Z A T Z y 0, c P A T P y = 0}. Again, we have ( s j x(µ) j + x ) j s(µ) j = nµ, j or ( x j x(µ) j j P ) + j Z ( s ) j = n. s(µ) j

2.3. CENTRAL PATHS OF LP AND SDP 45 Thus, we have and This implies that and Furthermore, or j P x j x(µ) j x j /n > 0, j P s(µ) j s j /n > 0, j Z. j P x(µ) j 0, j Z s(µ) j 0, j P. x j s j 1 x(µ) j s(µ) j j Z s j j Z x(µ) j s(µ) j. j P j Z However, ( j P x j )( j Z s j ) is the maximal value of the potential function over all interior point pairs on the optimal faces, and x(0) P and s(0) Z is one interior point pair on the optimal face. Thus, we must have = x(0) j s(0) j. j P j Z Therefore, j P x j j Z s j x(0) P = x P and s(0) Z = s Z, since x P and s Z are the unique maximizer pair of the potential function. This also implies that the limit point of the central path is unique. We usually define a neighborhood of the central path as { } N (η) = (x, y, s) F: Xs µe ηµ and µ = xt s, n where. can be any norm, or even a one-sided norm as We have the following theorem: x = min(0, min(x)).

46 CHAPTER 2. SEMIDEFINITE PROGRAMMING Theorem 2.9 Let (x, y, s) N (η) for constant 0 < η < 1. i) The N (η) {(x, y, s) : x T s nµ 0 } is bounded for any given µ 0 <. ii) Any limit point of N (η) as µ 0 is an optimal solution pair for (LP) and (LD). Moreover, for any j P x j (1 η)x j n where x is any optimal primal solution; for any j Z, s j (1 η)s j n where s is any optimal dual solution., 2.3.2 Central path for SDP Consider a SDP problem in Section 2.0.1 and Assume that F, i.e., both F p and F d. The central path can be expressed as { C = (X, y, S) F: } XS = µi, 0 < µ <, or a symmetric form { C = (X, y, S) F: } X.5 SX.5 = µi, 0 < µ <, where X.5 M n + is the square root matrix of X M n +, i.e., X.5 X.5 = X. We also see { C = (X, y, S) F: } ψ n (X, S) = n log n. When X and S are diagonal matrices, this definition is identical to LP. We also have the following corollary: Corollary 2.10 Let both (SDP) and (SDD) have interior feasible points. Then for any 0 < µ <, the central path point (X(µ), y(µ), S(µ)) exists and is unique. Moreover, i) the central path point (X(µ), S(µ)) is bounded where 0 < µ µ 0 for any given 0 < µ 0 <. ii) For 0 < µ < µ, C X(µ ) < C X(µ) and b T y(µ ) > b T y(µ).

2.4. NOTES 47 iii) (X(µ), S(µ)) converges to an optimal solution pair for (SDP) and (SDD), and the rank of the limit of X(µ) is maximal among all optimal solutions of (SDP) and the rank of the limit S(µ) is maximal among all optimal solutions of (SDD). 2.4 Notes General convex problems, such as membership, separation, validity, and optimization, can be solved by the central-section method; see Grötschel, Lovász and Schrijver [170]. Levin [244] and Newman [318] considered the center of gravity of a convex body; Elzinga and Moore [110] considered the center of the max-volume sphere contained in a convex body. A number of Russian mathematicians (for example, Tarasov, Khachiyan and Érlikh [403]) considered the center of the max-volume ellipsoid inscribing the body; Huard and Liêu [190, 191] considered a generic center in the body that maximizes a distance function; and Vaidya [438] considered the volumetric center, the maximizer of the determinant of the Hessian matrix of the logarithmic barrier function. See Kaiser, Morin and Trafalis [210] for a complete survey. Dyer and Frieze [104] proved that computing the volume of a convex polytope, either given as a list of facets or vertices, is itself #P -Hard. Furthermore, Elekes [109] has shown that no polynomial time algorithm can compute the volume of a convex body with less than exponential relative error. Bárány and Fürendi [42] further showed that for Ω R m, any polynomial time algorithm that gives an upper and lower bound on the volume of Ω, represented as V (Ω) and V (Ω), respectively, necessarily has an exponential gap between them. They showed V (Ω)/V (Ω) (cm/ log m) m, where c is a constant independent of m. In other words, there is no polynomial time algorithm that would compute V (Ω) and V (Ω) such that V (Ω)/V (Ω) < (cm/ log m) m. Recently, Dyer, Frieze and Kannan [105] developed a random polynomial time algorithm that can, with high probability, find a good estimate for the volume of Ω. Apparently, the result that every convex body contains a unique ellipsoid of maximal volume and is contained in a unique ellipsoid of minimal volume, was discovered independently by several mathematicians see, for example, Danzer, Grünbaum and Klee [94]. These authors attributed the first proof to K. Löwner. John [208] later proved the Theorem.

48 CHAPTER 2. SEMIDEFINITE PROGRAMMING Khachiyan and Todd [222] established a polynomial complexity bound for computing an approximate point of the center of the maximal inscribing ellipsoid if the convex body is represented by linear inequalities, The analytic center for a convex polyhedron given by linear inequalities was introduced by Huard [190], and later by Sonnevend [383]. The function d(y, Ω) is very similar to Huard s generic distance function, with one exception, where property (3) there was stated as If Ω Ω, then d(y, Ω ) d(y, Ω). The reason for the difference is that the distance function may return different values even if we have the same polytope but two different representations. The negative logarithmic function d(y, Ω), called the barrier function, was introduced by Frisch [126]. Todd [405] and Ye [465] showed that Karmarkar s potential function represents the logarithmic volume of a coordinate-aligned ellipsoid who contains the feasible region. The Karmarkar potential function in the standard form (LP) with a lower bound z for z was seen in Todd and Burrell [413], Anstreicher [24], Gay [133], and Ye and Kojima [477]. The primal potential function with ρ > 1 was proposed by Gonzaga [160], Freund [123], and Ye [466, 468]. The primal-dual potential function was proposed by Tanabe [400], and Todd and Ye [415]. Potential functions for LCP and SDP were studied by Kojima et al. [230, 228], Alizadeh [9], and Nesterov and Nemirovskii [327]. McLinden [267] earlier, then Bayer and Lagarias [45, 46], Megiddo [271], and Sonnevend [383], analyzed the central path for linear programming and convex optimization. Megiddo [271] derived the central path simply minimizing the primal with a logarithmic barrier function as in Fiacco and McCormick [116]. The central path for LCP, with more general matrix M, was given by Kojima et al. [227] and Güler [174]; the central path theory for SDP was first developed by Nesterov and Nemirovskii [327]. McLinden [267] proved Theorem 2.8 for the monotone LCP, which includes LP. 2.5 Exercises 2.1 Find the min-volume ellipsoid containing a half of the unit ball {x R n : x 1}. 2.2 Verify Examples 2.3, 2.4, and 2.5. 2.3 Compare and contrast the center of gravity of a polytope and its analytic center.

2.5. EXERCISES 49 2.4 Consider the maximum problem maximize f(x) = x e n x j s.t. e T x = n, x > 0 R n. Prove that its maximizer is achieved at x 1 = β and x 2 =... = x n = (n β)/(n 1) > 0 for some 1 < β < n. 2.5 Let Ω= {y R m : c A T y > 0} =, Ω = {y R m : c A T y > 0} =, and c c. Prove B(Ω ) B(Ω). 2.6 If Ω = {y : c A T y 0} is nonempty, prove the minimal value of its primal problem is 0; if Ω is bounded and has an interior, prove the interior of X Ω := {x R n : Ax = 0, x 0} is nonempty and x = 0 is the unique primal solution. 2.7 Let (LP) and (LD) have interior. Prove the dual potential function B n+1 (y, s, z), where z is a upper bound of z, represents the volume of a coordinate-aligned ellipsoid whose intersection with the affine set {x : Ax = b} contains the primal level set {x F p : c T x z}. 2.8 Let X, S M n be both positive definite. Then prove ψ n (X, S) = n log(x S) log(det(x) det(s)) n log n. 2.9 Consider linear programming and the level set Prove that Ψ(δ) := {(x, y, s) F: ψ n+ρ (x, s) δ}. Ψ(δ 1 ) Ψ(δ 2 ) if δ 1 δ 2, and for every δ Ψ(δ) is bounded and its closure ˆΨ(δ) has non-empty intersection with the solution set. 2.10 Consider the linear program max b T y s.t. 0 y 1 1, 0 y 2 1. Draw the feasible region the the (dual) potential level sets respectively, for {y : B 5 (y, s, 2) 0} and {y : B 5 (y, s, 2) 10},

50 CHAPTER 2. SEMIDEFINITE PROGRAMMING 1. b = (1; 0); 2. b = (1; 1)/2; 3. b = (2; 1)/3. 2.11 Consider the polytope {y R 2 : 0 y 1, 0 y 2 1, y 1 + y 2 z}. Describe how the analytic center changes as z decreases from 10 to 1. 2.12 Consider the linear program max b T y s.t. 0 y 1 1, 0 y 2 1. Draw the feasible region, central path, and solution set for 1. b = (1; 0); 2. b = (1; 1)/2; 3. b = (2; 1)/3. Finally, sketch a neighborhood for the third central path. 2.13 Prove (ii) of Theorem 2.8. 2.14 Prove Theorem 2.9.