On Conically Ordered Convex Programs

Size: px
Start display at page:

Download "On Conically Ordered Convex Programs"

Transcription

1 On Conically Ordered Convex Programs Shuzhong Zhang May 003; revised December 004 Abstract In this paper we study a special class of convex optimization problems called conically ordered convex programs COCP), where the feasible region is given as the level set of a vector-valued nonlinear mapping, expressed as a nonnegative combination of convex functions. The nonnegativity of the vectors is defined using a pre-described conic ordering. The new model extends the ordinary convex programming models where the feasible sets are the level sets of convex functions, and it also extends the famous linear conic optimization models. We introduce a condition on the barrier function for the order-defining cone, termed as the cone-consistent property. The relationship between the cone-consistent barriers and the self-concordance barriers is investigated. We prove that if the order-defining cone admits a self-concordant and cone-consistent barrier function, and moreover, if the constraint functions are all convex quadratic then the overall composite barrier function is self-concordant. The problem is thus solvable in polynomial time, following Nesterov and Nemirovskii, by means of the path-following method. If the constraint functions are not quadratic, but harmonically convex, then we propose a variant of Iri-Imai type potential reduction method. In addition to the self-concordance and the cone-consistence conditions, we assume that the barrier function for the order-defining cone has the property that the image of the cone under its Hessian matrix is contained in its dual cone. All these conditions are satisfied by the familiar self-scaled cones. Under these conditions we show that the Iri-Imai type potential reduction algorithm converges in polynomial time. Duality issues related to this class of optimization problems are discussed as well. Keywords: convex programming, conic ordering, self-concordant barriers, potential reduction. Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, Hong Kong. Research supported by Hong Kong RGC Earmarked Grants CUHK433/0E and CUHK474/03E.

2 The problem formulation Let K R l be a solid closed convex cone. Let h j K, j =,..., k, and h 0 R l. Consider the following optimization problem, which we shall call conically ordered convex program COCP) in this paper, P K ) minimize c T x subject to Ax = b h 0 f j x)h j K, where A R m n, c R n and b R m. If we use the conic ordering defined by K, i.e., for y, z R l, y K z if and only if z y K, then P K ) can be written as minimize subject to c T x Ax = b f j x)h j K h 0. Note that all the results in this paper straightforwardly extend to the setting where more than one conic constraint are present, namely, minimize subject to c T x Ax = b fjx)h i i j K i h i 0, i =,..., p. To keep the analysis simple, however, we shall restrict ourselves to only one conic constraint in the subsequent discussion. It is evident that P K ) is indeed a convex program as stated below. Lemma. If f i s are all convex functions, i =,..., k, then P K ) is a convex program. Since K is solid, i.e. K + K) = R l, and so any vector in R l can be written as difference of two vectors in K. Therefore, the ordinary linear conic constraint h 0 k x j h j K, where h j R l for all j, can be written as h 0 k x j h + j x jh j ) = h 0 k x j h + j k x j )h j K, where h j = h + j h j and h + j, h j K for all j; the latter expression is in the form of COCP. Obviously, if K = R l + and h j s are unit vectors, then P K ) is the usual convex program where the constraints are given by inequalities: f j x) 0, j =,..., k.

3 Another interesting special case of P K ) is the following P R l ) + minimize c T x subject to Ax = b f j x)h j h 0 where f j x) = e ιt j x+ι j0 with ι j R n and ι j0 R, and h j R l +, j = 0,,..., k. This problem is known as geometric program. Note that in its original form, a general geometric program is given as GP ) minimize subject to m c j f 0j t) m a ij f ij t) b i, i =,..., p f ij t) = b i, i = p +,..., p + q, t > 0,..., t n > 0 where c j > 0 and a ij > 0 and f ij t) = t ιij) t ιij) n, i = 0,,..., p + q, j =,..., m. Such f ij t) s are called monomials. Using the variable transformation x = log t, x = log t,..., x n = log t n the monomial f ij t) becomes e ιij) ) T x where ι ij) ) T = [ι ij),, ι ij) n ], i = 0,,..., p + q, j =,..., m. This way, GP ) is transformed into a COCP problem P R l + ). Next let us consider another special case of COCP P P SD ) minimize c T x subject to Ax = b H 0 f j x)h j 0, where H 0,..., H k 0 are l l positive semidefinite matrices, and f x),..., f k x) are convex functions. Consider the following barrier function F x) = log deth 0 f j x)h j ) ) 3

4 for the cone of l l positive semidefinite matrices S+. l A natural question to ask is: under what conditions is the barrier function F x) defined in ) self-concordant? For details on the theory of self-concordant barriers, one is referred to the excellent texts [5, 4, 7]. The notion of self-concordant barrier function was introduced by Nesterov and Nemirovskii in their seminal work [5]. For convenience, we shall include the definition below. First, a barrier convex) function F x) for the cone K is defined to have the property that F x) < for all x int K, and F x k ) as x k x where x is on the boundary of K. Moreover, it is called self-concordant Section.3. of [5]) if it further satisfies the property that 3 F x)[u, u, u] C F x)[u, u]) 3/ ) and F x)[u] C F x)[u, u]) / 3) for any x int K and any direction u R n. The positive constants C and C can be scaled so as to have only one degree of freedom. Usually C is set to be and C is referred to as the complexity parameter of the cone K. As noted by Renegar [7] see also []), just like convexity, the self-concordant property is in fact a line property, meaning that if a function is self-concordant along every line thus one-dimensional) restricted to its domain then the function is self-concordant in the whole domain, and vice versa. For this purpose, let us consider a given x in the interior of the feasible domain of P P SD ), and u R n be any given direction. Consider the one-dimensional function Xt) := H 0 f j x + tu)h j in the domain that Xt) 0. Obviously, X t) = f j x + tu)[u]h j = f j x + tu) T uh j, X t) = f j x + tu)[u, u]h j = u T f j x + tu)uh j 0, X t) = 3 f j x + tu)[u, u, u]h j, and [Xt) ] = Xt) X t)xt). 4

5 Let ft) = log det Xt). Applying the chain-rule to the composite function ft) we get ) f t) = tr Xt) X t) 4) ) ) f t) = tr Xt) X t)xt) X t) tr Xt) X t) 5) ) ) 3 ) f t) = 3tr Xt) X t)xt) X t) tr Xt) X t) tr Xt) X t). 6) Theorem. If f j x) s, j =,..., k, are convex and quadratic functions, then F x) as defined in ) is a self-concordant barrier. Proof. Let As discussed before, we need only to verify the self-concordant condition along a given line. M := Xt) / X t)xt) / and M := Xt) / X t)xt) / 0. 7) It follows from 4), 5) and 6) respectively that f t) = tr M, 8) f t) = tr M + tr M = M F + M / F, 9) and f t) = 3tr M M tr M 3 tr ) Xt) X t). 0) By the Cauchy-Schwarz inequality, we have f t) = tr M ltr M ) / lf t)) / which is 3) with C = l. Since f j s are all quadratic, we have X t) 0 and so 0) reduces to f t) = 3tr M M tr M 3. ) Clearly, tr M 3 /3 tr M )/ f t)) /. Hence, tr M 3 f t)) 3/. ) 5

6 Moreover, tr M M M F M F M F M / F M F + M / ) 3/ F 3 = 3 f t)) 3/, 3) where the last step is due to 9). Therefore, from ), ) and 3) we obtain f t) + 3)f t)) 3/, which is ) with C = + 3. Q.E.D. In case all f j s are affine linear, we have M = 0, and f t) f t)) 3/, which corresponds to the well-known fact that the logarithmic determinant barrier is self-concordant for Linear Matrix Inequalities. We note that if all f j s are convex quadratic functions then the feasible set of P P SD ) can also be described by an enlarged system of linear matrix inequalities using Schur s complement lemma. In particular, suppose that f j x) = n i= q T ij x q i) + q 0j, j =,..., k. Then, H 0 k f j x)h j 0 if and only if H 0 q 0j H j q T x q )H / q T x q )H / q T nk x q n)h / k q T x q )H / I I qnk T x q n)h / k I 0. However, the complexity parameter of the resulting LMI system becomes kn + )l, instead of l. The cone-consistent barriers The aim of this section is to derive the essential and general property of the barrier function that enabled Theorem. in the previous section. 6

7 Let us call a barrier function F x) to be consistent with the cone K denoted as C 3 -consistent in the sequel) if it satisfies the following condition: F x)[u, u]) / C3 F x)[u] 4) for all x int K and u K, where C 3 > 0 is a parameter. It is obvious that if K = R n + and F x) = n i= log x i then ) / n u F x)[u, u] = i x i= i n i= u i x i = F x)[u] for x > 0 and u 0. So 4) is satisfied with C 3 =. It is easy to verify that 4) is also satisfied by the logarithmic determinant barrier function for the positive semidefinite matrix cone. Similarly, it can be shown that the additional inequality 4) is satisfied by all self-scaled barrier functions for the self-scaled cones see Nesterov and Todd [6]) with C 3 = ; in a different but equivalent framework, the self-scaled barrier function is the log determinant barrier for the symmetric cones see Faybusovich [] or Sturm [8, 9]). One may therefore think that 4) is a property of the self-scaled cones only. However, this guess is incorrect. Let us consider the following cone in R 3 as an epigraph of the l 4 -norm in R : x K 4 := x x 0, x 4 x 4 + x 4 3) 0. x 3 The above cone is not self-scaled. Also, the barrier function B 4 x) = log x 4 x4 + x4 3 )) is not self-concordant, but it satisfies 4) with C 3 =. To see this, consider x = [x, x, x 3 ] T int K 4. We denote b 4 x) = x 4 x4 + x4 3 ) > 0 and compute that B 4 x) = b 4 x) 4x 3 4x 3 4x 3 3 and B 4 x) = B 4 x) B 4 x) T b 4 x) x x x 3. Take any y = [y, y, y 3 ] T K 4. We have y T B 4 x)y = B 4 x), y ) ) x b 4 x) y x y x 3y3. 7

8 Since x > x 4 + x4 3 and y y 4 + y4 3, we have x y x 4 + x4 3 )y4 + y4 3 ) x 4 y4 + x4 y4 3 + x4 3 y4 + x4 3 y4 3 x 4 y4 + x y x 3 y 3 + x4 3 y4 3 = x y + x 3y 3. Therefore, y T B 4 x)y B 4 x), y ). Inequality 4) thus holds for the barrier function B 4 x) with C 3 =. One may also suspect that the cone consistence condition 4) is implied by the self-concordance property. However, again this is not the case. In Zhang [] it is shown that if bx) is a convex quadratic function in R n, then the barrier function B h x) = logq pbx/p)) is a -logarithmically homogeneous self-concordant barrier for the convex cone K h = cl x = p q x R ++ R R n q pbx/p) 0. Consider y = [, + b0), 0 T ] T int K h, and x int K h. One computes that B h x) = q pbx/p) bx/p) + bx/p) T x/p) bx/p) and B h x) = bx/p) + bx/p) T x/p) bx/p) + bx/p) T T x/p) q pbx/p)) bx/p) bx/p) x/p) T bx/p)x/p) 0 x/p) T bx/p) + pq pbx/p)) bx/p)x/p) 0 bx/p) Therefore, B h x), y = + b0) bx/p) + bx/p)t x/p) q pbx/p) > 0 5) 8

9 and Thus, from 5) and 6) we have B h x)[y, y] = B h x), y ) + x/p)t bx/p)x/p). 6) pq pbx/p)) B h x)[y, y] B h x), y ) = + q pbx/p) x/p) T bx/p)x/p) p + b0) bx/p) + bx/p) T x/p)). The above quantity can not be uniformly bounded from above for all x int K h if bx) is strictly convex, because q can take arbitrarily large value while p and x are fixed. This implies that for the barrier function B h x), the cone-consistence condition 4) is not satisfied, although the function is self-concordant and -logarithmically homogeneous. These two examples show that the new condition 4) is a related but different property as compared with the self-concordance property, the self-scaled property, or the logarithmic homogeneous property of the barrier functions. Consider P K ). Let F ) be a consistent self-concordant barrier function for the cone K satisfying ), 3) and 4) altogether. Let and px) := h 0 f j x)h j int K fx) := F px)). 7) Let ξ R n be any given direction. For a chosen coordinate system we have fx)[ξ] = F px)), f j x) T ξ)h j. 8) Consequently, by the chain-rule, fx)[ξ, ξ] = F px))[ f j x) T ξ)h j, f j x) T ξ)h j ] F px)), ξ T f j x)ξ)h j 9) and 3 fx)[ξ, ξ, ξ] = 3 F px))[ f j x) T ξ)h j, f j x) T ξ)h j, f j x) T ξ)h j ] +3 F px))[ ξ T f j x)ξ)h j, f j x) T ξ)h j ] 9

10 k 3 f j x)[ξ, ξ, ξ]h j, F px)). 0) Since f j s are convex functions, ξ T f j x)ξ 0 for all j =,..., k, and so ξ T f j x)ξ)h j K. Moreover, because F px)) int K we obtain fx)[ξ, ξ] F px))[ f j x) T ξ)h j, f j x) T ξ)h j ] C F px)), f j x) T ξ)h j = C fx)[ξ] ) which is the second inequality required by the self-concordance 3). Therefore we can use condition 4) to obtain from 9) that fx)[ξ, ξ] F px))[ f j x) T ξ)h j, f j x) T ξ)h j ] + C F px))[ ξ T f j x)ξ)h j, ξ T f j x)ξ)h j ]. ) 3 Denote and We rewrite ) as ) / d := F px)) f j x) T ξ)h j ) / d := F px)) ξ T f j x)ξ)h j. fx)[ξ, ξ] d + [ ) d d ] 3 d 3. 3) C 3 C 3 Theorem. Suppose that F ) is a self-concordant barrier function for K, and moreover it is consistent with K. Suppose that f j x) s are convex quadratic functions, j =,..., k. Then fx) as defined in 7) is a self-concordant barrier function for P K ). 0

11 Proof. If f j s are all quadratic functions then it follows from 0) that 3 fx)[ξ, ξ, ξ] C fx)[ξ, ξ]) 3/ + 3d T d C fx)[ξ, ξ]) 3/ + 3 d d C fx)[ξ, ξ]) 3/ + C 3 3 fx)[ξ, ξ]) 3/ = C + C 3 3) fx)[ξ, ξ]) 3/, 4) where the third inequality is due to 3). Inequalities 4) and ) show the self-concordance of the function fx). Q.E.D. 3 A potential reduction approach Theorem. suggests that if the conic barrier F is self-concordant and consistent with the cone, and f j s are all convex quadratic, then the composite barrier function f is self-concordant. Therefore we can apply the general theory of self-concordant barriers developed by Nesterov and Nemirovskii [5]. However, there are interesting cases that are not included in this model. Consider for instance the following geometric semidefinite program: P P SD ) minimize f 0 t) subject to l i f i t) u i, i =,..., k + m t j > 0, j =,..., n H 0 f j t)h j 0, where H 0,..., H k 0 are l l positive semidefinite matrices, and f i t) = t ι i tι in is a monomial, i =,..., k + m. After the variable transformation x j = log t j, j =,..., n, the above problem can be turned into P P SD ) minimize ι T 0 x subject to log l i ι T i x log u i, i =,..., k + m H 0 e ιt j x H j 0. Clearly, P P SD ) is a conically ordered convex program. Unfortunately, in this case, the function e ιt j x is not quadratic, and so Theorem. does not apply: the barrier function fx) is not selfconcordant in general. However, e ιt j x s are still well structured. In the above particular case, e ιt j x = e ιt j x ι j ι T j. Therefore, it holds that e ιt j x λ e ιt j y for all x and y in its feasible region, where λ = max{u /l,..., u k /l k }.

12 This example motivates us to study algorithms that can solve the COCP problems without an overall easy computable self-concordant barriers. As a matter of fact, most known interior point methods for nonlinear convex programs with provable polynomial-time computational complexities are constructed within the framework of the central-path following using a self-concordant barrier function, except for a perhaps less known member of the IPM family called the Iri-Imai method [3], which is a variant of the potential reduction method. Zhang [], and Sturm and Zhang [0] studied the Iri-Imai method or the potential reduction method) applied to the so-called harmonically convex programs, and show that such algorithms also run in polynomial time. The main proof techniques in [], and also proofs in this section, are inspired by a paper of Iri []. In this section, we shall study an Iri-Imai type potential reduction method for the conically ordered convex program P K ). First we mention the notion of harmonically convex functions; see []. Definition 3. A twice differentiable function f ) is called λ-harmonically convex on the convex domain Ω if there is a constant λ such that 0 fx) λ fy) holds for all x, y Ω. It is easy to see that if λ > 0 then λ fy) fx) λ fy) for all x, y Ω. Obviously, the affine linear functions 0-harmonically convex), the convex quadratic functions -harmonically convex), and the strongly convex functions, are all examples of harmonically convex functions. It is shown in [] that the direct) sum of these functions characterizes the whole class of harmonically convex functions. If a convex program is to minimize a harmonically convex objective function, subject to harmonically convex inequality constraints, then [] and [0] show that such problem can be solved by Iri-Imai type potential reduction algorithms in polynomial time. Note that even if fx) is harmonically convex and negative on a domain Ω, it does not follow that log fx)) is selfconcordant. Consider for example fx, x ) = x +x +sinx ), which is the sum of two harmonically convex functions, x and x +sinx ), hence harmonically convex. However, logx x sinx )) is not self-concordant. Consider P K ). Let F ) be a barrier function for K. Moreover, let us impose the following conditions. Condition 3. F ) is ν-logarithmically homogeneous for K, namely for all y int K and t > 0. F ty) = F y) ν log t It is well known see e.g. [5]) that if F is ν-logarithmically homogeneous for K, then the following

13 identities hold where y int K and t > 0: F ty) = F y); 5) t F ty) = t F y); 6) F y)y = F y); 7) F y)) T y = ν. 8) Condition 3. F ) is a self-concordant barrier function for K. In particular, the following two inequalities hold 3 F y)[u, u, u] F y)[u, u]) 3/ and F y)[u] ν F y)[u, u]) /, for all y int K and u span K. Notice that the second inequality in Condition 3. follows from 7), 8), and also the Cauchy- Schwarz inequality; it is in fact implied by Condition 3.. Like 7), denote fx) := F px)), with px) = h 0 k f j x)h j. Clearly, f ) is a barrier function for the feasible set for P K ). Lemma 3. Suppose that F ) is a ν-logarithmically homogeneous barrier function for K Condition 3.). Denote F to be the feasible set for P K ). Then for all x rel int F and ξ R n. fx)[ξ] ν ) / fx)[ξ, ξ] Proof. As x rel int F we have px) = h 0 k f j x)h j int K. Relations 8) and 9) give that fx)[ξ] = F px)), f j x) T ξ)h j 3

14 and fx)[ξ, ξ] = F px))[ f j x) T ξ)h j, f j x) T ξ)h j ] F px)), ξ T f j x)ξ)h j. Since F px)) K and k ξ T f j x)ξ)h j K due to the convexity of f j s, we obtain from the above equation that fx)[ξ, ξ] F px))[ f j x) T ξ)h j, f j x) T ξ)h j ] ν F px)), f j x) T ξ)h j = ν fx)[ξ], where the second step is due to Condition 3.. Q.E.D. Suppose that P K ) has a finite optimal value denoted by v. Let v v. For a give parameter γ > 0, let us introduce Gx; v) := c T x v) γ expfx)). 9) Note that Gx; v) is an extension of the so-called multiplicative barrier function of Iri and Imai; see [3]. The following theorem extends Iri and Imai s result. Theorem 3.3 Suppose that F ) is a ν-logarithmically homogeneous barrier function for K Condition 3.). If γ ν + then Gx; v) is a convex function on int F; if γ > ν + then Gx; v) is a strictly convex function on int F. Proof. Differentiation yields Gx; v) Gx; v) = γ c T c + fx) x v 30) Gx; v) Gx; v) = Gx; v) Gx; v) Gx; γ v)t c T x v) cct + fx). 3) Take any 0 η R n. Denote a := c T η/c T x v), b := η T fx) and e := η T fx)η > 0. By Lemma 3. we have e b /ν. Using 30) and 3) it follows that Gx; v) ηt Gx; v)η = γa + b) γa + e 4

15 = γ γ)a + γab + b + e γ γ)a + γab + + /ν)b. The last expression is a quadratic form with respect to a and b, which is nonnegative as its discriminant is γ γ γ)+/ν) = γν+ γ)/ν 0. The quantity η T Gx; v)η is strict positive if γ > ν+, because in this case the discriminant of the quadratic form is negative while e > 0. The theorem is thus proven. Q.E.D. For simplicity of the analysis, from now on we shall set γ to be ν + ν. Condition 3.3 F ) is κ-consistent κ > 0) with the cone K, i.e., for all y int K and u K. F y)[u, u]) / κ F y)[u] Condition 3.4 All f j ) s, j =,..., k, are harmonically convex on F. In particular, there is λ > 0 such that f j x) λ f j x) for all x, y F and j =,..., k. Condition 3.5 F ) satisfies F y)[u, v] 0 for all y int K and u, v K. F y)k K for all y int K. In other words, A few comments are in order here, concerning the last condition. If K is a self-scaled cone and F ) is the corresponding self-scaled barrier function, then as a part of the definition it holds that F y)k = K for all y int K. In that respect, Condition 3.5 is naturally satisfied by all selfscaled cones with the corresponding self-scaled barrier functions. Thus, one may conjecture that Condition 3.5 is satisfied if and only if K is a self-scaled cone. However, this is not the case. The key is that we only require F y)k K here, rather than F y)k = K for all y int K. In fact, K can be any pointed closed convex cone. To see this, let C be a second order cone that properly strictly) contain K. Therefore, C is strictly) properly contained in K. Let F c be the corresponding self-scaled barrier function for C. Thus, for any y int K int C we have F c y)k F c y)c = C K. For a given barrier function F ) for K, consider F µ ) := F )+µf c ), where µ > 0 is a given constant. Clearly, F µ ) is also a barrier function for K. If µ is chosen sufficiently large then F µ y)k K for any y int K. 5

16 Lemma 3.4 Suppose that F ) satisfies Condition 3.5, and v i, i =,..., m, are m arbitrary vectors in K. Let y int K and F y) 0 and x F y) := x T F y)x. Then, m max t i v i i= 0 t i ˆt i, i =,..., m = m ˆt i v i. F y) i= F y) Proof. Under Condition 3.5, we have vi T F y)v j 0 for any i and j. The lemma is obvious since m m t i v i = t i t j vi T F y)v j. i= F y) i, Q.E.D. Theorem 3.5 Suppose that F ) satisfies Conditions 3., 3.3, and 3.5, and f j ) s, j =,..., k, satisfy Condition 3.4. Let x rel int F be any feasible but non-optimal point. Let γ = ν + ν. Let the Newton direction at x be Then d N x) := Gx; v )) Gx; v ). Gx; v ) T d N x)/gx; v ) = Gx; v ) T Gx; v )) Gx; v )/Gx; v ) δ, where δ := / + κ + λκ > 0. Proof. Let δx) := Gx; v ) T Gx; v )) Gx; v )/Gx; v ). Since Gx; v ) 0, by the Cauchy-Schwarz inequality we know that for any η 0 it holds δx) Gx; v ) T η/gx; v ) η T Gx; v )η/gx; v ). 3) Let us take η = x x 0. Then, noting c T x x )/c T x v ) = and using 30) and 3) we have Gx; v ) T x x )/Gx; v ) = γ + fx) T x x ). 33) By the mean-value theorem, there is x j between x and x, j =,..., k, such that px) px ) = f j x) T x x )h j + x x ) T f j x j )x x )h j. 6

17 By letting we have ũ = x x ) T f j x j )x x )h j K 34) fx) T x x ) = F px))[ f j x) T x x )h j ] = F px))[px) px ) ũ] where in the last step we used the property 8). = ν F px))[px ) + ũ], 35) Using the estimation 35) in 33) and the fact that px), px ), ũ K we have Gx; v ) T x x )/Gx; v ) = ν F px))[px ) + ũ] ν + /κ) F px))[px ) + ũ, px ) + ũ]. 36) Similarly, since F px))[px )] 0, we have Gx; v ) T x x )/Gx; v ) ν F px))[ũ] ν + /κ) F px))[ũ, ũ]. 37) On the other hand, ) x x ) T Gx; v )x x )/Gx; v ) Gx; v ) T x x )/Gx; v ) = γ + x x ) T fx)x x ) = γ + F px))[ f j x) T x x )h j, f j x) T x x )h j ] F px))[ x x ) T f j x) T x x )h j ] γ + F px))[px) px ) ũ, px) px ) ũ] + ν F px))[ x x ) T f j x) T x x )h j, x x ) T f j x) T x x )h j ] ) γ + F px))[px), px)] + F px))[px ) + ũ, px ) + ũ] + λ ν F px))[ũ, ũ] ) γ + ν + κ Gx; v ) T x x )/Gx; v ) ) + λκ ν Gx; v ) T x x )/Gx; v ) + κ + λκ) Gx; v ) T x x )/Gx; v )), 38) 7

18 where the first step is due to 3) pre and post multiplying both sides by x x ) T and x x respectively), and at the fourth step we used x x ) T f j x) T x x ) λx x ) T f j x j ) T x x ), j =,..., k, due to Condition 3.4, and then applied Lemma 3.4, and finally at the fifth and the sixth steps we used 36) and 37). The theorem follows by combining 3) and 38) to obtain δx) δ. Q.E.D. Now let us study the effect of applying the Newton method with line-minimization) to reduce the value of Gx; v) where v v. Denote d := Gx; v)) Gx; v). Multiplying d T on both sides of 30), and pre-multiplying d T and post-multiplying d on both sides of 3), and denoting a := c T d/c T x v) and, if there is no confusion, still denoting δx) := Gx; v) T Gx; v) ) Gx; v)/gx; v), we obtain Eliminating a from 39) and 40) we get δx) = γa + d T fx) 39) δx) δx) = γa + d T fx)d. 40) γδx) δx)) + δx) + d T fx)) = γd T fx)d. 4) Relations 8) and 9) yield fx) T d = F px))[ d T f j x)h j ] 4) d T fx)d = F px))[ d T f j x)h j, d T f j x)h j ] F px))[ d T f j x)dh j ]. 43) Since F ) is ν-logarithmically homogeneous and f j s are convex, we obtain from 4) and 43) that γδx) δx)) + δx) + d T fx)) γ F px))[ d T f j x)h j, d T f j x)h j ] γ F px))[ d T f j x)h j ] ν = γ d fx)) T. 44) ν 8

19 Noting that γ = ν + ν, because d T fx) is a real number, checking the discriminant of the above quadratic inequality in terms of d T fx) we conclude that The quadratic inequality 44) can be rearranged as with δx) ν + ν ν. 45) α x) d T fx) α x), 46) α x) := νδx) ν γ ν )δx)) δx) α x) := νδx) + ν γ ν )δx)) δx). Simplifying, we get α x) ν + ν ν ν + ν + ν ν ν =: α and α x) α x) α. Now we use the κ-consistency, Condition 3.3, to obtain from 4) and 43) that κ F px))[ d T f j x)dh j, d T f j x)dh j ] F px))[ d T f j x)dh j ] d T fx)d = γδx) δx)) + δx) + dt fx)) γ γδx) + δx) + d T fx)) γ γ ν + γ ν ) + α γ =: α 3, 47) where the second step is due to 43), the third step is due to 4), and the last step is due to 45) and 46). This yields F px))[ d T f j x)dh j, d T f j x)dh j ] κ α3. 48) 9

20 Along the same line, from 43) and 47) we get F px))[ d T f j x)h j, d T f j x)h j ] d T fx)d α 3. 49) Consider an iterative point x + td along the Newton direction d with step size t > 0. Using the mean-value theorem, there are x j [x, x + td], j =,..., k, such that Let ũ := k d T f j x j )dh j. px + td) px) = f j x + td) f j x)) h j = t d T f j x)h j t d T f j x j )dh j. 50) Remember that the norm induced by the Hessian of F at y is denoted by u F y) = u T F y)u. The lemma below is well known and it is a crucial property of the self-concordant barrier functions; see [4] and [7]. Lemma 3.6 Suppose that F ) is self-concordant, i.e., it satisfies Condition 3.. Suppose that px) int K. If y px) F px)) < then y int K. Moreover, for all u. u F y) u F px)), y px) F px)) In this notation, 48) can be written as d T f j x)dh j F px)) κα 3, 5) and 49) as d T f j x)h j α 3. 5) F px)) Lemma 3.7 Suppose that F ) satisfies Conditions 3., 3., 3.3, and 3.5, and f j ) s, j =,..., k, satisfy Condition 3.4. Let x rel int F be any feasible but non-optimal point. Let γ = ν + ν. Let the Newton direction at x be d = Gx; v) ) Gx; v) with v v. Let α 4 := α3 + + λκ). 0

21 Then, for any step-size 0 < t < α 4, the iterative point px + td) F. Proof. It follows from 50) that px + td) px) = t d T f j x)h j t ũ. By Lemma 3.4 and Condition 3.4, also noting 5) we have ũ F px)) λ d T f j x)dh j λκα 3. F px)) Using the above estimation and 5) we know that if 0 < t < α 4 then px + td) px) F px)) t d T f j x)h j + t ũ F px)) F px)) t α 3 + t λκα 3 <, because α 4 is the largest root of the quadratic equation t α 3 + t λκα 3 = 0. By Lemma 3.6, the result follows. Q.E.D. Remark that, in a similar way one can show that if 0 < t < α 5 with α 5 := α3 + + λκ), then t d T f j x)h j + t ũ F px)) t α 3 + t λκα 3 < /. 53) F px)) Theorem 3.8 Suppose that F ) satisfies Conditions 3., 3., 3.3, and 3.5, and f j ) s, j =,..., k, satisfy Condition 3.4. Let x rel int F be any feasible but non-optimal point. Let γ = ν + ν. Let the Newton direction at x be d = Gx; v) ) Gx; v). Suppose that Gx; v) T d/gx; v) δ. Then there exists β > 0 such that min log Gx + td; v) log Gx; v)) β. 0 t α 5

22 Proof. Denote xt) := x + td, and let a := ct d. Lemma 3.7 and the remark thereafter assure c T x v that if 0 < t < α 5 < α 4 ) then xt) F, and pxt)) px) F px)) < /. Now, consider a fixed 0 < t < α 5. By the mean-value theorem, there is 0 t t < α 5 such that the following estimation holds log Gx + td; v) log Gx; v) = γ log + at) + fxt)) fx) ) = γa + fx) T d)t + t a γ + a t) + dt fx t))d δx)t + t F px t)))[ d T f j x t))h j, d T f j x t))h j ] + t ν F px t)))[ d T f j x t))dh j, d T f j x t))dh j ] δt + t k d T f j x t))h j F px)) px t)) px) ) + k d T f j x t))dh j ν F px)) px t)) px). F px)) F px)) 54) Notice that d T f j x t))h j F px)) d T f j x)h j + λ d T f j x)dh j F px)) F px)) d T f j x)h j + λ 4 d T f j x)dh j F px)) F px)) α 3 + λ κ α 3/, where the last step follows from 5) and 5), and d T f j x t))dh j λ d T f j x)dh j F px)) F px)) The coefficient of the second order term in 54) can be further estimated as k d T f j x t))h j F px)) px t)) px) F px)) ) + ν λκα 3. k d T f j x t))dh j F px)) px t)) px) F px))

23 8α 3 + λ κ α 3 + λκα 3 ν =: α6. Using these estimations we rewrite 54) as follows log Gx + td; v) log Gx; v) δt + α 6 t. Therefore, if we take the step-length as t := min{ δ α 6, α 5 }, then log Gx + td; v) log Gx; v) β, where β := δ t α 6 t. Q.E.D. If K is a symmetric cone, say K = S l + and F p) = log det p, then we can improve the bound in 54). In that case, we have F p)[h] = tr p h F p)[h, h] = tr p hp h). Therefore, if p p F p) = F p)[ p p, p p] = p p p)p F = p pp I F < / then p p 3 p, and since h K we have Instead of the estimations in 54), we now have 0 < F p)[h] = tr p h tr p h = F p)[h]. log Gx + td; v) log Gx; v) = γ log + at) + fxt)) fx) ) = γa + fx) T d)t + t a γ + a t) + dt fx t))d δx)t + t F px t)))[ d T f j x t))h j, d T f j x t))h j ] t F px t)))[ d T f j x t))dh j ] δt + t k d T f j x t))h j F px)) px t)) px) ) λ F px))[ d T f j x)dh j ] F px)) δt + t 8α 3 + λ κ α 3 + λα 3 ). By letting α 7 := 8α 3 + λ κ α 3 + λα 3, and t := min{ δ α 7, α 5 } we have the estimation that log Gx + t d; v) log Gx; v) β with β := δ t α 7 t ). 3

24 One easily verifies that all the constants, α to α 7, δ, β and β, are rationally related to the problem parameters ν, κ and λ. This gives rise to a polynomial-time complexity proof for the following extended version of the Iri-Imai method or, potential reduction method). An Iri-Imai Type Potential Reduction Method for COCP Input data: An initial x 0 rel int F and an initial lower bound of the optimal value v 0 v. Output: A feasible solution x rel int F such that c T x v ɛ where ɛ > 0 is the desired precision. Step 0. Let x := x 0 and v := v 0. Go to Step. Step. Check whether c T x v ɛ or not. If yes, stop with x as the output; if no, continue with Step. Step. Compute the Newton direction at x d := Gx; v)) Gx; v). If Gx; v) T d/gx; v) < δ then increase the value v until Gx; v) T d/gx; v) = δ; otherwise continue with Step 3. Step 3. Apply the line-minimization for log Gx; v) along the direction d, starting from x. Let t = argmin t 0 log Gx + td; v). Let x := x + td, and go to Step. The next theorem is a consequence of Theorem 3.8 and the remarks thereafter. Theorem 3.9 Suppose that F ) satisfies Conditions 3., 3., 3.3, and 3.5, and f j ) s, j =,..., k, satisfy Condition 3.4. Let γ = ν + ν. Then the above described Iri-Imai type potential reduction algorithm terminates in a number of steps, which is bounded by a polynomial in terms of ν, κ, λ and log ɛ, provided that i) an optimal solution for P K) exists; ii) fx) L for all x F, where L is bounded by a polynomial in ν, κ and λ; iii) the initial potential function value log Gx 0 ; v 0 ) is bounded by a polynomial in ν, κ, λ, and log ɛ. 4

25 Proof. Let the iterates produced by the algorithm be {x 0, x,, x j, }, and the sequence of the lower bounds be {v 0, v,, v j, }. By Theorem 3.8 we have log Gx j ; v j ) log Gx j ; v j ) β log Gx 0 ; v 0 ) jβ where /β is bounded from above by a polynomial in ν, κ and λ. Thus, Therefore, if then it follows from 55) that logc T x j v j ) log Gx0 ; v 0 ) jβ fx j ) γ j log Gx0 ; v 0 ) + L + γ log/ɛ) β c T x j v j ɛ. log Gx0 ; v 0 ) jβ + L. 55) γ Q.E.D. We observe that Theorem 3.9 provides explicit computational complexity bounds for some new classes of convex programming problems. To give one such example, let us call a convex function fx) to be quadratically dominated if it can be split into two parts: fx) = qx) + hx) where qx) is convex quadratic and hx) is any other function whose Hessian is dominated by qx) = Q 0, i.e., there is 0 θ < such that θq hx) θq. Our example at the beginning of Section 3, fx) = x + sin x, is such a convex quadratically dominated function. Now consider the following COCP: K is a symmetric cone κ = ); f j s are convex quadratically dominated functions all with a uniform and constant bound θ). Since convex quadratically dominated functions are harmonically convex with λ = + θ)/ θ), we have λ = O), as θ is assumed to be fixed here. It follows that α = ν+ν ν α 3 = ν ν + ν+ ν γ ν + γ ν ) +α α 4 = α3 + +λκ) ν O ν) γ O) O) α 5 = α3 + +λκ) O) α 7 = 8α 3 + λ κ α3 + λα 3 O) δ = / + κ + λκ O) t = min{ δ α 7, α 5 } O) β = δ t α 7 t ) O). Therefore, the Iri-Imai type potential reduction algorithm would solve this problem in Oν log ɛ ) number of iterations to reach an ɛ-optimal solution if ɛ > 0 is small enough, as asserted by Theorem 3.9. Similarly, the geometric semidefinite program P P SD ) that we introduced at the beginning of this section can be solved in Om + k + l) log ɛ ) number of iterations if max i k u i /l i is a constant. 5

26 Even if K is not a symmetric cone assuming κ is fixed however) and f j s are either quadratically dominated convex functions or they are generated from monomials, since α 6 = 8α 3 + λ κ α 3 + λκα 3 ν = O ν) we can still apply Theorem 3.8 and Theorem 3.9 to conclude that these problems can be solved in no more than Oν.5 log ɛ ) number of iterations to reach an ɛ-optimal solution. 4 Duality and lower bounds What remains to be discussed is how to obtain a lower bound v v in order to apply the Iri-Imai type potential reduction method as introduced in the previous section. This naturally leads to the duality issues. Consider the following COCP problem: P K ) minimize c T x subject to Ax = b h 0 f j x)h j K, where A R m n, c R n and b R m, h, h,..., h k K, and f j ) s are convex functions. For this purpose, let us note that the conjugate of a convex function fx), which is also a convex function, is defined as f s) = sup{ s) T x fx) x dom f}, where dom f stands for the domain of the function f. Define a closed convex cone as follows C := cl p q x p > 0, q pfx/p) 0. Then, the dual of C can be computed as u C = cl v s v > 0, u vf s/v) 0. One is referred to [] for discussions on this and other related issues. Let K j := cl p j q j x j p j > 0, q j p j f j x j /p j ) 0, j =,..., k. 6

27 Then, K j = cl u j v j s j v j > 0, u j v j fj s j /v j ) 0, j =,..., k. In light of these new cones, we may rewrite our original COCP as P K ) minimize c T x subject to Ax = b q j x K j, j =,..., k h 0 q j h j K. The above description is in the form of linear conic optimization. Hence its dual problem can be found by standard procedures. Let us derive the dual using the Lagrangian multipliers. In particular, let y be the multipliers for the first set of equality constraints, and u j v j s j K j be the multipliers for the second set of conic) constraints. Let z K be the multiplier for the last conic constraint. Then the Lagrangian function is Lx, q; y, u, v, s, z) = c T x y T Ax b) u j + v j q j + s T j x) z T h 0 q j h j ), from which the dual problem can be derived as D K ) maximize b T y h T 0 z u j subject to A T y + s j = c u j h T j z K j, j =,..., k s j z K. 7

28 This problem can be further explicitly expressed as D K ) maximize b T y h T 0 z h T j z)fj s j /h T j z) subject to A T y + s j = c z K. Therefore, a lower bound v v for P K ) could be obtained from a finite feasible solution for the above problem D K ). If f j x) = xt Q j x with Q j 0 for all j, then fj s) = st Q j s, and the dual problem becomes maximize b T y h T 0 z subject to A T y + s j = c z K. s T j Q j s j /h T j z In the case of conically ordered geometric programming, where f j x) = e ιt j x for all j. It is easy to compute that Therefore, its dual is fj +, s j ) = s j ι j log s j ι j s j ι j, maximize subject to b T y h T 0 z A T y [ι,..., ι k ]w = c z K, w 0. if s j s j ι j ι j if s j = s j ι j ι j. ) w j log w j log h T j z + w j References [] L. Faybusovich, Jordan algebra, symmetric cones and interior point methods, Research Report, University of Notre Dame, 995. [] M. Iri, A proof of the polynomiality of the Iri-Imai method, Journal of Complexity 9, 69 90, 993. [3] M. Iri and H. Imai, A multiplicative barrier function method for linear programming, Algorithmica, ,

29 [4] Yu. Nesterov, Introductory lectures on convex programming, Kluwer Academic Publishers, Dordrecht, 004. [5] Yu. Nesterov and A. Nemirovskii, Interior point polynomial methods in convex programming, Studies in Applied Mathematics 3, SIAM, Philadelphia, PA, 994. [6] Yu. Nesterov and M.J. Todd, Self scaled barriers and interior point methods for convex programming, Mathematics of Operations Research, 4, 997. [7] J. Renegar, A mathematical view of interior-point methods in convex optimization, MPS-SIAM Series on Optimization, 00. [8] J.F. Sturm, Unifying semi-definite and second-order cone programming: symetric cones, Research Report, 998. [9] J.F. Sturm, Similarity and other special relations for symmetric cones, Linear Algebra and its Applications 3, 35 54, 000. [0] J.F. Sturm and S. Zhang, A potential reduction method for harmonically convex programming, Journal of Optimization Theory and Applications 84, 8 05, 995. [] S. Zhang, Convergence property of the Iri-Imai algorithm for some smooth convex programming problems, Journal of Optimization Theory and Applications 8, 38, 994. [] S. Zhang, A new self-dual embedding method for convex programming, Journal of Global Optimization 9, ,

Lecture 5. The Dual Cone and Dual Problem

Lecture 5. The Dual Cone and Dual Problem IE 8534 1 Lecture 5. The Dual Cone and Dual Problem IE 8534 2 For a convex cone K, its dual cone is defined as K = {y x, y 0, x K}. The inner-product can be replaced by x T y if the coordinates of the

More information

Supplement: Universal Self-Concordant Barrier Functions

Supplement: Universal Self-Concordant Barrier Functions IE 8534 1 Supplement: Universal Self-Concordant Barrier Functions IE 8534 2 Recall that a self-concordant barrier function for K is a barrier function satisfying 3 F (x)[h, h, h] 2( 2 F (x)[h, h]) 3/2,

More information

A New Self-Dual Embedding Method for Convex Programming

A New Self-Dual Embedding Method for Convex Programming A New Self-Dual Embedding Method for Convex Programming Shuzhong Zhang October 2001; revised October 2002 Abstract In this paper we introduce a conic optimization formulation to solve constrained convex

More information

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 1 Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 2 A system of linear equations may not have a solution. It is well known that either Ax = c has a solution, or A T y = 0, c

More information

An interior-point trust-region polynomial algorithm for convex programming

An interior-point trust-region polynomial algorithm for convex programming An interior-point trust-region polynomial algorithm for convex programming Ye LU and Ya-xiang YUAN Abstract. An interior-point trust-region algorithm is proposed for minimization of a convex quadratic

More information

Self-Concordant Barrier Functions for Convex Optimization

Self-Concordant Barrier Functions for Convex Optimization Appendix F Self-Concordant Barrier Functions for Convex Optimization F.1 Introduction In this Appendix we present a framework for developing polynomial-time algorithms for the solution of convex optimization

More information

On self-concordant barriers for generalized power cones

On self-concordant barriers for generalized power cones On self-concordant barriers for generalized power cones Scott Roy Lin Xiao January 30, 2018 Abstract In the study of interior-point methods for nonsymmetric conic optimization and their applications, Nesterov

More information

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming School of Mathematics T H E U N I V E R S I T Y O H F E D I N B U R G Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming Jacek Gondzio Email: J.Gondzio@ed.ac.uk URL: http://www.maths.ed.ac.uk/~gondzio

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization

More information

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization Iranian Journal of Operations Research Vol. 4, No. 1, 2013, pp. 88-107 Research Note A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization B. Kheirfam We

More information

Largest dual ellipsoids inscribed in dual cones

Largest dual ellipsoids inscribed in dual cones Largest dual ellipsoids inscribed in dual cones M. J. Todd June 23, 2005 Abstract Suppose x and s lie in the interiors of a cone K and its dual K respectively. We seek dual ellipsoidal norms such that

More information

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method L. Vandenberghe EE236C (Spring 2016) 1. Gradient method gradient method, first-order methods quadratic bounds on convex functions analysis of gradient method 1-1 Approximate course outline First-order

More information

A Value Estimation Approach to the Iri-Imai Method for Constrained Convex Optimization. April 2003; revisions on October 2003, and March 2004

A Value Estimation Approach to the Iri-Imai Method for Constrained Convex Optimization. April 2003; revisions on October 2003, and March 2004 A Value Estimation Approach to the Iri-Imai Method for Constrained Convex Optimization Szewan Lam, Duan Li, Shuzhong Zhang April 2003; revisions on October 2003, and March 2004 Abstract In this paper,

More information

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1 L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones definition spectral decomposition quadratic representation log-det barrier 18-1 Introduction This lecture: theoretical properties of the following

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

Nonsymmetric potential-reduction methods for general cones

Nonsymmetric potential-reduction methods for general cones CORE DISCUSSION PAPER 2006/34 Nonsymmetric potential-reduction methods for general cones Yu. Nesterov March 28, 2006 Abstract In this paper we propose two new nonsymmetric primal-dual potential-reduction

More information

5. Duality. Lagrangian

5. Duality. Lagrangian 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Constructing self-concordant barriers for convex cones

Constructing self-concordant barriers for convex cones CORE DISCUSSION PAPER 2006/30 Constructing self-concordant barriers for convex cones Yu. Nesterov March 21, 2006 Abstract In this paper we develop a technique for constructing self-concordant barriers

More information

Convex Optimization Boyd & Vandenberghe. 5. Duality

Convex Optimization Boyd & Vandenberghe. 5. Duality 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min. MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.

More information

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems AMSC 607 / CMSC 764 Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 4: Introduction to Interior Point Methods Dianne P. O Leary c 2008 Interior Point Methods We ll discuss

More information

What can be expressed via Conic Quadratic and Semidefinite Programming?

What can be expressed via Conic Quadratic and Semidefinite Programming? What can be expressed via Conic Quadratic and Semidefinite Programming? A. Nemirovski Faculty of Industrial Engineering and Management Technion Israel Institute of Technology Abstract Tremendous recent

More information

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization Yinyu Ye Department is Management Science & Engineering and Institute of Computational & Mathematical Engineering Stanford

More information

New stopping criteria for detecting infeasibility in conic optimization

New stopping criteria for detecting infeasibility in conic optimization Optimization Letters manuscript No. (will be inserted by the editor) New stopping criteria for detecting infeasibility in conic optimization Imre Pólik Tamás Terlaky Received: March 21, 2008/ Accepted:

More information

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS Sercan Yıldız syildiz@samsi.info in collaboration with Dávid Papp (NCSU) OPT Transition Workshop May 02, 2017 OUTLINE Polynomial optimization and

More information

12. Interior-point methods

12. Interior-point methods 12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity

More information

Lecture 6: Conic Optimization September 8

Lecture 6: Conic Optimization September 8 IE 598: Big Data Optimization Fall 2016 Lecture 6: Conic Optimization September 8 Lecturer: Niao He Scriber: Juan Xu Overview In this lecture, we finish up our previous discussion on optimality conditions

More information

A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming

A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming Chek Beng Chua Abstract This paper presents the new concept of second-order cone approximations for convex conic

More information

15. Conic optimization

15. Conic optimization L. Vandenberghe EE236C (Spring 216) 15. Conic optimization conic linear program examples modeling duality 15-1 Generalized (conic) inequalities Conic inequality: a constraint x K where K is a convex cone

More information

Primal-dual IPM with Asymmetric Barrier

Primal-dual IPM with Asymmetric Barrier Primal-dual IPM with Asymmetric Barrier Yurii Nesterov, CORE/INMA (UCL) September 29, 2008 (IFOR, ETHZ) Yu. Nesterov Primal-dual IPM with Asymmetric Barrier 1/28 Outline 1 Symmetric and asymmetric barriers

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond Tor Myklebust Levent Tunçel September 26, 2014 Convex Optimization in Conic Form (P) inf c, x A(x) = b, x

More information

ON IMPLEMENTATION OF A SELF-DUAL EMBEDDING METHOD FOR CONVEX PROGRAMMING

ON IMPLEMENTATION OF A SELF-DUAL EMBEDDING METHOD FOR CONVEX PROGRAMMING ON IMPLEMENTATION OF A SELF-DUAL EMBEDDING METHOD FOR CONVEX PROGRAMMING JOHNNY TAK WAI CHENG and SHUZHONG ZHANG Department of Systems Engineering and Engineering Management, The Chinese University of

More information

Interior Point Methods in Mathematical Programming

Interior Point Methods in Mathematical Programming Interior Point Methods in Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Brazil Journées en l honneur de Pierre Huard Paris, novembre 2008 01 00 11 00 000 000 000 000

More information

Applications of Linear Programming

Applications of Linear Programming Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal

More information

MATH 4211/6211 Optimization Basics of Optimization Problems

MATH 4211/6211 Optimization Basics of Optimization Problems MATH 4211/6211 Optimization Basics of Optimization Problems Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 A standard minimization

More information

Augmented self-concordant barriers and nonlinear optimization problems with finite complexity

Augmented self-concordant barriers and nonlinear optimization problems with finite complexity Augmented self-concordant barriers and nonlinear optimization problems with finite complexity Yu. Nesterov J.-Ph. Vial October 9, 2000 Abstract In this paper we study special barrier functions for the

More information

Second-order cone programming

Second-order cone programming Outline Second-order cone programming, PhD Lehigh University Department of Industrial and Systems Engineering February 10, 2009 Outline 1 Basic properties Spectral decomposition The cone of squares The

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Lecture: Duality.

Lecture: Duality. Lecture: Duality http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Introduction 2/35 Lagrange dual problem weak and strong

More information

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE Yugoslav Journal of Operations Research 24 (2014) Number 1, 35-51 DOI: 10.2298/YJOR120904016K A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE BEHROUZ

More information

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009 LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix

More information

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

The maximal stable set problem : Copositive programming and Semidefinite Relaxations The maximal stable set problem : Copositive programming and Semidefinite Relaxations Kartik Krishnan Department of Mathematical Sciences Rensselaer Polytechnic Institute Troy, NY 12180 USA kartis@rpi.edu

More information

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint Iranian Journal of Operations Research Vol. 2, No. 2, 20, pp. 29-34 A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint M. Salahi Semidefinite

More information

Nonlinear Programming Models

Nonlinear Programming Models Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:

More information

Local Self-concordance of Barrier Functions Based on Kernel-functions

Local Self-concordance of Barrier Functions Based on Kernel-functions Iranian Journal of Operations Research Vol. 3, No. 2, 2012, pp. 1-23 Local Self-concordance of Barrier Functions Based on Kernel-functions Y.Q. Bai 1, G. Lesaja 2, H. Mansouri 3, C. Roos *,4, M. Zangiabadi

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Cubic regularization of Newton s method for convex problems with constraints

Cubic regularization of Newton s method for convex problems with constraints CORE DISCUSSION PAPER 006/39 Cubic regularization of Newton s method for convex problems with constraints Yu. Nesterov March 31, 006 Abstract In this paper we derive efficiency estimates of the regularized

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

4. Convex optimization problems

4. Convex optimization problems Convex Optimization Boyd & Vandenberghe 4. Convex optimization problems optimization problem in standard form convex optimization problems quasiconvex optimization linear optimization quadratic optimization

More information

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008. 1 ECONOMICS 594: LECTURE NOTES CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS W. Erwin Diewert January 31, 2008. 1. Introduction Many economic problems have the following structure: (i) a linear function

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones Research Reports on Mathematical and Computing Sciences Series B : Operations Research Department of Mathematical and Computing Sciences Tokyo Institute of Technology 2-12-1 Oh-Okayama, Meguro-ku, Tokyo

More information

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization H. Mansouri M. Zangiabadi Y. Bai C. Roos Department of Mathematical Science, Shahrekord University, P.O. Box 115, Shahrekord,

More information

4. Algebra and Duality

4. Algebra and Duality 4-1 Algebra and Duality P. Parrilo and S. Lall, CDC 2003 2003.12.07.01 4. Algebra and Duality Example: non-convex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone

More information

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples

More information

Lecture 9 Sequential unconstrained minimization

Lecture 9 Sequential unconstrained minimization S. Boyd EE364 Lecture 9 Sequential unconstrained minimization brief history of SUMT & IP methods logarithmic barrier function central path UMT & SUMT complexity analysis feasibility phase generalized inequalities

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - A) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

A Potential Reduction Method for Harmonically Convex Programming 1

A Potential Reduction Method for Harmonically Convex Programming 1 A Potential Reduction Method for Harmonically Convex Programming 1 J. F. Sturm 2 and S. Zhang 3 Communicated by Z.Q. Luo 1. The authors like to thank Dr. Hans Nieuwenhuis for carefully reading this paper

More information

Full Newton step polynomial time methods for LO based on locally self concordant barrier functions

Full Newton step polynomial time methods for LO based on locally self concordant barrier functions Full Newton step polynomial time methods for LO based on locally self concordant barrier functions (work in progress) Kees Roos and Hossein Mansouri e-mail: [C.Roos,H.Mansouri]@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Inequality Constraints

Inequality Constraints Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the

More information

Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications

Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications Linear-quadratic control problem with a linear term on semiinfinite interval: theory and applications L. Faybusovich T. Mouktonglang Department of Mathematics, University of Notre Dame, Notre Dame, IN

More information

Interior Point Methods for Mathematical Programming

Interior Point Methods for Mathematical Programming Interior Point Methods for Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Florianópolis, Brazil EURO - 2013 Roma Our heroes Cauchy Newton Lagrange Early results Unconstrained

More information

Lecture: Convex Optimization Problems

Lecture: Convex Optimization Problems 1/36 Lecture: Convex Optimization Problems http://bicmr.pku.edu.cn/~wenzw/opt-2015-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Introduction 2/36 optimization

More information

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx. Two hours To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER CONVEX OPTIMIZATION - SOLUTIONS xx xxxx 27 xx:xx xx.xx Answer THREE of the FOUR questions. If

More information

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given. HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard

More information

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Lagrange Duality Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Lagrangian Dual function Dual

More information

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Distributed Newton Method for Network Utility Maximization, II: Convergence A Distributed Newton Method for Network Utility Maximization, II: Convergence Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie October 31, 2012 Abstract The existing distributed algorithms for Network Utility

More information

Interior-Point Methods

Interior-Point Methods Interior-Point Methods Stephen Wright University of Wisconsin-Madison Simons, Berkeley, August, 2017 Wright (UW-Madison) Interior-Point Methods August 2017 1 / 48 Outline Introduction: Problems and Fundamentals

More information

Complexity bounds for primal-dual methods minimizing the model of objective function

Complexity bounds for primal-dual methods minimizing the model of objective function Complexity bounds for primal-dual methods minimizing the model of objective function Yu. Nesterov July 4, 06 Abstract We provide Frank-Wolfe ( Conditional Gradients method with a convergence analysis allowing

More information

Introduction to Nonlinear Stochastic Programming

Introduction to Nonlinear Stochastic Programming School of Mathematics T H E U N I V E R S I T Y O H F R G E D I N B U Introduction to Nonlinear Stochastic Programming Jacek Gondzio Email: J.Gondzio@ed.ac.uk URL: http://www.maths.ed.ac.uk/~gondzio SPS

More information

EE 227A: Convex Optimization and Applications October 14, 2008

EE 227A: Convex Optimization and Applications October 14, 2008 EE 227A: Convex Optimization and Applications October 14, 2008 Lecture 13: SDP Duality Lecturer: Laurent El Ghaoui Reading assignment: Chapter 5 of BV. 13.1 Direct approach 13.1.1 Primal problem Consider

More information

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Nonlinear equations. Norms for R n. Convergence orders for iterative methods Nonlinear equations Norms for R n Assume that X is a vector space. A norm is a mapping X R with x such that for all x, y X, α R x = = x = αx = α x x + y x + y We define the following norms on the vector

More information

Advances in Convex Optimization: Theory, Algorithms, and Applications

Advances in Convex Optimization: Theory, Algorithms, and Applications Advances in Convex Optimization: Theory, Algorithms, and Applications Stephen Boyd Electrical Engineering Department Stanford University (joint work with Lieven Vandenberghe, UCLA) ISIT 02 ISIT 02 Lausanne

More information

Implicit Functions, Curves and Surfaces

Implicit Functions, Curves and Surfaces Chapter 11 Implicit Functions, Curves and Surfaces 11.1 Implicit Function Theorem Motivation. In many problems, objects or quantities of interest can only be described indirectly or implicitly. It is then

More information

Hessian Riemannian Gradient Flows in Convex Programming

Hessian Riemannian Gradient Flows in Convex Programming Hessian Riemannian Gradient Flows in Convex Programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic INTERNATIONAL CONFERENCE ON MODELING AND OPTIMIZATION MODOPT 2004 Universidad de La Frontera, Temuco,

More information

Optimisation in Higher Dimensions

Optimisation in Higher Dimensions CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION   henrion COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS Didier HENRION www.laas.fr/ henrion October 2006 Geometry of LMI sets Given symmetric matrices F i we want to characterize the shape in R n of the LMI set F

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms Agenda Interior Point Methods 1 Barrier functions 2 Analytic center 3 Central path 4 Barrier method 5 Primal-dual path following algorithms 6 Nesterov Todd scaling 7 Complexity analysis Interior point

More information

Functions of Several Variables

Functions of Several Variables Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Functions of Several Variables We now generalize the results from the previous section,

More information

Convex optimization problems. Optimization problem in standard form

Convex optimization problems. Optimization problem in standard form Convex optimization problems optimization problem in standard form convex optimization problems linear optimization quadratic optimization geometric programming quasiconvex optimization generalized inequality

More information

Convex Functions. Pontus Giselsson

Convex Functions. Pontus Giselsson Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum

More information

Lecture: Duality of LP, SOCP and SDP

Lecture: Duality of LP, SOCP and SDP 1/33 Lecture: Duality of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2017.html wenzw@pku.edu.cn Acknowledgement:

More information

An Introduction to Linear Matrix Inequalities. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

An Introduction to Linear Matrix Inequalities. Raktim Bhattacharya Aerospace Engineering, Texas A&M University An Introduction to Linear Matrix Inequalities Raktim Bhattacharya Aerospace Engineering, Texas A&M University Linear Matrix Inequalities What are they? Inequalities involving matrix variables Matrix variables

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

A Brief Review on Convex Optimization

A Brief Review on Convex Optimization A Brief Review on Convex Optimization 1 Convex set S R n is convex if x,y S, λ,µ 0, λ+µ = 1 λx+µy S geometrically: x,y S line segment through x,y S examples (one convex, two nonconvex sets): A Brief Review

More information

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, MATRIX SCALING, AND GORDAN'S THEOREM BAHMAN KALANTARI Abstract. It is a classical inequality that the minimum of

More information

Tilburg University. Multivariate Nonnegative Quadratic Mappings Luo, Z-Q; Sturm, J.F.; Zhang, S. Publication date: Link to publication

Tilburg University. Multivariate Nonnegative Quadratic Mappings Luo, Z-Q; Sturm, J.F.; Zhang, S. Publication date: Link to publication Tilburg University Multivariate Nonnegative Quadratic Mappings Luo, Z-Q; Sturm, J.F.; Zhang, S. Publication date: 2003 Link to publication Citation for published version (APA): Luo, Z-Q., Sturm, J. F.,

More information

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3 MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications

More information

On polyhedral approximations of the second-order cone Aharon Ben-Tal 1 and Arkadi Nemirovski 1 May 30, 1999

On polyhedral approximations of the second-order cone Aharon Ben-Tal 1 and Arkadi Nemirovski 1 May 30, 1999 On polyhedral approximations of the second-order cone Aharon Ben-Tal 1 and Arkadi Nemirovski 1 May 30, 1999 Abstract We demonstrate that a conic quadratic problem min x e T x Ax b, A l x b l 2 c T l x

More information

Newtonian Mechanics. Chapter Classical space-time

Newtonian Mechanics. Chapter Classical space-time Chapter 1 Newtonian Mechanics In these notes classical mechanics will be viewed as a mathematical model for the description of physical systems consisting of a certain (generally finite) number of particles

More information

Recursive Construction of Optimal Self-Concordant Barriers for Homogeneous Cones

Recursive Construction of Optimal Self-Concordant Barriers for Homogeneous Cones Recursive Construction of Optimal Self-Concordant Barriers for Homogeneous Cones OLENA SHEVCHENKO November 20, 2006 Abstract. In this paper, we give a recursive formula for optimal dual barrier functions

More information

EE/AA 578, Univ of Washington, Fall Duality

EE/AA 578, Univ of Washington, Fall Duality 7. Duality EE/AA 578, Univ of Washington, Fall 2016 Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Using Schur Complement Theorem to prove convexity of some SOC-functions

Using Schur Complement Theorem to prove convexity of some SOC-functions Journal of Nonlinear and Convex Analysis, vol. 13, no. 3, pp. 41-431, 01 Using Schur Complement Theorem to prove convexity of some SOC-functions Jein-Shan Chen 1 Department of Mathematics National Taiwan

More information