Bounding Probability of Small Deviation: A Fourth Moment Approach. December 13, 2007

Size: px
Start display at page:

Download "Bounding Probability of Small Deviation: A Fourth Moment Approach. December 13, 2007"

Transcription

1 Bounding Probability of Small Deviation: A Fourth Moment Approach Simai He, Jiawei Zhang, and Shuzhong Zhang December, 007 Abstract In this paper we study the problem of bounding the value of the probability distribution function of a random variable X at E[X] + a where a is a small quantity in comparison with E[X], by means of the second and the fourth moments of X In this particular context, many classical inequalities yield only trivial bounds By studying the primal-dual moments-generating conic optimization problems, we obtain upper bounds for Prob {X E[X] + a}, Prob {X 0}, and Prob {X a} respectively, where we assume the knowledge of the first, second and fourth moments of X These bounds are proved to be tightest possible As application, we demonstrate that the new probability bounds lead to a substantial sharpening and simplification of a recent result and its analysis by Feige ([7], 006); also, they lead to new properties of the distribution of the cut values for the max-cut problem We expect the new probability bounds to be useful in many other applications Keywords: probability of small deviation, fourth moment of a random variable, sum of random variables MSC subject classification: 60E5, 78M05, 60G50 Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, Hong Kong smhe@secuhkeduhk Department of Information, Operations, and Management Sciences, Stern School of Business, New York University, New York, USA jzhang@sternnyuedu Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, Hong Kong zhang@secuhkeduhk Research supported by Hong Kong RGC Earmarked Grants CUHK8505 and CUHK806

2 Introduction For a random variable X R, we consider the problem of upper bounding Prob {X E[X] + a} () for a given real a This problem has been studied extensively in the literature Based on available information about the distribution of X, various inequalities have been developed, including, the well-known Markov inequality and Chebyshev inequality Such inequalities () and () have been extremely useful However, these two inequalities by themselves could sometimes be too weak to yield useful results especially when a is small or zero This motivates us to develop stronger probability inequalities that can handle small deviations Our results In Section, we start our investigation by developing upper bounds for Prob {X 0} that are relatively simple functions of the first, second, and fourth moments of X In particular, we prove that, for any v > 0 Prob {X 0} ( 9 ( ) M + M v v M ) v () Here and throughout the paper, we denote M m = E[X m ] The above result is in Theorem of the current paper The bound provided by () has a relatively simple closed-form expression, and we have the freedom to choose any v > 0 in the bound Therefore, it is quite convenient to use this bound as long as the information about M, M, and M is available The study of this type of probability bounds is motivated by a lemma used in He et al [9], which is a special case of () when M = 0 with a specific choice of v The assumptions on the inequality () is minimal In fact, we do not even require the assumption that E[X] 0 That is to say, we can estimate the probability that X 0 even when E[X] 0 In fact, inequality () is non-trivial ie, the right hand side is less than, as long as E[X] < 0 or E[X] E[X ] E[X ] This is in contrast to many other probability inequalities in the literature, as we shall see in the next subsection The bound provided by () however, is not necessarily tight It is of interest to know whether or not the bound can be further improved We settle this issue by presenting in Theorem 8 a tight upper bound, which is thus the best possible bound, given the moments information As it turns out, the bound () is a very good one, in view of the tight bound; it is even tight under a certain condition After settling the issue of the probability bound for Prob {X 0}, it is natural to consider the bound for Prob {X a}, using the same moments information This extension is useful and is nontrivial to establish When E[X] = 0, we are able to provide a tight bound for Prob {X a}, using the information of M and M The result is presented in Theorem

3 Of course, inequality () may not be immediately applicable if the second and the fourth moments of X is not directly available Fortunately, in many applications, as we shall demonstrate in this paper, it is relatively straightforward to compute or bound the second and the fourth moments In Section and Section, we provide several examples to demonstrate the applicability of Theorem Our first example regards the sum of n independent random variables In particular, given n independent random variables, X, X,, X n, we provide upper bounds on { n [ n ] } Prob X i E X i + a, i =,, m i= i= The bounds are particularly useful when a is a relatively small non-negative real Here the random variables X i could be bounded from both sides, or from below only As a special case of this result, we obtain the following bound If X i is non-negative with expectation, then { n } Prob X i n i= This strengthens the main result of a recent paper by Feige [7] In [7], a weaker upper bound of / is proved by using a completely different approach, and the proof is considerably more involved and lengthy In Section we also apply Theorem to the well-known weighted maximum-cut problem Given an undirected graph G = (V, E) where each edge (u, v) has a weight w uv, we wish to partition the vertices of G into two sets S and S so as to maximize the total weight of the edges (u, v) such that u S and v S A simple solution to this problem is to independently and equiprobably assign each vertex of G to either S or S We denote the total weight of edges with end-points in different sets by W It is clear that the expected value of the W is exactly (u,v) E w u,v By applying Theorem, we can show that Prob W (u,v) E w u,v > 5 and Prob W ( ) V (u,v) E w u,v > % Both bounds seem to be new Furthermore, the second bound implies ( that) for any graph, there exists a cut so that the total weight of edges in the cut is at least V (u,v) E w u,v

4 Related Literature In the literature, there are several probability inequalities based on moment information example, if X assumes only non-negative values, then For Prob {X a} E[X] a () This is the well-known Markov inequality and gives the tightest possible bound when we know only that X is non-negative and has a given expectation If the standard deviation of X, denoted by σ, is also available, and t > 0, then we have Prob {X E[X] + tσ} + t () This inequality is often referred to as the (one-sided) Chebyshev inequality Both inequalities () and () have been extremely useful If we know the first three moments of X, it is shown in Bertsimas and Popescu [] that, Prob {X > ( + δ)e[x]} min ( C M C M +δ, +δ +δ ) DM, if δ > C DM +(C M δ) M, D M +(+δ)(c M δ) D M +(+C M )(C M δ), if δ C M, where CM = M M and D M M = M M M, and this bound is tight Tight bounds on Prob {X < M ( δ)m } and Prob { X M > δm } are also provided in [] These inequalities are potentially useful to bound small deviation probability as well, ie, when δ is small However, we noticed that in several applications that we consider in this paper, it is harder to estimate M than M Furthermore, the bound provided by inequality (5) could be as weak as Markov s and Chebyshev s bounds, for instance, for the problem considered by Feige [7] We shall discuss this in more details later Zelen [7] showed that, if the first four moments of X are known, then (5) Prob {X E[X] + tσ} ( + t + (t tκ ) ) κ κ for t κ + κ +, (6) where κ m = M m σ m Let There are also probability inequalities that uses absolute moments of the random variable X Cantelli [] showed that for m > 0, Prob { X E[X] a} ν m = E [(X E[X])] m ν m ν m ν m ν m + (a m ν m ) for a ( νm ν m ) /m

5 When m =, the above inequality is reduced to the well-known (two-sided) Chebyshev inequality Von Mises [] proved that, for m > k > 0, Prob { X a} J m ν m J m a m for a where J is the root, different from a, of the equation ( νm ν k ) /(m k), (J m a m )/(J k a k ) = (ν m a m )/(ν k a k ) Unfortunately, it is clear from the conditions provided in the above inequalities proved by Zelen [7], Cantelli [], and Von Mises [], that none of them is applicable for bounding probabilities when the deviation is very small In a recent paper, He et al [9] studied SDP relaxations for certain quadratic optimization problems The main results are to establish the gap between the SDP relaxations and the quadratic optimization problems As a key to their main results, they established the following inequality: Prob {X E[X]} 9 0 σ E[(X E[X]) ] This inequality is a special case of Theorem The current paper is partly motivated by [9] Our paper is also related to Berger [], which uses the fourth moment information to bound the absolute value of a random variable More specifically, it is shown in [] that, for all q > 0, E[ X ] ( ) E[X ] E[X ] q q This result has been used by Berger to bound the absolute value of a weighted sum of {+, } unbiased random variables, and achieve tight bounds for the total discrepancy of a set system Our results can be viewed as solutions to a special class of moment problems Moment problems concern about deriving bounds on the probability that a certain random variable belongs in a given set, given information on some of its moments The study on moment problems has a long history; see Bertsimas and Popescu [] for a brief review of this area The tight bounds derived in our paper use an optimization method and duality theory This duality approach was proposed independently and simultaneously by Isii [0] and Karlin [] Bertsimas and Popescu [] show that, for univariate random variables, the dual of the moment problem can be formulated as a semidefinite program (SDP) This result is important because SDP problem can be solved in polynomial time within any prescribed accuracy They also discuss the complexity of solving the dual moment problem for multivariate random variables Recent results on moment problems can also be found in [5], [5], and [] The work by Bertsimas and Popescu [] seems to have settled the moment problems for univariate random variables, ie, for the given information of the moments, one can compute the desired probability bound efficiently by solving an SDP However, such bounds may not be conveniently used because of the lack of simple closed-form expressions 5

6 The Moment Problem: Duality Approach Let us start our discussion by considering the problem: or equivalently ZP = max Prob {X 0} st E[X] = M E[X ] = M E[X ] = M, ZP = max F ( ) x 0 df (x) st x R df (x) = x R x df (x) = M x R x df (x) = M x R x df (x) = M, where the variable of this infinite dimensional optimization problem is the probability measure F ( ) The dual problem of (7) is given as follows: (7) Z D = min y 0 + M y + M y + M y st g(x) := y 0 + y x + y x + y x {x 0}, x R (8) We first define a feasible solution to the dual problem (8) Lemma For any u > v > 0, let c = (u+v) (u v) > 0 and d = v (u+v) > 0 Define y 0 = cu + du, y = du, y = d cu, y = c (9) Then (y 0, y, y, y ) is feasible to problem (8) if u + v Proof If (y 0, y, y, y ) is defined as in (9), then g(x) = cx + (d cu )x + dux + cu + du = c(x u ) + d(x + u) It is clear that g(x) 0 for all x R It is left to verify that g(x) for all x 0 We first observe that g(0) = u + u v u v (u + v) (u v) Thus, g(0) is reduced to v + uv u 0, which is true by the assumption that u + v Therefore, g(0) Notice that, g(x) = (x + u) (c(x u) + d ) 6

7 Since c(x u) + d > 0, we have g(x) = 0 if and only if x = u < 0 Thus x = u < 0 is the only global minimum solution of g(x), and thus one of the local minimum solutions Since g(x) is a polynomial with order four, it has at most two local minimum solutions, including x = u We denote the other local minimum solution by z If z < 0, then we must have that g(x) is increasing for x 0, and thus g(x) g(0) Therefore, we assume that z > 0 > u If follows that z must be the largest root to g (x) = 0 But g (x) = cx(x u ) + d(x + u) and the largest root to g (x) = 0 is u + u d c Therefore, z = u u + d c = u u + v(u v) = u v + u = v The last equality holds since u + v Now it is straightforward to verify that g(z) = g(v) = Finally, we observe that the global minimum solution to g(x) in [0, ) is either x = 0 or x = z Therefore, g(x) g(0) = g(z) = for all x 0 This completes the proof In Lemma, if we choose u = + v, then we have Corollary For any v > 0, define y 0 =, y = 8 9 ( )v, y = ( )v, y = 9 ( )v (0) Then (y 0, y, y, y ) is feasible to problem (8) with an objective value ( 9 ( ) M + M v v M ) v Corollary immediately leads to our first main result Theorem For any v > 0, Prob {X 0} 9 ( ) ( E[X] ) + E[X ] v v E[X ] v () In Theorem, we have the freedom to choose any v > 0 In particular, we could choose v that maximizes the function M + M v v M v Such a v can be obtained by solving the equation M v + M v M = 0, 7

8 if a solution exists Notice that, even if we choose the best v, the bound provided in Theorem is not necessarily tight In what follows, we develop a tight bound for Prob {X 0}, given the first, second, and the fourth moments of X We begin with the case where the bound in Theorem is tight Since when M = M or M = M the distribution X can be easily identified by the first two moments, and the result in Theorem 8 easily follows, therefore we assume M > M and M > M for the remaining part of this section and Let V min = α = M M M M ( )M + 7 > 0 Lemma If M /M M /M and α V min, then And the bound is tight Prob {X 0} 9 ( ) sup v>0 M + M 0 () ( E[X] + E[X ] v v E[X ] v Proof The inequality has been established in Theorem We need only to show that the bound is tight In view of Lemma, it is sufficient to find a feasible solution to problem (7) with an objective value that is equal to the right hand side of the bound If M < 0, define f(x) = M x + M x M Then it can be verified that f(x) is strictly increasing when x 0 By assumption, M /M M /M, and thus, ( f ( ) M ) = (M /M M ) 0 M On the other hand, by (), V min is a solution of the equation x ( )M x ( )M = 0 Thus ) f(v min ) = ( )M Vmin ( 7)M M V min + ( )M M = ( )M Vmin + ( )M (( ) )M Vmin + ( )M M = M M + ( )(D M )V min 0 where the last inequality holds because of the assumption that V min ( + )α By the monotonicity of f(x) when x 0, we must have V min ( ) M M Furthermore, there must exist a unique v [V min, ( ) M M ] such that f(v) = 0 For simplicity, in what follows, we assume v satisfy such a condition Also, let u = + v 8

9 We now define a random variable X = 0, with probability p q; v, with probability p := 6 u, with probability q := 9 ( )( M v + M v M v ); M 6 v 9 M v + M 9 v We show X defines a feasible solution to problem (7) First of all, by the fact f(v) = 0, or M = ( M v + M v ), we have q = ( ( ) M v + M ) v and p = M v + M v Therefore q 0 and p 0 since v ( ) M M Furthermore, p + q = ( ) M v + ( ) M v since v V min Therefore, X is indeed a well-defined random variable It is easy to check that ( E[X] = qu + pv = ( )( M v + M ) ( + v ) M v + v + = M, ( E[X ] = ( )( M v + M ) v ) = M, and ( E[X ] = ( )( M v + M ) v ) = M v + M v = M Therefore, X is feasible to problem (7) Finally, since u v > 0, we have ( + ( ) v + ( + ( ) v + Prob {X 0} = Prob {X u} = q = ( ) 9 This completes the proof of the lemma for the case M < 0 M v + M v + ( M v M v M v ) ) v v M v + M v M ) v For the case M 0 the proof is completely parallel, except that the solution for f(v) = 0 exists in range v [V min, M M ] The details are omitted here In order to get a tight bound for cases that are not covered in Lemma, we need to define different primal and dual variables, which are summarized in the following three lemmas ) v 9

10 Lemma 5 If M /M M /M and α V min, then And the bound is tight Prob {X 0} + α + M M + α + M α Proof Define From the assumption α s = M + α + αm z = α + M u = s+α v = s α V min, we have (5 + )α M α M 0 It follows that s = M + α + αm ( + )α and thus u = s + α + s α = + v, which also implies that u > v > 0 Thus, by Lemma, the function g(x) = c(x u ) + d(x + u) = cx + (d cu )x + dux + cu + du, with c = (u+v) (u v) > 0 and d = v > 0, (implicitly) defines a feasible solution to problem (u+v) (8) The corresponding dual objective value is where we use the fact that On the other hand, we define cm + (d cu )M + dum + cu + du = s + z s, X = { M = s α M α M = (M M )α + M c = s α d = s s α u (< 0), v (> 0), with probability q := s z with probability p := s+z We shall show that X is a feasible solution to problem (7) s ; s It is obvious that p + q = Also, by the fact that M M, we have s = M + α + αm α + M = z 0

11 Therefore, p, q 0 Thus, X is a well-defined random variable Furthermore, E[X] = uq + vp = s + α = M E[X ] = u q + v (s + α) p = s α M s s α M s + s α s + α + M s + (s α) s + α + M s = M, and E[X ] = u q + v (s + α) s α M (s α) s + α + M p = + 6 s 6 s = (M + αm )(M + α + αm ) αm (M + α + αm ) = (M M )α + M = M = s α M α Finally, Prob {X 0} = Prob {X = v} = s + z s, which is equal to the dual objective value This completes the proof of the Lemma Lemma 6 If M /M M /M and M < 0, then And the bound is tight Prob {X 0} M M Proof The inequality is the well-known Chebyshev inequality, which is known to be tight; see, for example, Bertsimas and Popescu [] Lemma 7 If M /M M /M and M > 0, then the trivial bound Prob {X 0} is actually tight Proof The primal solution X with objective value can be constructed this way: For any t M > 0, define random variable X t as a two point distribution as follows: { 0, with probability M X t = t t, with probability M t We have that EX t = M, EX t = tm and EX t = t M Consider function f(x) = x /M, which is convex when x > 0 Notice that M M, f(m ) = M and f(m ) M, the line passing through (M, M ) and (M, M ) intersect with function f(x) at some t M Thus there exists a p [0, ], such that p(m, M ) + ( p)(t, f(t)) = (M, M )

12 Let Y be the Bernoulli trial which takes the value with probability p, and let X = Y X M + ( Y )X t/m, where Y is independent to X t/m and X M, then (EX, EX, EX ) = p(ex M, EX M, EX M ) + ( p)(ex t/m, EX t/m, EX t/m ) = p(m, M, M ) + ( p)(m, t, f(t)) = (M, M, M ) Because X 0, this gives a feasible solution of the primal problem 7 with objective value Since is an upper bound for ZP, we conclude that Z P = and that X is an optimal primal solution For the dual problem (8), y 0 = y = y = y = 0 is obviously a feasible solution with objective value Because ZD Z P =, this is an optimal dual solution Lemma 6 and Lemma 7 indicate that when the fourth moment of X, ie, M, becomes sufficiently large, then the information will not be useful anymore in bounding the probability that X 0 The following theorem summarizes the results we obtained above Theorem 8 Prob {X 0} M M, if M M, if M M ( ) ( 9 sup v>0 M v + M M ) v v, if M M α+m, if M M +α +M α M M M M and M < 0; and M > 0; () < M and α M V min ; + < M M and α M V min, M where α M, V M M min ( )M + 7 M + M Furthermore, the bound is tight, ie, there exists an X such that the inequality () holds as an equality Now we consider a special case where E[X] = 0 As we have mentioned in the introduction, He et al [9] established the following inequality Prob {X E[X]} 9 0 σ E[(X E[X]) ], which has been a key to study an SDP relaxation for certain class of quadratic optimization problems Here we show that this inequlality can be strengthened by using Theorem Corollary 9 If E[X] = 0, then sup {Prob (X 0)} = X ( ) M M, if M M +, if M +M /M M ;

13 Proof Notice that if M = E[X] = 0, then V min = ( )M and α = M M M condition M M Therefore, The is equivalent to α V min The corollary follows by noting that ( M max v>0 v M ) v = 9 M M By applying Corollary 9, we can obtain a non-trivial bound for the probability X a when E[X] = 0, given the information on M and M Corollary 0 If E[X] = 0 and a 0, then Prob {X a} ( (M + a ) ) M + 6a M + a Proof Let Y be a random random variable independent to X, and X takes only one of the two values, a or a, each with probability half Let Z = X + Y Then Then, by Corollary 9, However, E[Z] = 0 E[Z ] = E[X ] + E[Y ] = M + a E[Z ] = E[X ] + 6E[X ]a + a = M + 6a M + a Prob {Z 0} = The desired inequality follows Prob {Z 0} ( (M + a ) ) M + 6a M + a Prob {X a} + Prob {X a} Prob {X a} The bound proved in Corollary 0 is not tight in general A tight bound is summarized in the following theorem Its proof, which is quite technical and similar to the proof of Theorem 8, is provided in the appendix Theorem Let K = M /M and L = M /a If E[X] = 0 and a 0, Then M M, if K L + +a L ; Prob {X a} M M M M, if K L + a +a L and L < ; + +M /M, if K L + L, L and L K + K + K K+ K ; min{p (v) v a}, otherwise

14 where P (v) = M + M (v + av + a ) + a v + av a + a v + a v + 6av + 9 v + (v + av + a v v) + (a+v) And the bounds are tight Small Deviation Bound for Sum of Independent Random Variables In this section, we consider the problem of bounding the probability of small deviations for sum of independent random variables In particular, consider n independent random variables X, X,, X n each with a mean of zero Let S = n i= X i We are interested in the probability that S < for some given constant For this purpose, we may directly apply Theorem Then we need to estimate E[S ] and E[S ] We may also apply Theorem 8 In this case, we need to estimate E[(S ) ] and E[(S ) ] We demonstrate below how this could be done We consider two cases In the first case, the random variables X i are uniformly bounded from both sides In the second case, we assume that the random variables are uniformly bounded only from below Given two nonnegative constants c and c, define s(c, c ) := max{c + c, c c, c + c c c (c c )} Our first result is summarized below Theorem Consider n independent random variables X, X,, X n Assume that > 0 is a given constant Also assume that E[X i ] = 0 and there exists two nonnegative constants c and c such that c X i c Let S = n i= X i, then where F (, c, c ) = ( ) 9 inf D>0 Prob {S < } F (, c, c ) F (c, c ), () ( 6(D + 9 ) D + (6 + s(c, c ))D + + (D + ) ) D + (6 + s(c, c ))D + and F (c, c ) = ( s(c, c ) + ) s(c, c ) + s(c, c ) +

15 Proof First of all, we can assume without loss of generality that X i follows a two point distribution for every i =,,, n In particular, given that E[X i ] = 0, we assume that there exists a i, b i 0, such that { b a i, with probability i X i = b i, with probability a i +b i a i a i +b i It follows that E[Xi ] = a ib i and E[Xi ] = a ib i (a i a ib i + b i ) Let denote the variance of S by D, ie, D = n i= E[X i ] = n i= a ib i Therefore, E[(S ) ] = D + Furthermore, E[(S ) ] = E[S ] E[S ] + 6 E[S ] + n = E[Xi ]E[Xj ] = n E[Xi ] + 6 E[Xi ] + 6 i= i<j i= i= n ( E[X i ] E[Xi ] (E[Xi ]) ) + D + 6 D + M i= = D + 6 D + + n E[Xi ] + n a i b i (a i + b i a i b i (b i a i )) i= Notice that a i + b i a ib i (b i a i ) is a convex function of a i when b i is fixed, and is convex in b i when a i is fixed Therefore, an optimal solution to the optimization problem max 0 a i c,0 b i c (a i + b i a i b i (b i a i )) is in the set {(0, 0), (0, c ), (c, 0), (c, c )} Thus, we conclude that a i + b i a i b i (b i a i ) s(c, c ) It then follows that E[(S ) ] D + 6 D + + s(c, c )D Thus by Theorem 8, we have for any v > 0 that Prob {S < 0} ( 9 ( ) v + (D + ) v D + 6 D + + s(c, c )D ) v In particular, we choose v such that Then we must have v = ( D + ) D + (6 + s(c, c ))D + Prob {S < 0} ( 9 ( 6(D ) + 9 ) D + (6 + s(c, c ))D + + (D + ) ) D + (6 + s(c, c ))D + F (, c, c ) 5

16 Furthermore, it is clear that F (, c, c ) ( inf ( (D + ) ) ) D>0 D + (6 + s(c, c ))D + ( (s(c, c ) + ) ) s(c, c ) + s(c, c ) + = F (c, c ), where the last inequality uses the fact that if we let t = D + (6 + x)d +, then D+ (D + ) = + xt( t) t + for any x > This completes the proof of the theorem x (x + ), Now we consider the case where the random variables X i are bounded from below only We obtain a similar result as Theorem Theorem Consider n independent random variables X, X,, X n Assume that > 0 is a given constant Also assume that E[X i ] = 0 and there exists a constant c > 0 such that X i c for every i Let S = n i= X i, then for any τ > 0, Prob {X < } e /τ F (, c, τ max(, c)) e /τ F (c, τ max(, c)) (5) Proof Once again, we assume without loss of generality that there exist a i, b i 0, such that { b a i, with probability i X i = a i +b i b i, with probability a i a i +b i By assumption a i c We also assume that without loss of generality that b b b n We consider a fixed τ > 0 and define N = max{0, max{k b k τ(a + a + + a k ), k n}} Let a = N i= a i; if N = 0, then let a = 0 If N < n, then for every i > N, b i b N+ τ N+ i= a i τ(a + a N+ ) τ(a + c ) For any i N, b i b N τa Thus, if N > 0, then { N } N Prob X i = a = Prob {X i = a i } i= = i= N i= ( a ) i a i + b i N e ai/(τa) = e /τ i= 6 N i= ( a ) i a i + τa

17 Let Y = n i=n+ X i Because for each i > N, a i c c(a + ) and b i max(, c)τ(a + ), by Theorem, we know that The proof is completed Prob {S < } { } Prob X i = a Prob {Y < a + } i<n e /τ F (, c, τ max{, c}) e /τ F (c, τ max{, c}) Theorem generalizes an inequality that was proved by Feige [7] In particular, if every X i is non-negative with expectation, then Feige proved that { n } Prob X i n + i= For this special case, Theorem implies a stronger result than the above inequality Corollary Consider n independent random variables X, X,, X n each with mean zero If X i for all i =,,, n, then we have that { n } Prob X i < e /5 ( ) 8 i= Proof We can apply Theorem with c = and = We choose τ = 5 and thus s(c, τ) = 5 In this case, ( ) 6(D + ) 9 F (, c, τ) = inf D>0 9 ( ) D + D + + (D + ) D + D + However, 6(D + ) 9 D + D + + (D + ) D + D + is a decreasing function of D when D 0 By letting D go to infinity, we have that By Theorem, we have F (, c, c ) 9 ( ) = Prob {X < } e /c F (, c, c ) = e /5 ( ), which completes the proof for the corollary 7

18 It would be interesting to see how strong a bound can be obtained if we apply Markov, Chebyshev, and Bertsimas-Popescu s (three-moments) inequality to the problem considered in Corollary Consider the following example, where all the X i s are iid distribution which take value 0 and with probability / of each Then for the random variable X = n i= X i, and δ = n, we have M = n, M = n + n, and M = n + n Since CM = n = δ, when n, the value f (CM, D M, δ) = n n+ approaches Therefore the three moments inequality alone is not good enough to yield a good bound for the problem Applications In many applications and rounding algorithms, the Chernoff type bounds or other similar inequalities can be applied to yield claims of the following spirit: If n > N(δ, ɛ), then for n independent samples X i ( i n) of a random variable X it follows that Prob {max i n X i ( δ)ex} ɛ However it follows from our analysis, the δ can be dropped and we can claim the following: Lemma If a random variable X has kurtosis κ = ) log( ɛ ) many samples, by Theorem, { Prob max X i EX i n E(X EX), then with n = + (E(X EX) ) (κ + } ɛ and { } Prob min X i EX ɛ i n This is to say that, when a distribution s Kurtosis κ can be estimated or upper bounded, then Θ(κ log(/ɛ)) many samples would guarantee that we are able to draw one whose value is at least as good as the expected value of the distribution, with high probability ɛ Proof Because for each i, Prob {X i EX} ( ) κ+, we have { } ( Prob max X i EX ( ) n ( ) exp n( ) i n κ + κ + Thus if n + (κ + ) log( ɛ ), we have Prob {max i n X i EX} ɛ The other inequality is symmetric Also, for sums of independent random variables we have the following: Lemma If X i are independent random variables with EX i = 0, EXi = D, EXi (κ + )(EXi ), then we have that { n } + κ Prob X i 0 n + κ n i= ) 8

19 Proof Let X = n i= X i, D X = Var(X) = nvar(x i ) = nd, τ = (κ + )D Then τ X = EX = n i= EX i + 6 i<j EX i EX j nτ + D X nd = (n n + n(κ + ))D Thus The other inequality follows by symmetry Prob {X EX} ( ) D X τ X + κ n Now we consider the weighted maximum cut problem In this problem, we are given an undirected graph G = (V, E) where each edge (u, v) has a weight w u,v, and the goal is to partition the vertices of G into two sets S and S so as to maximize the total weight of the edges (u, v) such that u S and v S This problem is NP-hard, but admits a polynomial time 0878-approximation algorithm; see Goemans and Williamson [6] Prior to the celebrated result of Goemans and Williamson, the best known approximation ratio for the maximum cut problem was / for the weighted version, and + δ for the unweighted version, where δ denotes the maximum degree of a vertex It is well-known that, a simple /-approximation algorithm can be obtained by independently and equiprobably assigning each vertex of G to either S or S Indeed, if we denote the total weight of edges with end-points in different sets by W, then it is clear that E[W ] = (u,v) E which of course is no less than half of the maximum weight w u,v := W tot (6) Equation (6) has a stronger implication That is, for any graph, there exists a cut so that weight of the cut is at least half of the total weight of the edges of the graph However, two interesting questions remain: There are O( V ) many cuts for a graph Among all the possible cuts, how many of them have a weight larger than W tot /? Is it possible to show that there always exists a cut with a weight higher than αw tot for some α > /? When the graph is unweighted, the answer to the second question is yes with an α = + n and this bound is the best possible; see Haglin and Venkatesan [8] The result is obtained by proving the existence of a matching of certain size, which also gives a linear time algorithm to find a cut with a weight larger than ( + n )W tot Now we shall answer the above two questions for a general weighted graph by using the simple randomized algorithm described earlier, together with the moment bound developed in this paper 9

20 We slightly formalizes the randomized algorithm as follows We define V independent random binary variables X,, X V, so that for each node i V, X i takes value or with probability half Thus, X i = indicate node i is assigned to the set S, and vice versa Then we have For convenience, we also define W = i<j X i X j w i,j Y = W W tot so that E[Y ] = 0 We now estimate the second and the fourth moments of random variable Y It can also be shown that E[Y ] = E w i,j X i X j = wi,j V W tot (7) i<j i<j E[Y ] 5(E[Y ]) (8) Therefore, it follows immediately from Corollary 9 that { Prob W } W tot = Prob { Y 0} ( ) (E[Y ]) E[Y ] 5 Denote = ( E[Y ]/ 5 ) / > E[Y ]/ and let Z = t Y with t 0 Then we have E[Z] = t, E[Z ] = E[Y ] + t, and E[Z ] = E[Y ] t E[Y ] + 6t E[Y ] + t E[Y ] + t (E[Y ]) / (E[Y ]) / + 6t E[Y ] + t E[Y ] + 5t (E[Y ]) / + 6t E[Y ] + t 5E[Y ] + 5t (E[Y ]) / + 6t E[Y ] + t ( = 5 + (5) / t + 6 t + ) 5 5 t (E[Y ]) Thus, by Theorem 8, we have, for any v > 0, Prob {Y t } = Prob {Z 0} 9 ( ) In particular, if we choose v = 0 and t = 00, then Prob {Y t } > % ( E[Z] ) + E[Z ] v v E[Z ] v It follows that Prob { ( W ) } W tot > % V To summarize, we have proved the following: 0

21 Theorem For any weighted graph, the following two statements are true ) Among all possible cuts of the graph, at least 5 > % of them will have a cut value larger than half of the total weight of the edges of the graph ( ) ) There exists a cut whose weight is at least V times the total weight of the edges of the graph References [] A Ben-Tal and A Nemirovski Lectures on Modern Convex Optimization: Analysis, Algorithm, and Engineering Applications MPS/SIAM Ser Optim, SIAM, Philadelphia, 00 [] B Berger The Fourth Moment Method SIAM Journal on Computing 6, pp 88 07, 999 [] FP Cantelli Intorno ad un teorema fundamentale della teoria del rischio Boll Assoc Attuar Ital (Milan), pp, 90 [] D Bertsimas and I Popescu Optimal Inequality in Probability Theory: A Convex Optimization Approach SIAM Journal on Optimzation, 005 [5] D Bertsimas and I Popescu On the Relation Between Option and Stock prices: A Convex Optimization Approach Operation Research 50 No, pp 58 7, 00 [6] MX Goemans and DP Williamson Improved Approximation Algorithms for Maximum Cut and Satisfiability Problems using Semidefinite Programming Journal of the ACM, pp 5 5, 995 [7] U Feige On Sums of Independent Random Variables with Unbounded Variances, and Estimating the Average Degree in a Graph SICOMP, (006) [8] DJ Haglin and SM Venkatesan Approximation and Intractability Results for the Maximum Cut Problem and Its Variants IEEE Transactions on Computers 0, pp 0, 99 [9] S He, ZQ Luo, J Nie, and S Zhang Semidefnite Relaxation Bounds for Indefinite Homogeneous Quadratic Optimization Working Paper, 007 [0] K Isii The extrema of probaiblity determined by generalized moments I Bounded random variables Ann Insti Statist Math,, pp 6 68, 960 [] S Karlin and WJ Studden Tchebysheff Systems: With Applications in Analysis and Statistics Pure Appl Math 5, Interscience, John Wiley and Sons, New York, 966 [] JB Lasserre Bounds on measures satisfying moment conditions Ann Appl Probab, pp 7, 00

22 [] R Von Mises The Limits of a Distribution Function if Two Expected Values Are Given Ann Math Statist 0, pp 99 0, 99 [] Yu Nesterov Structure of Non-Negative Polynomial and Optimization Problems Preprint DP 979, Louvain-la-Neuve, Belgium, 997 [5] I Popescu A Semidefinite Programming Approach to Optimal Moment Bounds for Convex Classes of Distributions Mathematics of Operation Research, 005 [6] J Smith Generalized Chebyshev Inequalities: Theory and Applications in Decision Analysis Operations Research,, , 995 [7] M Zelen Bounds on a Distribution Function That Are Functions of Moments to Order Four J Res Nat Bur Stand, 5, pp 77 8, 95 A Proof of Theorem For Theorem, the primal problem is ZP = max Prob {X a} st E[X] = 0 E[X ] = M E[X ] = M, or equivalently ZP = max F ( ) x a df (x) st x R df (x) = 0 x R x df (x) = M x R x df (x)µ = M x R x df (x) = M (A9) Its dual problem in this setting can be written as Z D = min y y + M y + M y st g(x) := y 0 + y x + y x + y x {x a}, x R (A0) Lemma A For any u > v > a, let c = (u+v) (u v) > 0 and d = v (u+v) > 0 Define y 0 = cu + du, y = du, y = d cu, y = c Then (y 0, y, y, y ) is feasible to problem (A0) if u v + v + (a+v) (A)

23 Proof If (y 0, y, y, y ) is defined by (A), then g(x) = cx + (d cu )x + dux + cu + du = c(x u ) + d(x + u) It is clear that g(x) 0 for all x R It is left to verify that g(x) for all x a The condition u v + v + (a+v) implies that We first observe that u uv (v + a) g(a) c = (a u ) + (a + u) (uv v ) (u + v) (u v) = a a (u uv + v ) + av(u uv) + v (v + uv u ) = (a v ) (u uv)(v a) (a v ) (v + a) (v a) = 0 Therefore g(a) Notice that, g(x) = (x + u) (c(x u) + d ) Since c(x u) + d > 0, we have g(x) = 0 if and only if x = u < 0 Thus x = u < 0 is the only global minimum solution of g(x), and thus one of the local minimum solutions Since g(x) is a polynomial with order four, it has at most two local minimum solutions, including x = u We denote the other local minimum solution by z If z < a, then we must have that g(x) is increasing for x a, and thus g(x) g(a) Therefore, we assume that z > 0 > u, and z must be the largest root to g (x) = 0 But g (x) = cx(x u ) + d(x + u) and the largest root to g (x) = 0 is u + u d c Therefore, The last equality holds since z = u u + d c = u u + v(u v) = u v + u = v u v + v + (a + v) Now it is straightforward to verify that g(z) = g(v) = v Finally, we observe that the global minimum solution to g(x) in [0, ) is either x = 0 or x = z Therefore, g(x) g(0) = g(z) = for all x 0 This completes the proof

24 Proof of Theorem When M = M, the distribution has to be { M, with probability X = ; M, with probability, Since Prob {X a} = Theorem holds under this condition { 0, if L < ;, if L, If M > M, the given condition of the moment information implies that the strong duality holds Thus problem (A9) is equivalent to ZP = max F ( ) x a df (x) st x R df (x) = 0 x R x df (x) = M x R x df (x) = M x R x df (x) M (A) Case : When K L + L, we define { M X = a, with probability a, with probability L which is always well defined Furthermore, E[X] = M a L+ + a L L+ = al L+ + a L L+ = 0, L+ ; L+, E[X ] = M a L+ + a L L+ = a L L+ + a L L+ = a L = M, E[X ] = a L L+ + a L L+ = a L(L L + ) = M L (L L + ) M K = M Therefore X is feasible to problem (A) with objective value Prob {X a} = L/(L + ) We now define ( ) ax + M g(x) = a, + M which is feasible to problem (A0) The corresponding dual objective value is ( a a + M ) ( ) M M + a = + M L (L + ) + L (L + ) = L L + Therefore, when K L + L the inequality in Theorem holds and is tight Case : When L < and K L + L, define [ a ] M g(x) = a a (x a ) +, M + M

25 which is a feasible solution of problem (A0) The corresponding dual objective value is = = = ( a ) M a a M + (a M )(M a M ) (M a M ) M + M (a a M + M ) M + (a a M + M ) ( a ) ( M a a (M M a M M a M ) + M + M a a M + M + M a a M + M M M (a a M + M ) ( (M M + (a M ) ) M M a a M + M Now we define v = a M M a M, p = M M M M a +a and v, with probability q = p(+a/v) ; X = v, with probability r = p( a/v) ; a, with probability p ) Since a M M = L M KM ( L)M > 0, the value v is well defined We observe that p = M M M M a + a = M M M M + (M a ), thus 0 < p < The condition L < and K L + L implies that which is equivalent to a (M M ) + (a M ) M ( ) (K ) M 5 ( L) + K L L ( ) ( L) M 5 ( L) L + L (L L + ) = M 5 ( L) L = a M (a M ), a v = a (a M ) a M M ( (a M ) ) = M M ( ) p Since a v + a v p, 5

26 we have q, r 0 Notice that p + q + r =, the distribution X is well defined Furthermore, E[X] = (q r)v + pa = pa + pa = 0, E[X ] = (q + r)v + pa = ( p)v + pa = (a M )(a M M ) + a (M M ) a a M + M = M, E[X ] = (q + r)v + pa = (a M M ) + a (M M ) a a M + M = M Thus X is feasible to problem (A9) Since v = the corresponding dual objective value is define a a + M a M a M < a, Prob {X a} = p = M M M M a + a Therefore, the inequality in Theorem is tight when L < and K L + L Case : When L, K L + L and L K + K + K u = v = M M +M M + M M M M +M M M M = M K++ K = M K+ K p = + +K = K++ K K+ q = +K = K+ K K+ K+ K, Because M M for any distribution, these values are well defined It follows from definition that u > v > 0 From the assumption K L + L and L, we have v = = K + K M M K + + K M L+ L + L L M = L = a The assumption L K + K + K K+ K implies that ( K + K + L ) K + K + K = ( K + + K ) K 6

27 Therefore u(u v) (a + v), which implies that u v + v + (a+v) It follows from Lemma A that the function with and g(x) = c(x u ) + d(x + u) = cx + (d cu )x + dux + cu + du c = (u + v) (u v) = > 0 (K + )M (K + )(K ) d = v (u + v) = K + K M (K + ) K + > 0 defines a feasible solution to problem (A0) Denote t = K + K, then d = cm (t K + ) and u = K++t The corresponding dual objective value is cm + (d cu )M + cu + du = cm (K (K + + t) + t K + (K + + t) + ( = cm t + K + ) t K + K = cm t t + K + = + t (K + ) = p + K + + t ) (t K + ) Now we define X = { u, with probability q; v, with probability p, which is always feasible since K > 0 Furthermore, E[X] = pv qu = 0, E[X ] = pv + qu = pv(u + v) = M, E[X ] = pv + qu = pv(u + v)(u uv + v ) = M K = M Therefore, the inequality in Theorem is tight when L, K L + L and K + K + K K + K L Case : Now we consider the case when L, K L + L and < K + K + K K + K L 7

28 Our main goal is to prove there exists a ˆv a and the corresponding û, so they can generate a feasible dual solution, and also satisfies the following conditions (which are the crucial conditions for the feasibility of the primal solution, and for the primal objective to match the dual objective value): aû M ˆvû; M + M (ˆv + aˆv + a ) + a ˆv + aˆv = (û + ˆv) (û ˆv)(M + aˆv) a + û Define W (K) = S(K) = V (K) = K + K + K K + K + K ; K+ K ; K++ K U(K) = ; u(v) = v + v v t(v) = + (a+v) + (a+v) ; K+ K ; Because K, there exists a unique b such that K = b +, and b K + K + K = (b ) K + + K b, which is a monotonically increasing function of b Since L+ L K = b + L, and L, b, we can conclude that L b If b, then KL b (b + b ) If b <, then KL b + b ( (b ) b ) Therefore, the assumptions L, K L + L and L < W (K) guarantees that KL, and L Also, since b is monotonically increasing to K, we have that W (K) is a monotonically increasing function of K Since W (K) is a continuous function with W () =, there exists a K 0 < K such that W (K 0 ) = L From Lemma A, for any v a > 0, by abusing the notation, let u = u(v) and t = t(v), the function g v (x) = c(x u ) + d(x + u) with d = v and c = is feasible for (u+v) (u+v) (u v) problem (A0) Notice that d = cv(u v) and u uv = (a+v), the corresponding dual objective value is: P (v) = M + M (v(u v) u ) + u + v(u v)u (u + v) (u v) = M + M (v + av + a ) + a v + av (u + v) (u v) 8

29 Let and then we have and f(v) = M + M (v + av + a ) + a v + av h(v) = (u + v) (u v), f (v) = (v + a)(m + av), h (v) = (u + v) (u ) + (u + v) (u v)(u + ) Therefore, = (u + v) ((u v)u + u v) v = (u + v) (a + v) + + = (u + v) (a + t(v)) a + v v + (a+v) + v + v + (a + v) v P (a) = h (a)f(a) h(a)f (a) h (a) = 7a ( M + 6a M + a ) 7a 8a(M + a ) h (a) = 7a (a M M ) h (a) = 7a5 M ( KL) h (a) 0 Also for all v, (u + v)(a + t) (v + a)(a + u) = ( v + t)(a + t) (v + a)(a + v + t) = t a va v = 0 Let { Then because we have Since u 0 v 0, we have that v 0 = V (K 0 ) M ; u 0 = U(K 0 ) M a = L M = W (K 0 ) M, K0 + + K 0 u 0 (u 0 v 0 ) = M K0 u 0 = v 0 v (a + v 0) 9 = u(v 0 ) = (a + v 0)

30 Notice that a + v 0 = S(K 0 ) M and a + t(v 0 ) = a + u 0 v 0 = (S(K 0 ) + K 0 ) M, we have (f(v 0 ) + M K 0 M )h (v 0 ) h(v 0 )f (v 0 ) = (u 0 + v 0 ) M 5/ [(S(K 0 ) + K 0 ) ( K 0 + V (K 0 ) + S(K 0 ) + V (K 0 ) (S(K 0 ) V (K 0 ) ) ) (U(K 0 ) V (K 0 ) )(V (K 0 ) + S(K 0 ))V (K 0 )(U(K 0 ) + W (K 0 )) ] [ = (u 0 + v 0 ) M 5/ (S(K 0 ) + K 0 ) K0 + K 0 K 0 + V (K 0 ) K0 + K 0 V (K 0 )(S(K 0 ) + K 0 )(S(K 0 ) + K 0 + ] K 0 ) [ = (u 0 + v 0 ) M 5/ V (K 0 ) K0 + K 0 (S(K 0 ) + K 0 ) K 0 + (S(K 0 ) + K 0 )(S(K 0 ) + K 0 + ] K 0 ) = 0 Therefore, P (v 0 ) = h (v 0 )f(v 0 ) h(v 0 )f (v 0 ) h (v 0 ) = (K 0 K) h (v 0 )M h (v 0 ) < 0 Define v = ( L + L L )a, and the corresponding u = u(v ) = La Let L + L = r, we have that f(v ) + M (L + L )M = a [ (L L + L) + L ( L + L + 6(L + )r + r (L + ) + ) + ( L + L + (L + )r ) (r L ) ] = a ( (6L + 0L + )r (9L + L + L + ) ) Therefore, (f(v ) + M (L + L )M )h (v ) h(v )f (v ) = (u + v ) a 5 [( (6L + 0L + )r (9L + L + L + ) ) (L + r) (r )(L + r)( + r L )(L + r L )] = (u + v ) a 5 [ (7L + 6L + 5L + 9)r (5L + L + L + 7L + ) (7L + 6L + 5L + 9)r + (5L + L + L + 7L + ) ] = 0, which implies that P (v ) = h (v )f(v ) h(v )f (v ) h (v ) = (L + L L KL )a h (v ) h (v ) 0 0

31 Since L, K 0 and K 0 L, we have that v LV (max(, /L)a LV (K 0 )a = v 0 Since L, we have v a Because P (v) is a continuous function, there exists a ˆv [max(a, v 0 ), v ] such that P (ˆv) = 0 Let û = u(ˆv) and ˆt = t(ˆv) Because u(v) is monotonically increasing function of v, we have that Since Since f (ˆv)h(ˆv) = h (ˆv)f(ˆv), it follows that f(ˆv) M + aˆv aû au = M = u 0 v 0 ûˆv f(ˆv) h(ˆv) = (ˆv + a) f = (ˆv + a) (ˆv) h (ˆv) = (û ˆv )(ˆv + a) a + ˆt (û + ˆv)(a + ˆt) = (ˆv + a)(a + û), we have f(ˆv) M + aˆv = (û ˆv )(ˆv + a) a + ˆt Therefore the corresponding dual objective is ˆP = P (ˆv) = = (û ˆv )(û + ˆv) a + û M + aˆv (a + û)(û + ˆv) By the definition of t and u, it s straightforward to prove that Therefore, (v + au a )(a + u) = ( v + 5av + 5 a v) + (a + av + v )t = (u + v) (u v) M = f(ˆv) + (aˆv + a ˆv + M (ˆv + aˆv + a ) = aˆv + a ˆv + M (ˆv + aˆv + a ) (M + aˆv) (û + ˆv) (û ˆv) a + û = aˆv + a ˆv + M (ˆv + aˆv + a ) (M + aˆv) v + au a = M (ˆv + a (ˆv + a) + aû + aˆv) + aˆv(û ûˆv aû) = M (ˆv + a + û ûˆv aû + aˆv) + aˆv(û ûˆv aû) = M (ˆv + a ) a ˆv + (M + aˆv)(û a)(û ˆv) = M (ˆv + a ) a ˆv + ( ˆP )(û a )(û ˆv ) = (M ( ˆP )û )(ˆv + a ) ˆP a ˆv + ( ˆP )û Now we define distribution X as X = ˆv, with probability q = M ˆP a ( ˆP )û ˆv a ; û, with probability p = ˆP ; a, with probability r = ˆP ˆv +( ˆP )û M ˆv a

32 Because it follows that Therefore, q, r 0 Since aû M ûˆv, û M û a ˆP û M û ˆv f(ˆv) = (û ˆv )(û + ˆv)(M + av) a + û 0, and u > v, we have that ˆP, therefore p 0 and the distribution X is well defined Furthermore, E[X] = qˆv pû + ra = M a ˆP ˆv ( ˆP )û ( ˆv + a ˆP )û = 0, E[X ] = qˆv + pû + râ = M, E[X ] = qˆv + pû + ra = (M ( ˆP )û )(ˆv + a ) ˆP a ˆv + ( ˆP )û = M Therefore, X is feasible to problem (A9) Finally, since ˆv a, the primal objective value is Prob {X a} = q + r = ˆP, which matches the dual objective value for dual feasible solution corresponds to ˆv Therefore, we have that Z P = Z D = ˆP By Lemma A, any v a corresponds to a dual feasible solution with objective value P (v), therefore P (v) ˆP, and ˆP = P (ˆv) = min{p (v) v a}

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3 MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications

More information

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization Agenda Applications of semidefinite programming 1 Control and system theory 2 Combinatorial and nonconvex optimization 3 Spectral estimation & super-resolution Control and system theory SDP in wide use

More information

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad Quadratic Maximization and Semidenite Relaxation Shuzhong Zhang Econometric Institute Erasmus University P.O. Box 1738 3000 DR Rotterdam The Netherlands email: zhang@few.eur.nl fax: +31-10-408916 August,

More information

6.854J / J Advanced Algorithms Fall 2008

6.854J / J Advanced Algorithms Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.85J / 8.5J Advanced Algorithms Fall 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 8.5/6.85 Advanced Algorithms

More information

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs LP-Duality ( Approximation Algorithms by V. Vazirani, Chapter 12) - Well-characterized problems, min-max relations, approximate certificates - LP problems in the standard form, primal and dual linear programs

More information

Lecture 8: The Goemans-Williamson MAXCUT algorithm

Lecture 8: The Goemans-Williamson MAXCUT algorithm IU Summer School Lecture 8: The Goemans-Williamson MAXCUT algorithm Lecturer: Igor Gorodezky The Goemans-Williamson algorithm is an approximation algorithm for MAX-CUT based on semidefinite programming.

More information

Nonnegative k-sums, fractional covers, and probability of small deviations

Nonnegative k-sums, fractional covers, and probability of small deviations Nonnegative k-sums, fractional covers, and probability of small deviations Noga Alon Hao Huang Benny Sudakov Abstract More than twenty years ago, Manickam, Miklós, and Singhi conjectured that for any integers

More information

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS Sercan Yıldız syildiz@samsi.info in collaboration with Dávid Papp (NCSU) OPT Transition Workshop May 02, 2017 OUTLINE Polynomial optimization and

More information

A Semidefinite Relaxation Scheme for Multivariate Quartic Polynomial Optimization With Quadratic Constraints

A Semidefinite Relaxation Scheme for Multivariate Quartic Polynomial Optimization With Quadratic Constraints A Semidefinite Relaxation Scheme for Multivariate Quartic Polynomial Optimization With Quadratic Constraints Zhi-Quan Luo and Shuzhong Zhang April 3, 009 Abstract We present a general semidefinite relaxation

More information

Handout 6: Some Applications of Conic Linear Programming

Handout 6: Some Applications of Conic Linear Programming ENGG 550: Foundations of Optimization 08 9 First Term Handout 6: Some Applications of Conic Linear Programming Instructor: Anthony Man Cho So November, 08 Introduction Conic linear programming CLP, and

More information

A Note on KKT Points of Homogeneous Programs 1

A Note on KKT Points of Homogeneous Programs 1 A Note on KKT Points of Homogeneous Programs 1 Y. B. Zhao 2 and D. Li 3 Abstract. Homogeneous programming is an important class of optimization problems. The purpose of this note is to give a truly equivalent

More information

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 1 Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 2 A system of linear equations may not have a solution. It is well known that either Ax = c has a solution, or A T y = 0, c

More information

BBM402-Lecture 20: LP Duality

BBM402-Lecture 20: LP Duality BBM402-Lecture 20: LP Duality Lecturer: Lale Özkahya Resources for the presentation: https://courses.engr.illinois.edu/cs473/fa2016/lectures.html An easy LP? which is compact form for max cx subject to

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization

More information

Convex Optimization 1

Convex Optimization 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski Convex Optimization 1 Many optimization objectives generated

More information

Complex Quadratic Optimization and Semidefinite Programming

Complex Quadratic Optimization and Semidefinite Programming Complex Quadratic Optimization and Semidefinite Programming Shuzhong Zhang Yongwei Huang August 4; revised April 5 Abstract In this paper we study the approximation algorithms for a class of discrete quadratic

More information

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones Research Reports on Mathematical and Computing Sciences Series B : Operations Research Department of Mathematical and Computing Sciences Tokyo Institute of Technology 2-12-1 Oh-Okayama, Meguro-ku, Tokyo

More information

A Unified Theorem on SDP Rank Reduction. yyye

A Unified Theorem on SDP Rank Reduction.   yyye SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford

More information

5. Duality. Lagrangian

5. Duality. Lagrangian 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Lecture 5. The Dual Cone and Dual Problem

Lecture 5. The Dual Cone and Dual Problem IE 8534 1 Lecture 5. The Dual Cone and Dual Problem IE 8534 2 For a convex cone K, its dual cone is defined as K = {y x, y 0, x K}. The inner-product can be replaced by x T y if the coordinates of the

More information

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given. HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard

More information

An 0.5-Approximation Algorithm for MAX DICUT with Given Sizes of Parts

An 0.5-Approximation Algorithm for MAX DICUT with Given Sizes of Parts An 0.5-Approximation Algorithm for MAX DICUT with Given Sizes of Parts Alexander Ageev Refael Hassin Maxim Sviridenko Abstract Given a directed graph G and an edge weight function w : E(G) R +, themaximumdirectedcutproblem(max

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 35, No., May 010, pp. 84 305 issn 0364-765X eissn 156-5471 10 350 084 informs doi 10.187/moor.1090.0440 010 INFORMS On the Power of Robust Solutions in Two-Stage

More information

On self-concordant barriers for generalized power cones

On self-concordant barriers for generalized power cones On self-concordant barriers for generalized power cones Scott Roy Lin Xiao January 30, 2018 Abstract In the study of interior-point methods for nonsymmetric conic optimization and their applications, Nesterov

More information

Approximation Algorithms for Homogeneous Polynomial Optimization with Quadratic Constraints

Approximation Algorithms for Homogeneous Polynomial Optimization with Quadratic Constraints Approximation Algorithms for Homogeneous Polynomial Optimization with Quadratic Constraints Simai HE Zhening LI Shuzhong ZHANG July 19, 010 Abstract In this paper, we consider approximation algorithms

More information

Lecture 8. Strong Duality Results. September 22, 2008

Lecture 8. Strong Duality Results. September 22, 2008 Strong Duality Results September 22, 2008 Outline Lecture 8 Slater Condition and its Variations Convex Objective with Linear Inequality Constraints Quadratic Objective over Quadratic Constraints Representation

More information

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut CMPUT 675: Approximation Algorithms Fall 011 Lecture 17 (Nov 3, 011 ): Approximation via rounding SDP: Max-Cut Lecturer: Mohammad R. Salavatipour Scribe: based on older notes 17.1 Approximation Algorithm

More information

Convex Optimization Boyd & Vandenberghe. 5. Duality

Convex Optimization Boyd & Vandenberghe. 5. Duality 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Introduction to Semidefinite Programming I: Basic properties a

Introduction to Semidefinite Programming I: Basic properties a Introduction to Semidefinite Programming I: Basic properties and variations on the Goemans-Williamson approximation algorithm for max-cut MFO seminar on Semidefinite Programming May 30, 2010 Semidefinite

More information

On distributional robust probability functions and their computations

On distributional robust probability functions and their computations On distributional robust probability functions and their computations Man Hong WONG a,, Shuzhong ZHANG b a Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong,

More information

Lecture 8: Semidefinite programs for fidelity and optimal measurements

Lecture 8: Semidefinite programs for fidelity and optimal measurements CS 766/QIC 80 Theory of Quantum Information (Fall 0) Lecture 8: Semidefinite programs for fidelity and optimal measurements This lecture is devoted to two examples of semidefinite programs: one is for

More information

Robust linear optimization under general norms

Robust linear optimization under general norms Operations Research Letters 3 (004) 50 56 Operations Research Letters www.elsevier.com/locate/dsw Robust linear optimization under general norms Dimitris Bertsimas a; ;, Dessislava Pachamanova b, Melvyn

More information

New stopping criteria for detecting infeasibility in conic optimization

New stopping criteria for detecting infeasibility in conic optimization Optimization Letters manuscript No. (will be inserted by the editor) New stopping criteria for detecting infeasibility in conic optimization Imre Pólik Tamás Terlaky Received: March 21, 2008/ Accepted:

More information

1. Introduction. Consider the following quadratic binary optimization problem,

1. Introduction. Consider the following quadratic binary optimization problem, ON DUALITY GAP IN BINARY QUADRATIC PROGRAMMING XIAOLING SUN, CHUNLI LIU, DUAN LI, AND JIANJUN GAO Abstract. We present in this paper new results on the duality gap between the binary quadratic optimization

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs 2015 American Control Conference Palmer House Hilton July 1-3, 2015. Chicago, IL, USA Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca and Eilyan Bitar

More information

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization H. Mansouri M. Zangiabadi Y. Bai C. Roos Department of Mathematical Science, Shahrekord University, P.O. Box 115, Shahrekord,

More information

1 Review of last lecture and introduction

1 Review of last lecture and introduction Semidefinite Programming Lecture 10 OR 637 Spring 2008 April 16, 2008 (Wednesday) Instructor: Michael Jeremy Todd Scribe: Yogeshwer (Yogi) Sharma 1 Review of last lecture and introduction Let us first

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

1 The independent set problem

1 The independent set problem ORF 523 Lecture 11 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Tuesday, March 29, 2016 When in doubt on the accuracy of these notes, please cross chec with the instructor

More information

Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems

Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems V. Jeyakumar and G. Li Revised Version:August 31, 2012 Abstract An exact semidefinite linear programming (SDP) relaxation

More information

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Convex relaxation. In example below, we have N = 6, and the cut we are considering Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution

More information

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 3 Convergence (Chapter 5) Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let

More information

Relaxations and Randomized Methods for Nonconvex QCQPs

Relaxations and Randomized Methods for Nonconvex QCQPs Relaxations and Randomized Methods for Nonconvex QCQPs Alexandre d Aspremont, Stephen Boyd EE392o, Stanford University Autumn, 2003 Introduction While some special classes of nonconvex problems can be

More information

Robust Farkas Lemma for Uncertain Linear Systems with Applications

Robust Farkas Lemma for Uncertain Linear Systems with Applications Robust Farkas Lemma for Uncertain Linear Systems with Applications V. Jeyakumar and G. Li Revised Version: July 8, 2010 Abstract We present a robust Farkas lemma, which provides a new generalization of

More information

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method CSC2411 - Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method Notes taken by Stefan Mathe April 28, 2007 Summary: Throughout the course, we have seen the importance

More information

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Convex relaxation. In example below, we have N = 6, and the cut we are considering Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution

More information

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function Zhongyi Liu, Wenyu Sun Abstract This paper proposes an infeasible interior-point algorithm with

More information

Lecture: Duality.

Lecture: Duality. Lecture: Duality http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Introduction 2/35 Lagrange dual problem weak and strong

More information

IE 521 Convex Optimization

IE 521 Convex Optimization Lecture 14: and Applications 11th March 2019 Outline LP SOCP SDP LP SOCP SDP 1 / 21 Conic LP SOCP SDP Primal Conic Program: min c T x s.t. Ax K b (CP) : b T y s.t. A T y = c (CD) y K 0 Theorem. (Strong

More information

SDP Relaxations for MAXCUT

SDP Relaxations for MAXCUT SDP Relaxations for MAXCUT from Random Hyperplanes to Sum-of-Squares Certificates CATS @ UMD March 3, 2017 Ahmed Abdelkader MAXCUT SDP SOS March 3, 2017 1 / 27 Overview 1 MAXCUT, Hardness and UGC 2 LP

More information

Notes 6 : First and second moment methods

Notes 6 : First and second moment methods Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative

More information

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint Iranian Journal of Operations Research Vol. 2, No. 2, 20, pp. 29-34 A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint M. Salahi Semidefinite

More information

4. Algebra and Duality

4. Algebra and Duality 4-1 Algebra and Duality P. Parrilo and S. Lall, CDC 2003 2003.12.07.01 4. Algebra and Duality Example: non-convex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone

More information

Solutions to Exercises

Solutions to Exercises 1/13 Solutions to Exercises The exercises referred to as WS 1.1(a), and so forth, are from the course book: Williamson and Shmoys, The Design of Approximation Algorithms, Cambridge University Press, 2011,

More information

Lecture: Duality of LP, SOCP and SDP

Lecture: Duality of LP, SOCP and SDP 1/33 Lecture: Duality of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2017.html wenzw@pku.edu.cn Acknowledgement:

More information

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009 LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix

More information

CS261: A Second Course in Algorithms Lecture #9: Linear Programming Duality (Part 2)

CS261: A Second Course in Algorithms Lecture #9: Linear Programming Duality (Part 2) CS261: A Second Course in Algorithms Lecture #9: Linear Programming Duality (Part 2) Tim Roughgarden February 2, 2016 1 Recap This is our third lecture on linear programming, and the second on linear programming

More information

Distributionally Robust Discrete Optimization with Entropic Value-at-Risk

Distributionally Robust Discrete Optimization with Entropic Value-at-Risk Distributionally Robust Discrete Optimization with Entropic Value-at-Risk Daniel Zhuoyu Long Department of SEEM, The Chinese University of Hong Kong, zylong@se.cuhk.edu.hk Jin Qi NUS Business School, National

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. xx, No. x, Xxxxxxx 00x, pp. xxx xxx ISSN 0364-765X EISSN 156-5471 0x xx0x 0xxx informs DOI 10.187/moor.xxxx.xxxx c 00x INFORMS On the Power of Robust Solutions in

More information

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres Front. Math. China 2017, 12(6): 1409 1426 https://doi.org/10.1007/s11464-017-0644-1 Approximation algorithms for nonnegative polynomial optimization problems over unit spheres Xinzhen ZHANG 1, Guanglu

More information

Lecture: Examples of LP, SOCP and SDP

Lecture: Examples of LP, SOCP and SDP 1/34 Lecture: Examples of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Lagrange Duality Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Lagrangian Dual function Dual

More information

Nonsymmetric potential-reduction methods for general cones

Nonsymmetric potential-reduction methods for general cones CORE DISCUSSION PAPER 2006/34 Nonsymmetric potential-reduction methods for general cones Yu. Nesterov March 28, 2006 Abstract In this paper we propose two new nonsymmetric primal-dual potential-reduction

More information

Cubic regularization of Newton s method for convex problems with constraints

Cubic regularization of Newton s method for convex problems with constraints CORE DISCUSSION PAPER 006/39 Cubic regularization of Newton s method for convex problems with constraints Yu. Nesterov March 31, 006 Abstract In this paper we derive efficiency estimates of the regularized

More information

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers. Chapter 3 Duality in Banach Space Modern optimization theory largely centers around the interplay of a normed vector space and its corresponding dual. The notion of duality is important for the following

More information

The Trust Region Subproblem with Non-Intersecting Linear Constraints

The Trust Region Subproblem with Non-Intersecting Linear Constraints The Trust Region Subproblem with Non-Intersecting Linear Constraints Samuel Burer Boshi Yang February 21, 2013 Abstract This paper studies an extended trust region subproblem (etrs in which the trust region

More information

Duality of LPs and Applications

Duality of LPs and Applications Lecture 6 Duality of LPs and Applications Last lecture we introduced duality of linear programs. We saw how to form duals, and proved both the weak and strong duality theorems. In this lecture we will

More information

Largest dual ellipsoids inscribed in dual cones

Largest dual ellipsoids inscribed in dual cones Largest dual ellipsoids inscribed in dual cones M. J. Todd June 23, 2005 Abstract Suppose x and s lie in the interiors of a cone K and its dual K respectively. We seek dual ellipsoidal norms such that

More information

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence

More information

Convex and Semidefinite Programming for Approximation

Convex and Semidefinite Programming for Approximation Convex and Semidefinite Programming for Approximation We have seen linear programming based methods to solve NP-hard problems. One perspective on this is that linear programming is a meta-method since

More information

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS Yugoslav Journal of Operations Research 25 (205), Number, 57 72 DOI: 0.2298/YJOR3055034A A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM FOR P (κ)-horizontal LINEAR COMPLEMENTARITY PROBLEMS Soodabeh

More information

Solving Dual Problems

Solving Dual Problems Lecture 20 Solving Dual Problems We consider a constrained problem where, in addition to the constraint set X, there are also inequality and linear equality constraints. Specifically the minimization problem

More information

Lecture 6: Conic Optimization September 8

Lecture 6: Conic Optimization September 8 IE 598: Big Data Optimization Fall 2016 Lecture 6: Conic Optimization September 8 Lecturer: Niao He Scriber: Juan Xu Overview In this lecture, we finish up our previous discussion on optimality conditions

More information

Conic Linear Optimization and its Dual. yyye

Conic Linear Optimization and its Dual.   yyye Conic Linear Optimization and Appl. MS&E314 Lecture Note #04 1 Conic Linear Optimization and its Dual Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization A Second Full-Newton Step On Infeasible Interior-Point Algorithm for Linear Optimization H. Mansouri C. Roos August 1, 005 July 1, 005 Department of Electrical Engineering, Mathematics and Computer Science,

More information

3.10 Lagrangian relaxation

3.10 Lagrangian relaxation 3.10 Lagrangian relaxation Consider a generic ILP problem min {c t x : Ax b, Dx d, x Z n } with integer coefficients. Suppose Dx d are the complicating constraints. Often the linear relaxation and the

More information

On the mean connected induced subgraph order of cographs

On the mean connected induced subgraph order of cographs AUSTRALASIAN JOURNAL OF COMBINATORICS Volume 71(1) (018), Pages 161 183 On the mean connected induced subgraph order of cographs Matthew E Kroeker Lucas Mol Ortrud R Oellermann University of Winnipeg Winnipeg,

More information

EE/AA 578, Univ of Washington, Fall Duality

EE/AA 578, Univ of Washington, Fall Duality 7. Duality EE/AA 578, Univ of Washington, Fall 2016 Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Downloaded 03/01/17 to Redistribution subject to SIAM license or copyright; see

Downloaded 03/01/17 to Redistribution subject to SIAM license or copyright; see SIAM J. DISCRETE MATH. Vol. 31, No. 1, pp. 335 382 c 2017 Society for Industrial and Applied Mathematics PARTITION CONSTRAINED COVERING OF A SYMMETRIC CROSSING SUPERMODULAR FUNCTION BY A GRAPH ATTILA BERNÁTH,

More information

arxiv: v1 [math.oc] 23 Nov 2012

arxiv: v1 [math.oc] 23 Nov 2012 arxiv:1211.5406v1 [math.oc] 23 Nov 2012 The equivalence between doubly nonnegative relaxation and semidefinite relaxation for binary quadratic programming problems Abstract Chuan-Hao Guo a,, Yan-Qin Bai

More information

Distributionally robust optimization techniques in batch bayesian optimisation

Distributionally robust optimization techniques in batch bayesian optimisation Distributionally robust optimization techniques in batch bayesian optimisation Nikitas Rontsis June 13, 2016 1 Introduction This report is concerned with performing batch bayesian optimization of an unknown

More information

Strongly chordal and chordal bipartite graphs are sandwich monotone

Strongly chordal and chordal bipartite graphs are sandwich monotone Strongly chordal and chordal bipartite graphs are sandwich monotone Pinar Heggernes Federico Mancini Charis Papadopoulos R. Sritharan Abstract A graph class is sandwich monotone if, for every pair of its

More information

A note on network reliability

A note on network reliability A note on network reliability Noga Alon Institute for Advanced Study, Princeton, NJ 08540 and Department of Mathematics Tel Aviv University, Tel Aviv, Israel Let G = (V, E) be a loopless undirected multigraph,

More information

Lecture 2 Sep 5, 2017

Lecture 2 Sep 5, 2017 CS 388R: Randomized Algorithms Fall 2017 Lecture 2 Sep 5, 2017 Prof. Eric Price Scribe: V. Orestis Papadigenopoulos and Patrick Rall NOTE: THESE NOTES HAVE NOT BEEN EDITED OR CHECKED FOR CORRECTNESS 1

More information

On the sufficiency of finite support duals in semi-infinite linear programming

On the sufficiency of finite support duals in semi-infinite linear programming On the sufficiency of finite support duals in semi-infinite linear programming Amitabh Basu a, Kipp Martin b, Christopher Thomas Ryan b a The Johns Hopkins University b University of Chicago, Booth School

More information

Approximation Algorithms for Maximum. Coverage and Max Cut with Given Sizes of. Parts? A. A. Ageev and M. I. Sviridenko

Approximation Algorithms for Maximum. Coverage and Max Cut with Given Sizes of. Parts? A. A. Ageev and M. I. Sviridenko Approximation Algorithms for Maximum Coverage and Max Cut with Given Sizes of Parts? A. A. Ageev and M. I. Sviridenko Sobolev Institute of Mathematics pr. Koptyuga 4, 630090, Novosibirsk, Russia fageev,svirg@math.nsc.ru

More information

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming

More information

On the Sandwich Theorem and a approximation algorithm for MAX CUT

On the Sandwich Theorem and a approximation algorithm for MAX CUT On the Sandwich Theorem and a 0.878-approximation algorithm for MAX CUT Kees Roos Technische Universiteit Delft Faculteit Electrotechniek. Wiskunde en Informatica e-mail: C.Roos@its.tudelft.nl URL: http://ssor.twi.tudelft.nl/

More information

Katarzyna Mieczkowska

Katarzyna Mieczkowska Katarzyna Mieczkowska Uniwersytet A. Mickiewicza w Poznaniu Erdős conjecture on matchings in hypergraphs Praca semestralna nr 1 (semestr letni 010/11 Opiekun pracy: Tomasz Łuczak ERDŐS CONJECTURE ON MATCHINGS

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Chapter 13. Convex and Concave. Josef Leydold Mathematical Methods WS 2018/19 13 Convex and Concave 1 / 44

Chapter 13. Convex and Concave. Josef Leydold Mathematical Methods WS 2018/19 13 Convex and Concave 1 / 44 Chapter 13 Convex and Concave Josef Leydold Mathematical Methods WS 2018/19 13 Convex and Concave 1 / 44 Monotone Function Function f is called monotonically increasing, if x 1 x 2 f (x 1 ) f (x 2 ) It

More information

Iteration-complexity of first-order penalty methods for convex programming

Iteration-complexity of first-order penalty methods for convex programming Iteration-complexity of first-order penalty methods for convex programming Guanghui Lan Renato D.C. Monteiro July 24, 2008 Abstract This paper considers a special but broad class of convex programing CP)

More information

Out-colourings of Digraphs

Out-colourings of Digraphs Out-colourings of Digraphs N. Alon J. Bang-Jensen S. Bessy July 13, 2017 Abstract We study vertex colourings of digraphs so that no out-neighbourhood is monochromatic and call such a colouring an out-colouring.

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize

More information

Semidefinite Programming

Semidefinite Programming Semidefinite Programming Notes by Bernd Sturmfels for the lecture on June 26, 208, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra The transition from linear algebra to nonlinear algebra has

More information

Positive Semi-definite programing and applications for approximation

Positive Semi-definite programing and applications for approximation Combinatorial Optimization 1 Positive Semi-definite programing and applications for approximation Guy Kortsarz Combinatorial Optimization 2 Positive Sem-Definite (PSD) matrices, a definition Note that

More information

CS229T/STATS231: Statistical Learning Theory. Lecturer: Tengyu Ma Lecture 11 Scribe: Jongho Kim, Jamie Kang October 29th, 2018

CS229T/STATS231: Statistical Learning Theory. Lecturer: Tengyu Ma Lecture 11 Scribe: Jongho Kim, Jamie Kang October 29th, 2018 CS229T/STATS231: Statistical Learning Theory Lecturer: Tengyu Ma Lecture 11 Scribe: Jongho Kim, Jamie Kang October 29th, 2018 1 Overview This lecture mainly covers Recall the statistical theory of GANs

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION   henrion COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS Didier HENRION www.laas.fr/ henrion October 2006 Geometry of LMI sets Given symmetric matrices F i we want to characterize the shape in R n of the LMI set F

More information