Approximate Constraint Satisfaction Requires Large LP Relaxations

Size: px
Start display at page:

Download "Approximate Constraint Satisfaction Requires Large LP Relaxations"

Transcription

1 Approximate Constraint Satisfaction Requires Large LP Relaxations oah Fleming April 19, 2018 Linear programming is a very powerful tool for attacking optimization problems. Techniques such as the ellipsoi metho have shown that linear programs are solvable in polynomial time. Furthermore, it is known linear programming is P-complete. Therefore, if one was to show that some P-har problem amitte a polynomial-size linear program, then P = P. In an attempt to rule out this approach, Yannakakis [4] gave a framework for proving lower bouns on a large class of linear programs known as extene formulations. Consier the 3XOR n problem on n variables. It s not P-har, but it will serve as a goo running example. An instance of Π 3XOR n consists of m parity constraints {P 1,..., P m }, P l : {±1} n {0, 1} where P l (x) := x i x j x k = a l, for i, j, k [n] an a l {±1} n ; the goal is to maximize the number of constraints satisfie. ote that Π can also be represente uniquely as a multilinear polynomial over {±1} n by taking the Fourier expansion. We can rewrite each P l (x) = x i x j x k = a l as P l (x) := ( 1) 1 a l 2 x i x j x k. The value of Π on some assignment x { 1, 1} n is given by Π(x) = 1 m P i (x), i [m] which is the fraction of constraints satisfie by assignment x. We will enote by opt(π) = max Π(x), x { 1,1} n the largest fraction of constraints of Π satisfiable by any assignment x { 1, 1} n. If we want to express this as a linear program, then we nee to linearize this function. To o this, we can associate some orering to the 2 ( n 3) possible 3XORn constraints, P 1,..., P 2( n 3). 1

2 A natural way of linearizing such a function is to associate with each 3XOR n instance Π on m vertices, a vector Π R 2(n 3), where the ith entry is 1/m if Π contains contains constraint P i an 0 otherwise. Similarly, we can associate with each assignment x { 1, 1} n a vector x R 2(n 3) in which xi = 1 if P i (x) = 1 an 0 otherwise. This satisfies, for every 3XOR n instance Π an assignment x { 1, 1} n, that Π, x = Π(x). This lens itself to a natural linear program: let P R 2(n 3) be the convex hull of all x for x { 1, 1} n ; the linear program is given by L(Π) = max y, Π. y P This polytope P has vertices corresponing to the points x for x { 1, 1} n an facets corresponing to the encoings Π of all 3XOR n instances Π. Therefore, the value returne by optimizing over P will be opt(π). Unfortunately, the polytope P has an exponential number of facets an therefore cannot be optimize over efficiently. One possible way to overcome this issue is to fin some new polytope P in a higher imensional space R n with much fewer facets an such that there is a linear projection from P own to P. We coul then optimize over the new polytope P instea of optimizing over P. Such a polytope P is known as an extene formulation of the polytope P. The size of an extene formulation is the number of facets of the polytope, while the extension complexity of the base polytope P, enote xc(p) is the smallest extene formulation of P. We stress that an extene formulation P epens only on the input size an not the particular instance Π 3XOR n that we want to compute; the instance Π is efine only in he objective function. Yannakakis gave a beautiful characterization of the extension complexity of a polytope in terms of the non-negative rank of its slack matrix. Consier a linear program P computing 3XOR n. The slack matrix M S has rows corresponing to the instances Π 3XOR n, an columns corresponing to the vertices x of P. The entry at some row, column ( Π, x) is the slack between that vertex an that instance, M S Π, x := L(Π) x, Π, where L(Π) = max y P y, Π. The non-negative rank of a matrix M, enote rk + (M) is the smallest imension r such that M can be written as a prouct of two non-negative matrices F an V with inner-imension r. Theorem 1. (Yannakakis [4]) For any polytope P, xc(p) + 1 = rk + (P) The Proof relies on Farkas Lemma. 2

3 = Figure 1: Representation of Theorem 1, the ecomposition of the slack matrix into two non-negative matrices with inner imension r. Lemma 1. (Farkas Lemma) Let P be a with facets efine by inequalities {A 1 x b 1,..., A m x b r } an let Cx be an inequality that is vali for P (that is, every point α P satisfies Cα ) then there exists λ 0,..., λ r R 0 such that Cx = λ 0 + r λ(b i A i x) We will only prove the forwar irection, since it is all that we will nee. Proof. (of Theorem 1) Let P be an extene formulation of P such that P has r facets, efine by inequalities A 1 x b 1,..., A r x b r. Observe that for every Π 3XOR n, the inequality opt(π) Π, y 0 is vali for the polytope P, for every y P an furthermore, that opt(π) = L(Π) because P computes 3XOR n exactly. Applying Farkas Lemma, we can write L(Π) Π, y = λ 0 ( Π) + r λ i ( Π) (b i A i, y ), (1) for some λ 0 ( Π),..., λ r ( Π) R 0. ow, because there is a linear projection from P to P, there is a vertex v of P that projects to each vertex x of P. We will restrict to these vertices, L(Π) Π, x = λ 0 ( Π) + r λ i ( Π) (b i A i, v ). (2) Furthermore, the x are in one-to-one corresponence with the x { 1, 1} n, we can rewrite this as each b i A i, v as a non-negative function q i : { 1, 1} n R 0, where q i (x) = b i A i, v. 3

4 Therefore, we can rewrite equation 2 as L(Π) Π, x = λ0 ( Π) + r λ i ( Π) q i (x); this is the slack between vertex x an instance Π. We now construct the non-negative matrices V an F with inner imension r + 1. Let the rows of V be inexe by the Π for Π 3XOR n an the columns of F be inexe by the x for x { 1, 1} n. The Πth row of V will be the vector [λ 0 ( Π),..., λ r ( Π)] corresponing to Π. The ith row of F is the truth table encoing of q i, where the (i, j)th entry of F is the evaluation of q i (x). The (r + 1)st row of F is the all 1 vector. This can be seen in figure 1. Therefore, the inner prouct between V Π an F x is r λ 0 ( Π) + λ i ( Π) q i (x) = L(Π) Π, x. ote: Because the extene formulation computes Π 3XOR n exactly, L(Π) = opt(π). Furthermore because the rows an columns of the slack matrix are in one-to-one corresponence between x { 1, 1} n an Π 3XOR n, the ( x, Π)th entry of the slack matrix is equivalent to M S Π,x = opt(π) Π(x), because, x, Π = Π(x). Therefore, the slack matrix will be the same for any base polytope P, the particular linearization is irrelevant. Therefore, more generally, we can efine an extene formulation that exactly computes 3XOR n as a polytope P R n such that 1. for every assignment x { 1, 1} n there is a vector x P an for every instance Π 3XOR n there is a vector Π R such that Π(x) = x, Π 2. opt(π) = max y P y, Π for every Π 3XOR n. The extension complexity of 3XOR n, xc(3xor n ) is then the smallest extene formulation for 3XOR n. The key fact from Theorem 1 that we will use is that if an extene formulation P of size r computes 3XOR n then, for every instance Π, there exists a representation L(Π) Π = λ 0 (Π) + r λ i q i (Π), where each q i is a slack function of P. We will call this representation an extene formulation witness, because it witnesses that P computes Π. From now on, we will write λ i (Π) as simply λ i, where the epenence on Π is implicit. 4

5 Recall that the egree- Sherali-Aams hierarchy computes an instance Π 3XOR n if opt(π) Π can be written as a non-negative linear combination of -juntas, q i, opt(π) Π = i I λ i q i. This representations is superficially similar, an one might woner if there is a way to approximate an extene formulation witness with a Sherali-Aams witness. Obviously, it woul be too much to hope for that each of the non-negative functions in the extene formulation witness coul be well approximate by a non-negative junta. Surprisingly, Chan, Lee, Raghavenra an Steurer [1] showe that after a specialize ranom restriction, the resulting q i can be well approximate by non-negative -juntas. Using this, they are able to lift Sherali-Aams lower bouns to extension lower complexity lower bouns. This transformation works for the class of constraint satisfaction problems (CSP), but we will prove it for the special case of 3XOR n. Theorem 2. ([1]) Suppose that the (n)-roun Sherali-Aams relaxation cannot compute 3XOR n, then for all sufficiently large n, no extene formulation of size at most n (n)2 can compute 3XOR for some = n 10(n) We begin with the family of 3XOR instances over variables, an some extene formulation P of size r. By Yannakakis Theorem above, each instance Π 3XOR, can be written as L(Π) Π = r λ iq i, where each q i is a function { 1, 1} R 0. Our goal is to write (a restriction of) L(Π) Π as a non-negative linear combination of -juntas plus some small error term. The proof procees in three steps. 1. First, we show that we can restrict our attention to q i that are sufficiently smooth (the infinity norm of these functions is boune). 2. Then, we show that each of these q i can be approximate by an 0.2 -junta q i, such that the error on the low egree Fourier coefficients of q i q i is small. Here we crucially use the fact that egree- Sherali-Aams can only reason about monomials of egree up to. This step will incur some error, but we will show that this error goes to 0 as n goes to infinity. 3. Up until now, this proof has worke for any instance Π 3XOR. We will now fix a particular instance which will allow us to make the connection to Sherali-Aams lower bouns. Let Π 0 3XOR n be a har instance for Sherali-Aams on n variables. To obtain the instance Π 3XOR, we plant Π 0 at ranom insie a larger space of variables by picking a subset of n of the variables an efining the constraints of Π 0 on them; the remaining n variables will remain unconstraine. Finally, we argue that with high probability, the set of significant coorinates of q i when restricte to the variables on which Π 0 is efine is at most, an so the existence of this extene formulation implies that egree- Sherali-Aams computes this instance exactly. 5

6 Fourier Analysis We will nee several tools from Fourier analysis. We will efine the inner prouct between two n-variable functions f an g as f, g := E x { 1,1} n[f(x)g(x)], where the expectation is taken over the uniform istribution on { 1, 1} n. The Fourier representation of a function f : { 1, 1} n R is its unique representation over the basis of parity functions χ α := i α x i for α [n]. We can represent f over this basis as f = α [n] ˆf(α)χ α, where the fourier coefficient ˆf(α) is efine as f in the χ α irection, ˆf(α) := f, χ α. Intuitively, ˆf(α) measures the correlation of the variables Πi α x i. Throughout this, we will use the functions regular representation an its Fourier representation interchangeably. Furthermore, if f is non-negative an E x { 1,1} n[f(x)] = 1, then we can treat the Fourier coefficients of f as a istribution over { 1, 1} n. Step 1: Smooth Slack Functions We will now prove the main theorem by following the three steps lai out previously. Again, suppose that we have an extene formulation P of size r /2 which computes 3XOR exactly. By Yannakakis Theorem, for any Π 3XOR, we can write Π as a sum of nonnegative slack functions, r L(Π) Π = λ 0 + λ i q i, where λ i 0 an q i : { 1, 1} R 0. Furthermore, we can normalize each q i an write it as q i (x) = γ i q i (x) for some γ i R 0 such that E[q i ] = 1. That is, Define the set L(Π) Π = λ 0 + r (λ i γ i ) q i. Q := {i : q i }, of the q i which are fairly smooth. Recall that is the egree of the Sherali-Aams proof we are trying to obtain. We will show that restricting attention to the set of functions Q will only incur a small aitive error. We can ecompose the previous sum into L(Π) Π = λ 0 + (λ i γ i ) q i + (λ j γ j ) q j. i Q 6 j [r]\q

7 Because the value of opt(π) Π(x) [0, 1] for every x { 1, 1} an Π 3XOR an because λ j, γ j 0 an q j is non-negative an E[q j ] = 1, we must have λ j γ j for every instance Π 3XOR. Because of this, j [r]\q (λ jγ j )q j cannot be very large an we will treat it as some small aitive error term, which we will enote by ε(π). Later, we will boun its value, L(Π) Π = λ 0 + i Q(λ i γ i ) q i + ε(π). Step 2: Approximate Functionals by High-Degree Juntas The aim now is to show that the smooth slack functions q i for i Q can be well approximate by high-egree juntas. For this, we will use a ensity version of Chang s Lemma. The proof follows from the entropic proof of Chang s Lemma in Impagliazzo, Moore an Russell [2]. Lemma 2. (Chang s Lemma) Let q be a ensity with entropy at least t for some t 0, let σ > 0 an efine R = {α : ˆf(α) σ2 t }. Then R spans a space of imension less than 2t/σ 2 A consequence of Chang s Lemma is the following. Lemma 3. If q i has entropy at least log, then for any σ > 0, there exists a set J(q i ) [] with J(q i ) 22 log σ 2 such that for every α J(q i ) with α, we have ˆq(α) σ. Proof. Consier S = { α : ˆq(α) σ} an let S be the maximal set of linearly inepenent elements in S. The ensity version of Chang s Lemma states that, after setting t = log, that S 2σ 2 log. Let J(q i ) = α S α, then J(q i ) 2 2 log /σ 2 because each α contains at most elements (it follows by linear inepenence that for all α J(q i ) with α, that ˆq i (α) sigma. This lemma says that we can ecompose any high-entropy q i into two parts q i an e i, where q i = ˆq i (α)χ α, an e i = ˆq i (α)χ α. α J(q i ) α []\J(q i ) That is, q i epens only on the set of variables in J(q i ) an in e i, the Fourier coefficients of correlations up to egree- are very small. Beyon egree- we have no control over the magnitue of the Fourier coefficients in e i. However, recall that the egree- Sherali-Aams hierarchy can only perceive correlations of egree up to. Therefore, because our en goal is to convert this into a Sherali-Aams proof, this is a non-issue for us. Therefore, if we coul ensure that each of the q i were -junta that is, that J(q i ), an that the extra error e i was small, then the proof woul be finishe. We woul have 7

8 arrive at a representation of L(Π) Π consisting of -juntas plus some small aitive error. Unfortunately, because we nee i Q e i to ten to 0 as n, it turns out that the largest ( ) 16n that we will be able to set σ, an still achieve this is σ = 2 log 1/2. Uner Lemma 3, this only guarantees that each q i is an ( /8n)-junta, which is approximately 0.2 when the final numbers are plugge in. Finally, we verify that each q i with i Q inee has high enough entropy to satisfy the hypothesis of Lemma 3: H(q i ) = ( ) q i (x) 2 2 log ( ) q i (x) 2 log q i (x) 2 q i x { 1,1} x { 1,1} ( ) q i (x) 2 log = log, 2 x { 1,1} where we use the fact that E[q i ] = 1. So far we have achieve a representation of the form L(Π) Π = λ 0 + i Q(λ i γ i ) (q i + e i ) + ε(π), where e i are error terms whose Fourier coefficients corresponing to egree-up-to- correlations are boune by σ, an q i are 0.2 -juntas. Step 3: Ranom Restriction to a Har Instance for Sherali-Aams The final step is to reuce the 0.2 -juntas to -Juntas. To o this, we will employ a special ranom restriction which will restrict to an instance Π 0 3XOR n for which we have Sherali- Aams lower bouns. ote that until this point, the steps of the proof have not relie on the particular instance of 3XOR. We will now restrict attention to a particular sub-family of instances. Consier an instance Π 0 of 3XOR n on n variables, where n is much smaller than (Π 0 shoul be thought of as a har instance for Sherali-Aams). To create our instance Π, we will ranomly plant Π 0 insie a larger space of unconstraine variables by picking a subset S of n variables an efining the constraints of Π 0 on them. The iea is that since the only constraints in Π are those corresponing to Π 0, L(Π) = opt(π) = opt(π 0 ). ow, because each of the junta q i epen on at most 0.2 variables, then if we restrict to the variables of Π 0, with high probability only a small fraction of the variables on which q i epens will remain. This can be seen in figure 2. This will be one in the following lemma; recall that in step 2, using Chang s Lemma, we ecompose q i = q i + e i. 8

9 [] var(q 1 ') var(q 4 ') S var(q 2 ') var(q 3 ') var(q 5 ') Figure 2: The intersection between the variable space of each 0.2 -junta q i an the restricte set S on which we will plant Π 0. Lemma 4. There exists a set S [] of size n such that for each q i with i Q, there is a set J(q) S with J(q) such that for all α S \ J(q) with α. ˆq(α) ( 16n 2 log ) 1/2, For the proof, we will nee the following inequality. Let X 1,..., X n be i.i.. {0, 1}- ranom variables, with E[X i ] = p. Then [ n ] Pr X i t (pn) t (3) Proof. We will choose the set S as follows: 1. Uniformly at ranom, pick a partition of [] into sets S 1,..., S n, each of size /n. 2. For each variable i [n], pick a variable v i from S i uniformly at ranom. 3. Let S = {v i : i [n]} 9

10 In step 2 we argue, using Lemma 3, that we coul ecompose q i = q i + e i, where each q i is an /8n-junta which epens on a set of coorinates J(q i ), an for every α [] ( ) 1/2. with α, e i (α) We will show that with some positive probability, the 16n 2 log n intersection of each of the sets J(q i ) with the set S is at most. For each variable l J(q i ), let X l be the event l S. Then, E[q i ] = n/ because we are choosing each element of S uniformly at ranom, an so Pr[ J(q i) S ] = Pr ( X l n ) 1 J(q i) 8, /2 l J(q i ) where the secon inequality follows from inequality 3 above. Finally, because we have assume that our original extene formulation is of size at most /2, we have that Q /2, an so taking a union boun over all J(q i ) for i Q completes the proof. Finally, we construct the instance Π 3XOR as follows: Let Π 0 be an instance of 3XOR n on n variables. Apply Lemma 4 to obtain a subset S = {v 1,..., v n } []. Define the constraints of Π as the constrains of Π 0 efine on the variables {v 1,..., v n }; the remaining n variables are left unconstraine. Let E be the egree- Sherali-Aams pseuo-expectation which achieves the optimal value on the Π 0, E = SA [Π 0 ] = max Ẽ[Π 0 ], Ẽ P E where we think of Π(x) as its representation as a multilinear polynomial so that we can apply E to it. Furthermore, we can represent each of the functions q i as a multilinear polynomial by taking its Fourier transform. We will think of q i as having that representation from now on so that we can apply E to them. We now plant E on the set of variables S, that is, we efine E on the variables in S an exten it to have Fourier coefficient 0 on all terms outsie of S. To o this, we note that Π is unconstraine on variables outsie of S an therefore, we efine the unerlying pseuo-istribution to be uniform on all variables on outsie of S. Applying it to both sies of equation?? we arrive at E [L(Π) Π(x)] = λ 0 + i Q λ i γ i (E [q i] + E [e i ]) + E [ε(π)] L(Π) SA [Π 0 ] = λ 0 + i Q λ i γ i (E [q i] + E [e i ]) + E [ε(π)], ow, because E gives non-zero value only on the variables of S, we have that E [q i] = E [q i S ], an so, by Lemma 4, we know that q i epens only on at most variables in S, so it is a non-negative -junta. Therefore, E [q i] 0, an so L(Π) SA [Π 0 ] + i Q λ i γ i E [e i ] + E [ε(π)]. 10

11 ow, if we can show the error terms, i Q λ iγ i E [e i ] + E [ε(π)] go to 0 as n, then we will arrive at a representation of the form L(Π) SA [Π 0 ], an so Sherali-Aams lower bouns will imply extension complexity lower bouns. Bouning Error Terms All that is left is to show that the error terms go to 0 as n. We begin with bouning E [ε(π)] = (λ j γ j ) E [q j ], (4) j [r]\q the error term that we obtaine from Step 1. We will nee a simple fact about pseuoexpectations Claim 1. For any egree- pseuo-expectation E in its Fourier representation as a multilinear polynomial over { 1, 1} n, we have E, α [n] E [χ α ] i=0 ( n i). Proof. The Fourier representation of E is E = α E [χ α ]χ α, where χ α = i α x i. We know that because E is a pseuo-expectation that E [χ α ] 1 (this follows because a pseuo-expectation is the expectation over a pseuo-istribution). Because it is a egree- pseuo-expectation, there can be at most ( n i=0 ) non-zero Fourier coefficients, each having absolute value at most 1. We can apply this claim as follows: Viewing q j as its Fourier representation as a multilinear polynomial, an noting that E assigns values to monomials of egree at most we can write E [q j ] E i=0 ( ) n, the first inequality follows because the Fourier representation of E is E = α E [χ α ] an so the Fourier coefficient corresponing to the monomial χ α is E [χ α ], the value that E assigns to χ α. Therefore, we can view E as a vector whose αth place is the Fourier coefficient E [χ α ]. Similarly, we view q j as a vector, where the αth entry is the Fourier coefficient ˆf(α), which is the coefficient of χ α in the representation of q j as a multilinear polynomial. Therefore, taking the inner prouct between these two vectors E, q j gives the evaluation E [q j ]. Because q i is a ensity, that is E[q j ] = 1 an q j 0, we can represent it as a istribution over assignments x { 1, 1} n, an associate with each such assignment a 11

12 set which contains element i if x i = 1. Then, E, q j = E α qi [E [α] E During step 1, we argue that because the q j, λ j, γ j 0, that q i for every i [r] \ Q, an because L(Π) Π 0 [0, 1], we must have (λ j γ j ) < for each j [r] \ Q. Putting all of this together, we have E [ε(π)] = j [r]\q (λ j γ j ) E [q j ] ( ) n ( en ) r /2, i where the secon inequality follows because [r] \ Q r, an the final inequality follows because r, the size of the extene formulation, is at most /2. Finally, we boun the error term i Q (λ iγ i ) E [e i ] in a similar way, by noting that E [e i ] α S i=0 E [χ α ] ê i (α) = ( ) ( n 16n 2 log α S: α ) 1/2 ( en E [χ α ] ê i (α) α S: α ) ( 16n 2 log ) 1/2, ( ) 16n E 2 1/2 log [χ α ] where the equality follows because E [χ α ] = 0 for all α >, the thir inequality follows from the boun we got on the size of the egree-up-to- Fourier coefficients of e i from Lemma 4, an the fourth inequality follows from Claim 1. Therefore, we have ( en (λ i γ i ) E [e i ] i Q ) ( ) 16n 2 1/2 log ( en ) ( ) 16n 2 1/2 log (λ i γ i ), i Q where the final inequality follows from the observation that r (λ iγ i ) 1. To see this, note that opt(π) Π [0, 1], an E[q i ] = 1, an apply an expectation over assignments in { 1, 1} n to both sies of opt Π(x) = λ 0 + r (λ iγ i ) q i. Finishing Up Finally, putting everything together, we have L(Π) SA [Π 0 ] = λ 0 + i Q λ i γ i (E [q i] + E [e i ]) + E [ε(π)] λ 0 + ( en i γ i ) E i Q(λ [q i] ) ( ) 16n 2 1/2 log ( en ) /2. Therefore, we arrive at an expression of the form L(Π) SA [Π 0 ] err n. 12

13 We now show that err n := ( ) ( en 4 n log our value for = n 10 we have 1/4 ) ( en ) ( 4 10n log n err n = n ( 5/2 e 4 10n log n + 1 = n 3/2 = o(1). + ( ) en /2 goes to 0 as n. Plugging in ), ) + n 5 Therefore, this theorem lifts Sherali-Aams egree lower bouns of up to to extene formulation lower bouns. Unfortunately, because we set nee to set = n 10, the best lower boun that we can achieve this way (lifting a Sherali-Aams lower boun of egree Ω(n) is o( log log log ). The bottleneck, which limits Chan et al. to only obtaining quasi-polynomial size lower bouns is the application of Chang s Lemma in step 2. Later, Kothari, Meka an Raghavenra [3] overcame this barrier, obtaining truly exponential lower bouns. References [1] Siu On Chan, James R. Lee, Prasa Raghavenra, an Davi Steurer. Approximate constraint satisfaction requires large LP relaxations. J. ACM, 63(4):34:1 34:22, [2] Russell Impagliazzo, Cristopher Moore, an Alexaner Russell. An entropic proof of chang s inequality. SIAM J. Discrete Math., 28(1): , [3] Aam R. Klivans, Pravesh K. Kothari, an Raghu Meka. Efficient algorithms for outlierrobust regression. CoRR, abs/ , [4] Mihalis Yannakakis. Expressing combinatorial optimization problems by linear programs (extene abstract). In Proceeings of the 20th Annual ACM Symposium on Theory of Computing, May 2-4, 1988, Chicago, Illinois, USA, pages , 1988., 13

Permanent vs. Determinant

Permanent vs. Determinant Permanent vs. Determinant Frank Ban Introuction A major problem in theoretical computer science is the Permanent vs. Determinant problem. It asks: given an n by n matrix of ineterminates A = (a i,j ) an

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Discrete Mathematics

Discrete Mathematics Discrete Mathematics 309 (009) 86 869 Contents lists available at ScienceDirect Discrete Mathematics journal homepage: wwwelseviercom/locate/isc Profile vectors in the lattice of subspaces Dániel Gerbner

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Lecture 5. Symmetric Shearer s Lemma

Lecture 5. Symmetric Shearer s Lemma Stanfor University Spring 208 Math 233: Non-constructive methos in combinatorics Instructor: Jan Vonrák Lecture ate: January 23, 208 Original scribe: Erik Bates Lecture 5 Symmetric Shearer s Lemma Here

More information

Self-normalized Martingale Tail Inequality

Self-normalized Martingale Tail Inequality Online-to-Confience-Set Conversions an Application to Sparse Stochastic Banits A Self-normalize Martingale Tail Inequality The self-normalize martingale tail inequality that we present here is the scalar-value

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Quantum Mechanics in Three Dimensions

Quantum Mechanics in Three Dimensions Physics 342 Lecture 20 Quantum Mechanics in Three Dimensions Lecture 20 Physics 342 Quantum Mechanics I Monay, March 24th, 2008 We begin our spherical solutions with the simplest possible case zero potential.

More information

PDE Notes, Lecture #11

PDE Notes, Lecture #11 PDE Notes, Lecture # from Professor Jalal Shatah s Lectures Febuary 9th, 2009 Sobolev Spaces Recall that for u L loc we can efine the weak erivative Du by Du, φ := udφ φ C0 If v L loc such that Du, φ =

More information

7.1 Support Vector Machine

7.1 Support Vector Machine 67577 Intro. to Machine Learning Fall semester, 006/7 Lecture 7: Support Vector Machines an Kernel Functions II Lecturer: Amnon Shashua Scribe: Amnon Shashua 7. Support Vector Machine We return now to

More information

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell Lower bounds on the size of semidefinite relaxations David Steurer Cornell James R. Lee Washington Prasad Raghavendra Berkeley Institute for Advanced Study, November 2015 overview of results unconditional

More information

On combinatorial approaches to compressed sensing

On combinatorial approaches to compressed sensing On combinatorial approaches to compresse sensing Abolreza Abolhosseini Moghaam an Hayer Raha Department of Electrical an Computer Engineering, Michigan State University, East Lansing, MI, U.S. Emails:{abolhos,raha}@msu.eu

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

linear programming and approximate constraint satisfaction

linear programming and approximate constraint satisfaction linear programming and approximate constraint satisfaction Siu On Chan MSR New England James R. Lee University of Washington Prasad Raghavendra UC Berkeley David Steurer Cornell results MAIN THEOREM: Any

More information

Lecture 12: November 6, 2013

Lecture 12: November 6, 2013 Information an Coing Theory Autumn 204 Lecturer: Mahur Tulsiani Lecture 2: November 6, 203 Scribe: Davi Kim Recall: We were looking at coes of the form C : F k q F n q, where q is prime, k is the message

More information

A Sketch of Menshikov s Theorem

A Sketch of Menshikov s Theorem A Sketch of Menshikov s Theorem Thomas Bao March 14, 2010 Abstract Let Λ be an infinite, locally finite oriente multi-graph with C Λ finite an strongly connecte, an let p

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Technical Report TTI-TR-2008-5 Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri UC San Diego Sham M. Kakae Toyota Technological Institute at Chicago ABSTRACT Clustering ata in

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

Chromatic number for a generalization of Cartesian product graphs

Chromatic number for a generalization of Cartesian product graphs Chromatic number for a generalization of Cartesian prouct graphs Daniel Král Douglas B. West Abstract Let G be a class of graphs. The -fol gri over G, enote G, is the family of graphs obtaine from -imensional

More information

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks A PAC-Bayesian Approach to Spectrally-Normalize Margin Bouns for Neural Networks Behnam Neyshabur, Srinah Bhojanapalli, Davi McAllester, Nathan Srebro Toyota Technological Institute at Chicago {bneyshabur,

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes Leaving Ranomness to Nature: -Dimensional Prouct Coes through the lens of Generalize-LDPC coes Tavor Baharav, Kannan Ramchanran Dept. of Electrical Engineering an Computer Sciences, U.C. Berkeley {tavorb,

More information

Implicit Differentiation

Implicit Differentiation Implicit Differentiation Thus far, the functions we have been concerne with have been efine explicitly. A function is efine explicitly if the output is given irectly in terms of the input. For instance,

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

26.1 Metropolis method

26.1 Metropolis method CS880: Approximations Algorithms Scribe: Dave Anrzejewski Lecturer: Shuchi Chawla Topic: Metropolis metho, volume estimation Date: 4/26/07 The previous lecture iscusse they some of the key concepts of

More information

12.11 Laplace s Equation in Cylindrical and

12.11 Laplace s Equation in Cylindrical and SEC. 2. Laplace s Equation in Cylinrical an Spherical Coorinates. Potential 593 2. Laplace s Equation in Cylinrical an Spherical Coorinates. Potential One of the most important PDEs in physics an engineering

More information

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS ALINA BUCUR, CHANTAL DAVID, BROOKE FEIGON, MATILDE LALÍN 1 Introuction In this note, we stuy the fluctuations in the number

More information

Lecture 6 : Dimensionality Reduction

Lecture 6 : Dimensionality Reduction CPS290: Algorithmic Founations of Data Science February 3, 207 Lecture 6 : Dimensionality Reuction Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will consier the roblem of maing

More information

Monte Carlo Methods with Reduced Error

Monte Carlo Methods with Reduced Error Monte Carlo Methos with Reuce Error As has been shown, the probable error in Monte Carlo algorithms when no information about the smoothness of the function is use is Dξ r N = c N. It is important for

More information

arxiv: v1 [math.mg] 10 Apr 2018

arxiv: v1 [math.mg] 10 Apr 2018 ON THE VOLUME BOUND IN THE DVORETZKY ROGERS LEMMA FERENC FODOR, MÁRTON NASZÓDI, AND TAMÁS ZARNÓCZ arxiv:1804.03444v1 [math.mg] 10 Apr 2018 Abstract. The classical Dvoretzky Rogers lemma provies a eterministic

More information

Agmon Kolmogorov Inequalities on l 2 (Z d )

Agmon Kolmogorov Inequalities on l 2 (Z d ) Journal of Mathematics Research; Vol. 6, No. ; 04 ISSN 96-9795 E-ISSN 96-9809 Publishe by Canaian Center of Science an Eucation Agmon Kolmogorov Inequalities on l (Z ) Arman Sahovic Mathematics Department,

More information

Math 342 Partial Differential Equations «Viktor Grigoryan

Math 342 Partial Differential Equations «Viktor Grigoryan Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite

More information

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d A new proof of the sharpness of the phase transition for Bernoulli percolation on Z Hugo Duminil-Copin an Vincent Tassion October 8, 205 Abstract We provie a new proof of the sharpness of the phase transition

More information

On colour-blind distinguishing colour pallets in regular graphs

On colour-blind distinguishing colour pallets in regular graphs J Comb Optim (2014 28:348 357 DOI 10.1007/s10878-012-9556-x On colour-blin istinguishing colour pallets in regular graphs Jakub Przybyło Publishe online: 25 October 2012 The Author(s 2012. This article

More information

McMaster University. Advanced Optimization Laboratory. Title: The Central Path Visits all the Vertices of the Klee-Minty Cube.

McMaster University. Advanced Optimization Laboratory. Title: The Central Path Visits all the Vertices of the Klee-Minty Cube. McMaster University Avance Optimization Laboratory Title: The Central Path Visits all the Vertices of the Klee-Minty Cube Authors: Antoine Deza, Eissa Nematollahi, Reza Peyghami an Tamás Terlaky AvOl-Report

More information

Rank, Trace, Determinant, Transpose an Inverse of a Matrix Let A be an n n square matrix: A = a11 a1 a1n a1 a an a n1 a n a nn nn where is the jth col

Rank, Trace, Determinant, Transpose an Inverse of a Matrix Let A be an n n square matrix: A = a11 a1 a1n a1 a an a n1 a n a nn nn where is the jth col Review of Linear Algebra { E18 Hanout Vectors an Their Inner Proucts Let X an Y be two vectors: an Their inner prouct is ene as X =[x1; ;x n ] T Y =[y1; ;y n ] T (X; Y ) = X T Y = x k y k k=1 where T an

More information

Topic 7: Convergence of Random Variables

Topic 7: Convergence of Random Variables Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information

More information

Qubit channels that achieve capacity with two states

Qubit channels that achieve capacity with two states Qubit channels that achieve capacity with two states Dominic W. Berry Department of Physics, The University of Queenslan, Brisbane, Queenslan 4072, Australia Receive 22 December 2004; publishe 22 March

More information

Witten s Proof of Morse Inequalities

Witten s Proof of Morse Inequalities Witten s Proof of Morse Inequalities by Igor Prokhorenkov Let M be a smooth, compact, oriente manifol with imension n. A Morse function is a smooth function f : M R such that all of its critical points

More information

Iterated Point-Line Configurations Grow Doubly-Exponentially

Iterated Point-Line Configurations Grow Doubly-Exponentially Iterate Point-Line Configurations Grow Doubly-Exponentially Joshua Cooper an Mark Walters July 9, 008 Abstract Begin with a set of four points in the real plane in general position. A to this collection

More information

Ramsey numbers of some bipartite graphs versus complete graphs

Ramsey numbers of some bipartite graphs versus complete graphs Ramsey numbers of some bipartite graphs versus complete graphs Tao Jiang, Michael Salerno Miami University, Oxfor, OH 45056, USA Abstract. The Ramsey number r(h, K n ) is the smallest positive integer

More information

Center of Gravity and Center of Mass

Center of Gravity and Center of Mass Center of Gravity an Center of Mass 1 Introuction. Center of mass an center of gravity closely parallel each other: they both work the same way. Center of mass is the more important, but center of gravity

More information

Calculus and optimization

Calculus and optimization Calculus an optimization These notes essentially correspon to mathematical appenix 2 in the text. 1 Functions of a single variable Now that we have e ne functions we turn our attention to calculus. A function

More information

Euler equations for multiple integrals

Euler equations for multiple integrals Euler equations for multiple integrals January 22, 2013 Contents 1 Reminer of multivariable calculus 2 1.1 Vector ifferentiation......................... 2 1.2 Matrix ifferentiation........................

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

The Exact Form and General Integrating Factors

The Exact Form and General Integrating Factors 7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily

More information

Sharp Thresholds. Zachary Hamaker. March 15, 2010

Sharp Thresholds. Zachary Hamaker. March 15, 2010 Sharp Threshols Zachary Hamaker March 15, 2010 Abstract The Kolmogorov Zero-One law states that for tail events on infinite-imensional probability spaces, the probability must be either zero or one. Behavior

More information

Final Exam Study Guide and Practice Problems Solutions

Final Exam Study Guide and Practice Problems Solutions Final Exam Stuy Guie an Practice Problems Solutions Note: These problems are just some of the types of problems that might appear on the exam. However, to fully prepare for the exam, in aition to making

More information

LECTURE NOTES ON DVORETZKY S THEOREM

LECTURE NOTES ON DVORETZKY S THEOREM LECTURE NOTES ON DVORETZKY S THEOREM STEVEN HEILMAN Abstract. We present the first half of the paper [S]. In particular, the results below, unless otherwise state, shoul be attribute to G. Schechtman.

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y Ph195a lecture notes, 1/3/01 Density operators for spin- 1 ensembles So far in our iscussion of spin- 1 systems, we have restricte our attention to the case of pure states an Hamiltonian evolution. Toay

More information

REAL ANALYSIS I HOMEWORK 5

REAL ANALYSIS I HOMEWORK 5 REAL ANALYSIS I HOMEWORK 5 CİHAN BAHRAN The questions are from Stein an Shakarchi s text, Chapter 3. 1. Suppose ϕ is an integrable function on R with R ϕ(x)x = 1. Let K δ(x) = δ ϕ(x/δ), δ > 0. (a) Prove

More information

Solutions to Math 41 Second Exam November 4, 2010

Solutions to Math 41 Second Exam November 4, 2010 Solutions to Math 41 Secon Exam November 4, 2010 1. (13 points) Differentiate, using the metho of your choice. (a) p(t) = ln(sec t + tan t) + log 2 (2 + t) (4 points) Using the rule for the erivative of

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

A FURTHER REFINEMENT OF MORDELL S BOUND ON EXPONENTIAL SUMS

A FURTHER REFINEMENT OF MORDELL S BOUND ON EXPONENTIAL SUMS A FURTHER REFINEMENT OF MORDELL S BOUND ON EXPONENTIAL SUMS TODD COCHRANE, JEREMY COFFELT, AND CHRISTOPHER PINNER 1. Introuction For a prime p, integer Laurent polynomial (1.1) f(x) = a 1 x k 1 + + a r

More information

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Ashish Goel Michael Kapralov Sanjeev Khanna Abstract We consier the well-stuie problem of fining a perfect matching in -regular bipartite

More information

On the Expansion of Group based Lifts

On the Expansion of Group based Lifts On the Expansion of Group base Lifts Naman Agarwal, Karthekeyan Chanrasekaran, Alexanra Kolla, Vivek Maan July 7, 015 Abstract A k-lift of an n-vertex base graph G is a graph H on n k vertices, where each

More information

Table of Common Derivatives By David Abraham

Table of Common Derivatives By David Abraham Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec

More information

Function Spaces. 1 Hilbert Spaces

Function Spaces. 1 Hilbert Spaces Function Spaces A function space is a set of functions F that has some structure. Often a nonparametric regression function or classifier is chosen to lie in some function space, where the assume structure

More information

Algorithms and matching lower bounds for approximately-convex optimization

Algorithms and matching lower bounds for approximately-convex optimization Algorithms an matching lower bouns for approximately-convex optimization Yuanzhi Li Department of Computer Science Princeton University Princeton, NJ, 08450 yuanzhil@cs.princeton.eu Anrej Risteski Department

More information

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS MARK SCHACHNER Abstract. When consiere as an algebraic space, the set of arithmetic functions equippe with the operations of pointwise aition an

More information

Linear and quadratic approximation

Linear and quadratic approximation Linear an quaratic approximation November 11, 2013 Definition: Suppose f is a function that is ifferentiable on an interval I containing the point a. The linear approximation to f at a is the linear function

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

MA 2232 Lecture 08 - Review of Log and Exponential Functions and Exponential Growth

MA 2232 Lecture 08 - Review of Log and Exponential Functions and Exponential Growth MA 2232 Lecture 08 - Review of Log an Exponential Functions an Exponential Growth Friay, February 2, 2018. Objectives: Review log an exponential functions, their erivative an integration formulas. Exponential

More information

Lecture 22. Lecturer: Michel X. Goemans Scribe: Alantha Newman (2004), Ankur Moitra (2009)

Lecture 22. Lecturer: Michel X. Goemans Scribe: Alantha Newman (2004), Ankur Moitra (2009) 8.438 Avance Combinatorial Optimization Lecture Lecturer: Michel X. Goemans Scribe: Alantha Newman (004), Ankur Moitra (009) MultiFlows an Disjoint Paths Here we will survey a number of variants of isjoint

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

Review of Differentiation and Integration for Ordinary Differential Equations

Review of Differentiation and Integration for Ordinary Differential Equations Schreyer Fall 208 Review of Differentiation an Integration for Orinary Differential Equations In this course you will be expecte to be able to ifferentiate an integrate quickly an accurately. Many stuents

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a

More information

Modelling and simulation of dependence structures in nonlife insurance with Bernstein copulas

Modelling and simulation of dependence structures in nonlife insurance with Bernstein copulas Moelling an simulation of epenence structures in nonlife insurance with Bernstein copulas Prof. Dr. Dietmar Pfeifer Dept. of Mathematics, University of Olenburg an AON Benfiel, Hamburg Dr. Doreen Straßburger

More information

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback Journal of Machine Learning Research 8 07) - Submitte /6; Publishe 5/7 An Optimal Algorithm for Banit an Zero-Orer Convex Optimization with wo-point Feeback Oha Shamir Department of Computer Science an

More information

Expected Value of Partial Perfect Information

Expected Value of Partial Perfect Information Expecte Value of Partial Perfect Information Mike Giles 1, Takashi Goa 2, Howar Thom 3 Wei Fang 1, Zhenru Wang 1 1 Mathematical Institute, University of Oxfor 2 School of Engineering, University of Tokyo

More information

arxiv: v2 [cs.ds] 11 May 2016

arxiv: v2 [cs.ds] 11 May 2016 Optimizing Star-Convex Functions Jasper C.H. Lee Paul Valiant arxiv:5.04466v2 [cs.ds] May 206 Department of Computer Science Brown University {jasperchlee,paul_valiant}@brown.eu May 3, 206 Abstract We

More information

The chromatic number of graph powers

The chromatic number of graph powers Combinatorics, Probability an Computing (19XX) 00, 000 000. c 19XX Cambrige University Press Printe in the Unite Kingom The chromatic number of graph powers N O G A A L O N 1 an B O J A N M O H A R 1 Department

More information

Logarithmic spurious regressions

Logarithmic spurious regressions Logarithmic spurious regressions Robert M. e Jong Michigan State University February 5, 22 Abstract Spurious regressions, i.e. regressions in which an integrate process is regresse on another integrate

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

Analysis IV, Assignment 4

Analysis IV, Assignment 4 Analysis IV, Assignment 4 Prof. John Toth Winter 23 Exercise Let f C () an perioic with f(x+2) f(x). Let a n f(t)e int t an (S N f)(x) N n N then f(x ) lim (S Nf)(x ). N a n e inx. If f is continuously

More information

Lecture XII. where Φ is called the potential function. Let us introduce spherical coordinates defined through the relations

Lecture XII. where Φ is called the potential function. Let us introduce spherical coordinates defined through the relations Lecture XII Abstract We introuce the Laplace equation in spherical coorinates an apply the metho of separation of variables to solve it. This will generate three linear orinary secon orer ifferential equations:

More information

Math 1B, lecture 8: Integration by parts

Math 1B, lecture 8: Integration by parts Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores

More information

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides Reference 1: Transformations of Graphs an En Behavior of Polynomial Graphs Transformations of graphs aitive constant constant on the outsie g(x) = + c Make graph of g by aing c to the y-values on the graph

More information

Resistant Polynomials and Stronger Lower Bounds for Depth-Three Arithmetical Formulas

Resistant Polynomials and Stronger Lower Bounds for Depth-Three Arithmetical Formulas Resistant Polynomials an Stronger Lower Bouns for Depth-Three Arithmetical Formulas Maurice J. Jansen University at Buffalo Kenneth W.Regan University at Buffalo Abstract We erive quaratic lower bouns

More information

Problem Sheet 2: Eigenvalues and eigenvectors and their use in solving linear ODEs

Problem Sheet 2: Eigenvalues and eigenvectors and their use in solving linear ODEs Problem Sheet 2: Eigenvalues an eigenvectors an their use in solving linear ODEs If you fin any typos/errors in this problem sheet please email jk28@icacuk The material in this problem sheet is not examinable

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Many problems in physics, engineering, and chemistry fall in a general class of equations of the form. d dx. d dx

Many problems in physics, engineering, and chemistry fall in a general class of equations of the form. d dx. d dx Math 53 Notes on turm-liouville equations Many problems in physics, engineering, an chemistry fall in a general class of equations of the form w(x)p(x) u ] + (q(x) λ) u = w(x) on an interval a, b], plus

More information

DIFFERENTIAL GEOMETRY OF CURVES AND SURFACES 7. Geodesics and the Theorem of Gauss-Bonnet

DIFFERENTIAL GEOMETRY OF CURVES AND SURFACES 7. Geodesics and the Theorem of Gauss-Bonnet A P Q O B DIFFERENTIAL GEOMETRY OF CURVES AND SURFACES 7. Geoesics an the Theorem of Gauss-Bonnet 7.. Geoesics on a Surface. The goal of this section is to give an answer to the following question. Question.

More information

Counting Lattice Points in Polytopes: The Ehrhart Theory

Counting Lattice Points in Polytopes: The Ehrhart Theory 3 Counting Lattice Points in Polytopes: The Ehrhart Theory Ubi materia, ibi geometria. Johannes Kepler (1571 1630) Given the profusion of examples that gave rise to the polynomial behavior of the integer-point

More information

Diagonalization of Matrices Dr. E. Jacobs

Diagonalization of Matrices Dr. E. Jacobs Diagonalization of Matrices Dr. E. Jacobs One of the very interesting lessons in this course is how certain algebraic techniques can be use to solve ifferential equations. The purpose of these notes is

More information

Lecture 2: Correlated Topic Model

Lecture 2: Correlated Topic Model Probabilistic Moels for Unsupervise Learning Spring 203 Lecture 2: Correlate Topic Moel Inference for Correlate Topic Moel Yuan Yuan First of all, let us make some claims about the parameters an variables

More information

Chapter 4. Electrostatics of Macroscopic Media

Chapter 4. Electrostatics of Macroscopic Media Chapter 4. Electrostatics of Macroscopic Meia 4.1 Multipole Expansion Approximate potentials at large istances 3 x' x' (x') x x' x x Fig 4.1 We consier the potential in the far-fiel region (see Fig. 4.1

More information

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys Homewor Solutions EM, Mixture Moels, PCA, Dualitys CMU 0-75: Machine Learning Fall 05 http://www.cs.cmu.eu/~bapoczos/classes/ml075_05fall/ OUT: Oct 5, 05 DUE: Oct 9, 05, 0:0 AM An EM algorithm for a Mixture

More information

Chapter 1. Preliminaries

Chapter 1. Preliminaries Introduction This dissertation is a reading of chapter 4 in part I of the book : Integer and Combinatorial Optimization by George L. Nemhauser & Laurence A. Wolsey. The chapter elaborates links between

More information

Applications of the Wronskian to ordinary linear differential equations

Applications of the Wronskian to ordinary linear differential equations Physics 116C Fall 2011 Applications of the Wronskian to orinary linear ifferential equations Consier a of n continuous functions y i (x) [i = 1,2,3,...,n], each of which is ifferentiable at least n times.

More information

Sturm-Liouville Theory

Sturm-Liouville Theory LECTURE 5 Sturm-Liouville Theory In the three preceing lectures I emonstrate the utility of Fourier series in solving PDE/BVPs. As we ll now see, Fourier series are just the tip of the iceberg of the theory

More information

QF101: Quantitative Finance September 5, Week 3: Derivatives. Facilitator: Christopher Ting AY 2017/2018. f ( x + ) f(x) f(x) = lim

QF101: Quantitative Finance September 5, Week 3: Derivatives. Facilitator: Christopher Ting AY 2017/2018. f ( x + ) f(x) f(x) = lim QF101: Quantitative Finance September 5, 2017 Week 3: Derivatives Facilitator: Christopher Ting AY 2017/2018 I recoil with ismay an horror at this lamentable plague of functions which o not have erivatives.

More information