Robust polynomial regression up to the information theoretic limit

Size: px
Start display at page:

Download "Robust polynomial regression up to the information theoretic limit"

Transcription

1 58th Annual IEEE Syposiu on Foundations of Coputer Science Robust polynoial regression up to the inforation theoretic liit Daniel Kane Departent of Coputer Science and Engineering / Departent of Matheatics University of California, San Diego La Jolla, CA, USA dakane@ucsd.edu Sushrut Karalkar, Eric Price Departent of Coputer Science The University of Texas at Austin Austin, TX, USA {sushrutk, ecprice}@cs.utexas.edu Abstract We consider the proble of robust polynoial regression, where one receives saples that are usually within a sall additive error of a target polynoial, but have a chance of being arbitrary adversarial outliers. Previously, it was known how to efficiently estiate the target polynoial only when the outlier probability was subconstant in the degree of the target polynoial. We give an algorith that works for the entire feasible range of outlier probabilities, while siultaneously iproving other paraeters of the proble. We copleent our algorith, which gives a factor approxiation, with ipossibility results that show, for exaple, that a 1.09 approxiation is ipossible even with infinitely any saples. Keywords-polynoial regression, robust regression, robust recovery, learning, approxiation; I. INTRODUCTION Polynoial regression is the proble of finding a polynoial that passes near a collection of input points. The proble has been studied for 00 years with diverse applications [1], including coputer graphics [], achine learning [3], and statistics [4]. As with linear regression, the classic solution to polynoial regression is least squares, but this is not robust to outliers: a single outlier point can perturb the least squares solution by an arbitrary aount. Hence finding a robust ethod of polynoial regression is an iportant question. A version of this proble was foralized by Arora and Khot [5]. We want to learn a degree d polynoial p : [ 1, 1] R, and we observe (x i,y i ) where x i is drawn fro soe easure χ (say, unifor) and y i = p(x i )+w i for soe noise w i. Each saple is an outlier independently with probability at ost ρ; if the saple is not an outlier, then w i σ for soe σ. Other than the rando choice of outlier locations, the noise is adversarial. One would like to recover p with p p Cσ using as few saples as possible, with high probability. In other contexts an l requireent on the input noise would be a significant restriction, but outlier tolerance akes it uch less so. For exaple, independent bounded-variance noise fits in this fraework: by Chebyshev s inequality, the fraework applies with outlier chance ρ = and σ =10E[w i ]1/, which gives the ideal result up to constant factors. At the sae tie, the l requireent on the input allows for the strong l bound on the output. Arora and Khot [5] showed that ρ < 1/ is inforation-theoretically necessary for non-trivial recovery, and showed how to solve the proble using a linear progra when ρ = 0. For ρ > 0, the RANSAC [6] technique of fitting to a subsaple of the data works in polynoial tie when ρ log d d. Finding an efficient algorith for ρ log d d reained open until recently. Last year, Guruswai and Zuckeran [7] showed how to solve the proble in polynoial tie for ρ larger than log d d. Unfortunately, their result falls short of the ideal in several ways: it needs a low outlier rate (ρ < 1 p log d ), a bounded signal-to-noise ratio ( σ < d O(1) ), and has ( a super-constant ) approxiation factor 0.01). ( p p σ 1+ p σ It uses O(d log c d) saples fro the unifor easure, or O(d log c d) saples fro the Chebyshev easure. 1 x 1 These deficiencies ean that the algorith doesn t work when all the noise coes fro outliers (σ =0); the low outlier rate required also leads to a superconstant factor loss when reducing fro other noise odels. In this work, we give a siple algorith that avoids all these deficiencies: it works for all ρ<1/; it has no restrictions on σ; and it gets a constant approxiation factor C = + ɛ. Additionally, it only uses O(d ) saples fro the unifor easure, without any log factors, and O(d log d) saples fro the Chebyshev easure. We also give lower bounds for the approxiation factor C, indicating that one cannot achieve a 1+ɛ approxiation by showing that C > Our lower bounds are not the best possible, and it is possible that the true lower bounds are uch better. A. Algorithic results The proble forulation has randoly placed outliers, which is needed to ensure that we can estiate /17 $ IEEE DOI /FOCS

2 the polynoial locally around any point x. We use the following definition to encapsulate this requireent, after which the noise can be fully adversarial: Definition I.1. The size Chebyshev partition of [ 1, 1] is the set of intervals I j =[cos πj π(j 1), cos ] for j []. We say that a set S of saples (x i,y i ) is α-good for the partition if, in every interval I j, strictly less than an α fraction of the points x i I j are outliers. The size Chebyshev partition is the set of intervals between consecutive extrea of the Chebyshev polynoial of degree. Standard approxiation theory recoends sapling functions at the roots of the Chebyshev polynoial, so intuitively a good set of saples will allow fairly accurate estiation near each of these ideal sapling points. All our algoriths will work for any set of α- good saples, for α< 1 and sufficiently large. The saple coplexities are then what is required for S to be good with high probability. We first describe two siple algoriths that do not quite achieve the goal. We then describe how to blackbox cobine their results to get the full result. L1 regression.: Our first result is that l 1 regression alost solves the proble: it satisfies all the requireents, except that the resulting error p p is bounded in l 1 not l : Lea I.. Suppose the set S of saples is α-good for the size = O(d) Chebyshev partition, for constant α< 1. Then the result p of l 1 regression arg in degree-d n i=1 polynoial p satisfies p p 1 O α (1) σ. ean x i I j y i p(x i ) This has a nice parallel to the d =1case, where l 1 regression is the canonical robust estiator. As shown by the Chebyshev polynoial in Figure 1, l 1 regression ight not give a good solution to the l proble. However, converting to an l bound loses at ost an O(d ) factor. This eans this lea is already useful for an l bound: one of the requireents of [7] was that p σd O(1). We can avoid this by first coputing the l 1 regression estiate p (l1) and then applying [7] to the residual p p (l1), which always satisfies the condition. Median-based recovery.: Our second result is for a edian-based approach to the proble: take the edian ỹ j of the y values within each interval I j. Because the edian is robust, this will lie within the range of inlier y values for that interval. We don t know which x I j it corresponds to, but this turns out not to atter too y l 1 regression ay yield high l error Observation Scaled Chebyshev T 8 Zero polynoial x Figure 1: The blue observations are within σ =1of the red zero polynoial at each point, but are closer in l 1 to the green Chebyshev polynoial. This deonstrates that l 1 regression does not solve the l proble even for ρ =0. uch: assigning each ỹ j to an arbitrary x j I j and applying non-robust l regression gives a useful result. Lea I.3. Let ɛ, α < 1, and suppose the set S of saples is α-good for the size = O(d/ɛ) Chebyshev partition. Let x j I j be chosen arbitrarily, and let ỹ j =edian xi I j y i. Then the degree d polynoial p iniizing ax j [] p( x j) ỹ j satisfies p p ( + ɛ)σ + ɛ p. Without the additive ɛ p ter, which coes fro not knowing the x I j for ỹ j, this would be exactly what we want. If σ p, however, this is not so good but it still akes progress. Iterating the algoriths.: While neither of the above results is sufficient on its own, siply applying the iteratively on the residual left by previous rounds gives our full theore. The result of l 1 regression is a p (0) with p p (0) = O(d σ). Applying the edian recovery algorith to the residual points (x i,y i p (0) (x i )), we will get p so that p (1) = p (0) + p has p (1) p ( + ɛ)σ + ɛ O(d σ) If we continue applying the edian recovery algorith to the reaining residual, the l nor of the residual will continue to decrease exponentially. After r = O(log 1/ɛ d) rounds we will reach p (r) p ( + 4ɛ)σ 39

3 as desired 1. Since each ethod can be coputed efficiently with a linear progra, this gives our ain theore: Theore I.4. Let ɛ, α < 1, and suppose the set S of saples is α-good for the size = O(d/ɛ) Chebyshev partition. The result p of Algorith is a degree d polynoial satisfying p p ( + ɛ)σ. Its running tie is that of solving O(log 1/ɛ d) linear progras, each with O(d) variables and O( S ) constraints. We now apply this to the robust polynoial regression proble, where we receive points (x i,y i ) such that each x i is drawn fro soe distribution, and with probability ρ it is an outlier. An adversary then picks the y i such that, for each non-outlier i, y i p(x i ) σ. We observe the following: Corollary I.5. Consider the robust polynoial regression proble for constant outlier chance ρ < 1/, with points x i drawn fro the Chebyshev distribution 1 D c (x). Then O( d 1 x ɛ log d δɛ ) saples suffice to recover with probability 1 δ adegreed polynoial p with ax p(x) p(x) ( + ɛ)σ. x [ 1,1] If x i is drawn fro the unifor distribution instead, then O( d ɛ log 1 δ ) saples suffice for the sae result. In both cases, the recovery tie is polynoial in the saple size. B. Ipossibility results We show liitations on iproving any of the three paraeters used in Corollary I.5: the saple coplexity, the approxiation factor, and the requireent that ρ< 1/. Saple Coplexity.: Our result uses O(d ) saples fro the unifor distribution and O(d log d) saples fro the Chebyshev distribution. We show in Section V-A that both results are tight. In particular, we show that it is ipossible to get any constant approxiation with o(d ) saples fro the unifor distribution or o(d log d) saples fro the Chebyshev distribution. Approxiation factor.: Our approxiation factor is +ɛ. Even with infinitely any saples and no outliers, can one do better than a -approxiation for l regression? For coparison, in l regression a 1+ɛ approxiation is possible. For these lower bounds, it is convenient to consider the special case where the values y i are y(x i ) for a function y with p y σ. We show two lower bounds related to this question. 1 We reark that one could skip the initial l 1 iniization step, but the nuber of rounds would then depend on log( p ), which σ could be unbounded. Note that this only affects running tie and not the saple coplexity. Best fitting quadratic Figure : The lower bound for 1.09-approxiations via any recovery algorith. We give three quadratics that all lie within σ of each other for all x, and hence the observed y(x) can be identical regardless of which quadratic is p. The region containing points within 1.09σ of all three options just barely doesn t contain a quadratic itself. First, we show that l projection that is, the algorith that iniizes y p over degree-d polynoials p can have error arbitrarily close to σ. Since our algorith is an outlier-tolerant version of l projection, we should not expect to perfor better. Second, we show that no proper learning algorith can achieve a 1+ɛ guarantee. In particular, we show for d =that any algorith with ore than /3 success probability ust have C > 1.09; this is illustrated in Figure. For general d, we show that C>1+Ω(1/d 3 ). List decoding.: The ρ < 1/ requireent is obviously required for uniquely decoding p, since for ρ 1/ the observations could coe fro a ixture of two copletely different polynoials. But one could hope for a list decoding version of the result, outputting a sall set of polynoials such that one is close to the true output. Unfortunately, [5] showed that for a C- approxiate algorith, such a set would require size Ω( d/c) at least e event for ρ 1/. We iprove this lower bound to e Ω(d/C). At least for constant C and ρ, our result is tight: Theore I.4 iplies that if no saples were outliers, then = O(d) saples would suffice. Hence picking saples will result in Algorith outputting a polynoial consistent with the data with (1 ρ) = e O(d) probability. Repeating this would give a set of size e O(d) that works with high probability. 393

4 C. Related Work In addition to the work of [5], [7] discussed above, several papers have looked at siilar probles. When σ =0, the proble becoes the standard one of Reed- Soloon decoding with relative Haing distance ρ. Efficient algoriths such as Berlekap-Massey [8] give unique decoding for all ρ<1/. For ρ>1/, while unique decoding is ipossible, a polynoial size list decoding is possible with sufficiently any saples for all ρ<1 [9]. Other work has looked at robust estiation for distributions. In the field of robust statistics (see, e.g., [10]), as well as soe recent papers in theoretical coputer science (e.g. [11] [13]), one would like to estiate properties of a distribution fro saples that have a ρ chance of being adversarial. In soe such cases, list decoding for ρ>1/ is possible [13]. II. PRELIMINARIES For a function f :[ 1, 1] R and interval I [ 1, 1], we define the l q nor on the interval to be ( ) 1/q f I,q := f(x) q dx, I k where f I, := sup x I f(x). We also define the overall l q nor f q := f [ 1,1],q. We will need the following consequence of a generalization due to Nevai [14] of Bernstein s inequality fro l to l q. The proof of this lea is in Appendix A. Lea II.1. Let p be a degree d polynoial. Let I 1,...,I partition [ 1, 1] between the Chebyshev extrea cos πj, for soe d. Let r :[ 1, 1] R be piecewise constant, so that for each I k there exists an x k I k with r(x) =p(x k ) for all x I k. Then there exists a universal constant C such that, for any q 1, p r q C d p q. We recall the definition of the Chebyshev polynoials T d (x), which we will use extensively. Definition II.. The Chebyshev polynoials of the first kind T d (x) are defined by the following recurrence: T 0 (x) = 1,T 1 (x) = x and T n+1 (x) = xt n (x) T n 1 (x). They have the property that T d (cos θ) = cos(dθ) for all θ. III. l 1 REGRESSION Lea III.1. Suppose C d ɛ constant C. Then, for any set of saples x 1,...,x n with all S i = {j x j I i } nonepty, and any polynoial p of degree d, i for a large enough I i p(x j ) =(1± ɛ) p 1. S i j S i Proof: When all S i = 1, this is a restateent of Lea II.1 for q = 1. Otherwise, the LHS is the expectation of the S i =1case, when randoly drawing a single j in each S i. We will show a ore precise stateent than Lea I., which allows for a weaker l 1 version of the α-good requireent on the saples: Definition III.. For a set of saples (x 1,y 1 ),...,(x n,y n ), and the Chebyshev partition I 1,...,I, define S j = {i x i I j }. We say that the saples are (α, σ) close to a polynoial p in l 1 if, for we have e j := in ax p(x j) y j, S S j, S (1 α) S j j S j [] e j σ. If the saples are α-good, then each e j σ, so Ij e j ( )(ax e j ) σ and the saples are (α, σ) close to p in l 1. We now state Lea III.3 which is a ore precise stateent of Lea I.. Lea III.3. Let α<1/, and C d ɛ for a large enough constant C and soe ɛ (1 α)/4. Then, given any saples (α, σ) close to p in l 1, the degree d polynoial solution to the l 1 regression proble p =argin q j [] y i q(x i ) i S j has p p 1 σ 1 α. Proof: In each interval, we consider the α coordinates axiizing y i p(x i ) to be bad, and the rest to be good. Let the set of bad and good coordinates in S j be denoted by B j and G j respectively, and define obj(f) = I j j [] S j i S j y i f(x i ). By definition obj( p) obj(p). This gives us 0 obj( p) obj(p) = y i p(x i ) y i p(x i ) j [] i Sj i S j = y i p(x i ) y i p(x i ) j [] i G j + y i p(x i ) y i p(x i ). i B j j [] Fro the triangle inequality, we have y i p(x i ) p(x i ) p(x i ) y i p(x i ) and y i p(x i ) y i 394

5 p(x i ) p(x i ) p(x i ). This gives 0 j [] j [] ( p(x i ) p(x i ) y i p(x i ) ) i G j p(x i ) p(x i ) i B j Now, recall that our saples are (α, δ) close to p in l 1. This eans (by Definition III.) that for any good i, y i p(x i ) e j for a set of e j s that satisfy e j σ. Therefore, for these e j, 0 j [] j [] j [] j [] j [] j [] p(x i ) p(x i ) i G j p(x i ) p(x i ) i B j y i p(x i ) i G j p(x i ) p(x i ) i G j p(x i ) p(x i ) i B j ((1 α) S j e j ) Using Lea III.1 on p p we now get or Hence 0 (1 α)(1 ɛ) p p 1 α(1 + ɛ) p p 1 (1 α)σ (1 α ɛ) p p 1 (1 α)σ, p p 1 (1 α)σ 1 α ɛ. For ɛ 1 α, this gives the desired p p 1 σ 1 α. IV. l REGRESSION Given a set of α< 1 -good saples S, our goal is to find a degree p polynoial q with p p (+ɛ)σ. We start by proving Lea I.3, which we restate below for clarity: Lea I.3. Let ɛ, α < 1, and suppose the set S of saples is α-good for the size = O(d/ɛ) Chebyshev partition. Let x j I j be chosen arbitrarily, and let Algorith 1 Refineent ethod, analyzed in Lea I.3 for r =0 1: procedure REFINE(S = {(x i,y i )}, p) : ỹ j edian xi I j y i p(x i ) 3: Choose arbitrary x j I j 4: Fit degree d polynoial r iniizing r( x j ) ỹ j 5: p p + r 6: return p 7: end procedure Algorith Coplete recovery procedure 1: procedure APPROX(S) : p (0) result of l 1 regression 3: for i [0,O(log 1/ɛ d)] do 4: p (i+1) REFINE(S, p (i) ) 5: end for 6: return p (O(log 1/ɛ d)) 7: end procedure ỹ j =edian xi I j y i. Then the degree d polynoial p iniizing ax j [] p( x j) ỹ j satisfies p p ( + ɛ)σ + ɛ p. Proof: Since ore than half the points in any interval I j are such that y j p(x j ) σ, and p is continuous, there ust exist an x j I j satisfying ỹ j p(x j) σ. (1) We now define three piecewise-constant functions, r(x), r(x), and r(x), to be such that within each interval I j we have r(x) =p(x j ), r(x) = p( x j), and r(x) =ỹ j. By Lea II.1, for our choice of we have p r ɛ p and p r ɛ p. () We also have by (1) that r r σ, sobythe triangle inequality p r σ + ɛ p. (3) Now, our choice of p ensures that r r =ax j [] p( x j) ỹ j ax j [] p( x j) ỹ j p r σ + ɛ p. Cobining with () and (3) gives by the triangle inequality that p p σ +ɛ p + ɛ p. (4) 395

6 To finish the proof, we just need a bound on p. Note that (4) iplies p σ +(1+ɛ) p + ɛ p, and since ɛ 1/, this iplies p 4σ +(+4ɛ) p. Plugging into (4) and rescaling ɛ down by a constant factor gives the result. Before we prove Theore I.4, we briefly describe the algorith. Algorith first sets its initial estiate to the polynoial produced by l 1 regression. It then continues to refine the estiate it has by using Algorith 1 for O(log 1/ɛ d) iterations. Theore I.4. Let ɛ, α < 1, and suppose the set S of saples is α-good for the size = O(d/ɛ) Chebyshev partition. The result p of Algorith is a degree d polynoial satisfying p p ( + ɛ)σ. Its running tie is that of solving O(log 1/ɛ d) linear progras, each with O(d) variables and O( S ) constraints. Proof: Our algorith proceeds by iteratively iproving a polynoial approxiation p to p, so that p p iproves at each stage. Our notation will be borrowed fro the notation in the stateents of Algoriths and 1. Let p (t) (x) be the t th estiate of p found by Algorith and let e t (x) = (p p (t) )(x) be the error of the t th estiate p (t). r t is the polynoial r found by the t th iteration of Algorith 1, i.e. r t = arg in r r( x j ) ỹ j where ỹ j = edian xi I j y i p (t) (x i ), and the iniu is taken over all degree d polynoials. Lea I.3 now iplies r t (x) e t (x) ( + ɛ)σ + ɛ e t (x). Observe that r t (x) e t (x) =( p (t+1) (x) p (t) (x)) (p(x) p (t) (x)) = e t+1 (x), which gives us e t+1 (x) ( + ɛ)σ + ɛ e t (x). Proceeding by induction and using the geoetric series forula we see p p (t) +ɛ 1 ɛ σ+ɛt e 0 (+4ɛ)σ+ɛ t e 0. Rescaling ɛ, we see that in a nuber of iterations logarithic in the quality of our initial solution, we find a p such that p p ( + ɛ)σ. Finally, we analyze the quality of the initial solution produced by l 1 regression. By Lea I., e 0 1 O(σ). Applying the Markov brothers inequality to the degree d +1 polynoial Q(x) = x 1 e 0(u)du, weget e 0 = ax x [ 1,1] Q (x) (d +1) ax Q(x) x [ 1,1] (d +1) e 0 1 O(d σ). Hence the nuber of iterations required is O(log 1/ɛ d). Applying this to our rando outlier setting, we get Corollary I.5. Proof: It is enough to show that for = O( d ɛ ), drawing O( d ɛ log d δɛ ) saples fro the Chebyshev distribution or O( d ɛ log( 1 δ )) saples fro the unifor distribution gives us an α-good saple for soe α< 1 with probability 1 δ. The corollary then follows fro Theore I.4. Let k be the total nuber of saples taken, and let p j denote the probability that any sapled x i lies in the j th interval I j. With Chebyshev sapling, p j = 1 for all j. With unifor sapling, p j =Θ( in(j,+1 j) ). Let X j be the nuber of saples that appear in I j, and Y j be the nuber of these saples that are outliers. We have E[X j ]=kp j, so by a Chernoff bound, Pr[X j 1 kp j] e Ω(kpj). (5) The outliers are then chosen independently, with expectation ρx j,so Pr[Y j αx j X j ] e Ω((α ρ) X j). (6) Setting α = ρ+ 1, we have that conditioned on (5) not occurring, (6) occurs with at ost e Ω(kpj) probability. If neither occur, then less than an α fraction of the saples in I j are outliers. Hence, by a union bound over the intervals, the saples are α-good with probability at least 1 e Ω(kpj). (7) i=1 In the Chebyshev setting, we have p j = 1/ and k = O( log(/δ)), aking (7) at least 1 δ for appropriately chosen constants. In the unifor setting, we siilarly have e Ω(kpj) = j=1 e Ω( log(1/δ) in(j,1+ j) j=1 3δ, / ) aking (7) at least 1 6δ. Rescaling δ gives the result, that the saples are α-good for soe α<1/ with probability at least 1 δ. V. IMPOSSIBILITY RESULTS δ j j=1 A. Saple Coplexity Lea V.1. If the x i are sapled uniforly then it is not possible to get an O(1) approxiation in l nor to the original function in o(d ) saples and 1/4 failure probability. 396

7 1 of degree-d polynoials f S (x) = j S p b j (x) for S []. For any x [ 1, 1], let k x [] iniize b kx x. Then for any S [], f {kx} S(x) f S (x) f {kx} S(x)+α. Figure 3: p b (x) for d =50and b =0. Proof: Consider an algorith that gives a C- approxiation given s unifor saples with noise level σ =1and zero outliers, for C = O(1) and s = o(d ). We will construct two polynoials with l distance ore than C, but for which the saples have a constant chance of being indistinguishable. Define ( the ) polynoials g(x) = 0 and f(x) := T d x + α d where Td is the degree d Chebyshev polynoial of the first kind, for soe d 4 and constant α = 4 (C 1). By construction, f(x) 1 for x [ ] 1, 1 α d,but f(1) > C because for d 4 ( f(1) = T d 1+ α ) ( cosh = (d d arcosh 1+ α )) d > C. The final inequality above follows fro the fact that cosh(d arcosh(1 + x d )) 1.9(1 + x /8) at d = 4 and this function is increasing in both d and x. Since f g > C, no single answer can be a valid C- approxiate recovery of both f and g. Suppose one always observes saples of the for (x i, 0). These saples are within σ =1of both f(x) and g(x) if they lie in the region [ 1, 1 α d ]. The chance that all saples x i lie in this region is at least (1 α d ) s e sα/d, which is e o(1) > 1/. Hence there is at least a 1/ chance that the saples fro f and g are indistinguishable, so the algorith has a failure probability ore than 1/4. Our next lower bound will use the following leas, proven in Appendix A. Lea V.. Let d 1. For any point b [ 1, 1], there exists a degree d polynoial p b such that p b (b) =1, p b =1, and for all x [ 1, 1]. p b (x) d x b Lea V.3. For any d and α>0, let = d α/, and define b j = 1+ j for j []. Consider the set Lea V.4. For any distribution on sets x = (x 1,...,x s ) of s = o(d log d) saple points with independent outlier chance ρ =Ω(1), it is not possible to get a robust O(1) approxiation in l nor to the original function with 1/4 failure probability. Proof: Let = d/ 1C, and f S and k x be as in Lea V.3 for α = 1 3C. For any x, let L j := {i [s] k xi = j}. Since these sets are disjoint, there ust exist at least / different j for which L j s/ = o(log d). Let B [] contain these j. We say that agivenl j is outlier-full if all of the x i L j are outliers, which for j B happens with probability at least ρ Lj o(log d) 1/ d. Hence the probability that at least one L j is outlier-full is at least 1 (1 1/ d) B 1 e /( d) > Suppose the true polynoial p to be learned is f S for a uniforly rando S, and consider the following adversary. If at least one L j is outlier-full, she arbitrarily picks one such j and sets S to the syetric difference of S and {j }. She then flips a coin, and with 50% probability outputs (x i,f S (x i )) for each i, and otherwise outputs (x i,f S (x i )) for each i. This is valid for σ = 1 3C, because for each i [s], either k xi = j (in which case x i L j is an outlier) or {k xi } S = {k xi } S (in which case Lea V.3 iplies f S (x i ) f S (x i ) α = 1 3C ). Because the distribution on j is independent of S, the distribution of S is also unifor, so the algorith cannot distinguish whether it received f S or f S. But f S f S = f {i} =1> Cσ, so the algorith s output cannot satisfy both cases siultaneously. Hence, in the 99% of cases where one L j is outlier-full, the algorith will have 50% failure probability, for 49.5% > 1/4 overall. B. Approxiation Factor Any algorith that relies on the result of l projection cannot do significantly better. There are two functions p(x) and f(x) such that p(x) f(x) σ for all x [ 1, 1], however the l projection to the space of all degree 1 polynoials is alost σ away in l nor fro p. 397

8 σ f(x) p 1 (x)+1 q(x) p(x) p 1 (x) 1 (1 α) 1 ax + c σ(1 α) Figure 4: q(x) akes an error of ( α)σ with p(x). Lea V.5. Let d = 1, p(x) = σ be a constant function and f(x) = σ α ax (0,x (1 α)) where α< 1. Note that p(x) f(x) σ for all x [ 1, 1]. The result q =argin r f r of l projection of f to the space of degree-1 polynoials satisfies p(x) q(x) [ 1,1], ( α)σ. Proof: Observe that f(x) =0for x [ 1, (1 α)] and f(x) = ( σ α (x 1) + σ) in [1 α, 1]. The axiu difference between two linear equations in a closed interval is attained at the endpoints of that interval. This tells us q(x) =arginax{ r( 1), r(1 α), σ r(1) }. r Let q(x) =ax + b. q(x) will be such that f( 1) > q( 1), f(1 α) <q(1 α) and f(1) >q(1) (see Figure 4), and so we want arg in ax{a b, a + b αa, σ (a + b)}. (a,b) This function is iniized when all three ters are equal, which happens when a = σ and b = ασ. This gives q(x) =σx ασ, and q(x) p(x) [ 1,1], = ( α)σ. We now show that that one cannot hope for a proper (1+ɛ)-approxiate algorith, even with no outliers. We will present a set of polynoials such that no two are ore than -apart, but for which no single polynoial lies within α>1 of all polynoials in the set. Then an adversary with σ =1can output a function y(x) independent of the choice of polynoial in the set, forcing the algorith to by α-far when recovering soe polynoial in the set. Lea V.6. There exist three degree polynoials, all within of each other over [ 1, 1], such that any single quadratic function has l distance ore than 1.09 fro one of the three. p 3 (x) 1 v v 1 Figure 5: The quadratic that passes through the four points above does not satisfy the inequality ax + c p 1 (x)+1 in the range [ 1, v], for v = Proof: Consider the polynoials p 1 (x) =(x +1), p (x) =(1 x), p 3 (x) = 3+ (1 x ), and let v 1 denote 3+ (see Figure 5). Observe that p 1 and p are at distance fro each other at x =1, 1. Also p 1 is at distance fro p 3 at x = v, and siilarly p is at distance fro p 3 at x = v. Hence, any polynoial that wants to 1-approxiate all the p i s necessarily has to go through the points (1, 1), ( 1, 1), (v, v), ( v, v). By syetry, we know that any quadratic that goes through these will be of the for y = ax + c. Substituting these values in the equation and solving for a and c we see that a = 1 1 v+1 and c =1 v+1. Since the quadratic ax + c has to 1-approxiate the p i it ust be the case that ax + c p 1 (x)+1 = x + at all points. However this inequality is not satisfied in the interval ( 1, v) as shown in Figure 5. This is because p 1 (x)+1 is the line between two points on the curve ax + c, which is a concave function for a<0. Running a progra to optiize paraeters, we see that the best approxiation to these polynoials will ake an error of at least 1.09 with one of the polynoials (see Figure ). Lea V.7. There exist O(d) degree polynoials such that any degree d polynoial that tries to approxiate all these polynoials has to ake an error of Ω( 1 d 3 ) with at least one of the degree polynoials. Proof: We will consider a set of quadratic equations such that any two are at ost distance σ fro each other, and such that any polynoial that has to σ-approxiate the needs to perfor d oscillations. As no degree d polynoial can perfor ore than d oscillations, there is at least one oscillation that this polynoial does not perfor, and hence the approxi- 398

9 Figure 6: Construction for Lea V.7. On the left, any 1-approxiator necessarily has to go through the three points. On the right, placing translated copies of these functions force any 1-approxiation to perfor ore than d oscillations. Figure 7: One eleent of f S fro Lea V.3. ating polynoial will ake an error of at least 1+h where h is the height of the oscillation. Observe that (1 Ax ) A(x ± c) =1 Ac and this is achieved at x = ± c. This eans the set S 0 = {1 Ax,A(x c),a(x + c),ax + Ac } has every polynoial at ost at distance σ := 1 Ac fro each other. Any σ approxiator, hence, necessarily has to go through the three points as shown in Figure 6. Define S t = {1 A(x ct),a(x ct c),a(x ct + c),a(x ct) + Ac } Now observe that any σ approxiator to S = t [ 1.5 d,1.5 d] S t will necessarily have to perfor d oscillations. We will now define the paraeters such that these d oscillations take place in the range [ 1, 1] and every pair of functions is at distance at ost 1 Ac fro each other. Set A = 1 d and c = 1 4d. This ensures that every pair of eleents in S t are at ost at distance d fro each other. 3 We now show that if p S t and q S t for t<t, then p q d. Observe that it is enough 3 to check this for p =1 A(x ct) and q = A(x ct + c). Because of our choice of A, c 1 A(x ct) A(x ct + c) ( ( = 1 1 x t ) ( ) ) + x t 1 d 4d 4d 1 1 ( ) d 4 ((t t) 1) 16d d 3 Finally, observe that we force the approxiating polnoial to perfor one oscillation for every S t, and if c = 1 4d the polynoial has to perfor 8d 3 oscillations to σ-approxiate every polynoial in S in the interval [ 1, 1] because it has to perfor one oscillation in every interval of length 3c. Since no degree d polynoial can perfor d oscillations, there is at least one oscillation that it cannot perfor, and so the approxiating polynoial necessarily has to ake an error of at least 1+h with one of the polynoials in S, where h is the height of the oscillation, which is Ω( 1 d ). 3 C. List decoding We now show that if the probability of getting a bad saple were greater than 1, then it is not possible to find poly(d) polynoials of degree O(d) such that one of the polynoials is close to the original polynoial. Theore V.8. Consider any algorith for robust polynoial regression that returns a set L of polynoials fro saples with outlier chance ρ =1/, such that at least one eleent of L is an C-approxiation to the true answer with 3/4 probability. Then E[ L ] 3 4 Ω(d/ C). Proof: Note that we ay assue d/c is larger than a sufficiently large constant, since otherwise the result follows fro the fact that L 1 whenever the algorith succeeds. Let us then set = d/ 1C, and take f S fro Lea V.3 for α = 1/(3C). For any S [] and x [ 1, 1], the lea iplies either f {} (x) f S (x) f {} (x)+α or f [] (x) α f S (x) f [] (x). Therefore, when observing any polynoial f S for S [], the adversary for σ = α can ensure that any saple point x is observed as f {} (x) with 50% probability, and f [] (x) with 50% probability. This is because p(x) is always within α of one of these, so the adversary can output that one if x is an inlier, and the other one if x is an outlier. For this adversary, the algorith s input is independent of the choice of f S. 399

10 Hence its distribution on output L is also independent of f S. But each p L can only be a C-approxiation to at ost one f S. Hence, if S [] is chosen at rando, Pr[ p L : p f S Cσ] E L. This iplies the result. ACKNOWLEDGEMENTS We would like to thank user111 on MathOverflow [15] for pointing us to [14]. DK is supported in part by NSF Award CCF (CAREER) and a Sloan Fellowship. REFERENCES [1] J. Gergonne, The application of the ethod of least squares to the interpolation of sequences, Historia Matheatica, vol. 1, no. 4, pp , [] V. Pratt, Direct least-squares fitting of algebraic surfaces, SIGGRAPH Coput. Graph., vol. 1, no. 4, pp , Aug [Online]. Available: http: //doi.ac.org/ / [3] A. T. Kalai, A. R. Klivans, Y. Mansour, and R. A. Servedio, Agnostically learning halfspaces, SIAM Journal on Coputing, vol. 37, no. 6, pp , 008. [4] I. B. MacNeill, Properties of sequences of partial sus of polynoial regression residuals with applications to tests for change of regression at unknown ties, The Annals of Statistics, pp , [5] S. Arora and S. Khot, Fitting algebraic curves to noisy data, Journal of Coputer and Syste Sciences, vol. 67, no., pp , 003. [6] M. A. Fischler and R. C. Bolles, Rando saple consensus: a paradig for odel fitting with applications to iage analysis and autoated cartography, Counications of the ACM, vol. 4, no. 6, pp , [7] V. Guruswai and D. Zuckeran, Robust Fourier and polynoial curve fitting, FOCS, 016. [8] L. R. Welch and E. R. Berlekap, Error correction for algebraic block codes, Dec , us Patent 4,633,470. [9] V. Guruswai and M. Sudan, Iproved decoding of reed-soloon and algebraic-geoetric codes, in Foundations of Coputer Science, Proceedings. 39th Annual Syposiu on. IEEE, 1998, pp [10] P. J. Huber, Robust statistics. Springer, 011. [11] K. A. Lai, A. B. Rao, and S. Vepala, Agnostic estiation of ean and covariance, in Foundations of Coputer Science (FOCS), 016 IEEE 57th Annual Syposiu on. IEEE, 016, pp [1] I. Diakonikolas, G. Kaath, D. M. Kane, J. Li, A. Moitra, and A. Stewart, Robust estiators in high diensions without the coputational intractability, in Foundations of Coputer Science (FOCS), 016 IEEE 57th Annual Syposiu on. IEEE, 016, pp [13] M. Charikar, J. Steinhardt, and G. Valiant, Learning fro untrusted data, arxiv preprint arxiv: , 016. [14] P. G. Nevai, Bernstein s inequality in L p for 0 <p< 1, Journal of Approxiation Theory, vol. 7, no. 3, pp , [15] user111, L1 analog of bernstein s inequality, MathOverflow, url: ). [Online]. Available: http: //athoverflow.net/q/43447 [16] N. I. Achieser, Theory of approxiation. Courier Corporation, 013. APPENDIX To prove Lea II.1, we need a a generalization of Bernstein s inequality fro l to l q for all q > 0, which appears as Theore 5 of [14]. In the univariate case and setting its paraeters γ,γ=0, it states Lea A.1 (Special case of Theore 5 of [14]). There exists a universal constant C such that, for any degree d polynoial p(x) and any q>0, x p (x) q dx Cd q p(x) q dx. Lea II.1. Let p be a degree d polynoial. Let I 1,...,I partition [ 1, 1] between the Chebyshev extrea cos πj, for soe d. Let r :[ 1, 1] R be piecewise constant, so that for each I k there exists an x k I k with r(x) =p(x k ) for all x I k. Then there exists a universal constant C such that, for any q 1, p r q C d p q. Proof: For any individual I k and x k I k,wehave by Hölder s inequality that x q p(x) p(x k) q dx = p (y)dy dx x I k x I k x x x x q 1 p (y) q dydx x I k x 1 I k q x I k p (x) q dx. Hence p(x) r(x) q dx I k q x I k x I k p (x) q dx. (8) 400

11 1 or We first consider I,...I /, then separately consider the first interval I 1. The stateent holds for the intervals I /,...,I by syetry. For each interval I k with k, we have for all x I k I k =cos = k/ ( kπ ) cos sin(kπ/) ( ) (k +1)π k 1 x where we use the notation a b to denote that there exists a universal constant C such that a Cb. Hence p(x) r(x) q dx I k q p (x) q dx x I k x I k 1 (c) q 1 x p (x) q dx x I k and so by Lea A.1, for soe constant c, x I I / p(x) r(x) q dx 1 1 (c) q ( ) q d 1 x p (x) q dx 1 p q. (9) Now we consider the end I 1. By the Markov brothers inequality [16], p d p. Let x [ 1, 1] such that p(x ) = p, and let I = {y [ 1, 1] x y 1 d }.Wehavep(y) p / for all y I,so p q q I ( p /) q p q d q. Hence by (8), and using that I 1 =1 cos π =Θ( 1 ), x I 1 p(x) r(x) q dx I 1 q+1 p q ( d ) q+ p q q c d q (c) q+ p q For soe constant c. The sae holds for intervals I /,...,I by syetry, so cobining with (9) gives ( ) ( q ( ) ) d d p(x) r(x) q dx 1+ q+1 p q q c 1 p r q d ( 1+ ( ) ) /q d p q. When d, the first ter doinates giving the result. Lea V.. Let d 1. For any point b [ 1, 1], there exists a degree d polynoial p b such that p b (b) =1, p b =1, and for all x [ 1, 1]. p b (x) d x b Proof: We will show a stronger for of this lea for even d and b =0, giving p 0 (x) 1 (d +1) x. (10) By subtracting 1 fro odd d, this iplies the sae for general d with a 1/d rather than 1/(d +1) ter. Then for general b we take p b (x) =p 0 ((x b)/), which satisfies the lea. So it suffices to show (10) for even d and b =0. We choose p 0 (x) to be so that p(x) :=( 1) d/ T d+1(x) (d +1)x d/ cos((d +1)θ) p(cos θ) =( 1) (d +1)cosθ or, replacing θ with π θ and using that sin(d π +ψ) = ( 1) d/ sin ψ, p(sin θ) = sin((d +1)θ) (d +1)sinθ. Since sin((d +1)θ) 1, this iediately gives (10); we just need to show p(0) = p =1.ByL Hôpital s rule, we have (d +1)cos((d +1) 0) p(0) = =1. (d + 1) cos 0 If θ 1.1/(d+1), then sin θ 1/(d+1), and so (10) iplies p(sin θ) 1. Since p is syetric, all that reains is to show p(sin θ) 1 for 0 <θ<1.1/(d + 1). The axiu value of p(sin θ) will either appear at an endpoint of this interval which we have already shown is at ost 1 or at a zero of the derivative. We have +1)cos((d +1)θ) p(sin θ) =(d θ (d +1)sinθ sin((d +1)θ) (d +1)cosθ ((d +1)sinθ) (d +1)cos((d +1)θ)sinθ sin((d +1)θ)cosθ = (d +1)sin. θ For all 0 <ψ, we have the inequalities ψ ψ 3 /6 < sin ψ<ψand 1 ψ / < cos ψ<1 ψ /+ψ 4 /4. 401

12 Hence for 0 <θ<1.1/(d +1), the denoinator is positive and the nuerator is less than (d + 1)(1 (d +1) θ /+(d +1) 4 θ 4 /4)θ ( (d +1)θ (d +1) 3 θ 3 /6 ) (1 θ /) = θ 3 ( (d +1) 3 /+(d +1)/+(d +1) 3 /6 ) + θ 5 ( (d +1) 5 /4 (d +1) 3 /1 ) 5 18 (θ(d +1))3 +(θ(d +1)) 5 /4 0.(θ(d +1)) 3 < 0. Thus p(sin θ) does not have a local axiu on (0, 1.1/(d +1)),so p 1, finishing the proof. Lea V.3. For any d and α>0, let = d α/, and define b j = 1+ j for j []. Consider the set of degree-d polynoials f S (x) = j S p b j (x) for S []. For any x [ 1, 1], let k x [] iniize b kx x. Then for any S [], f {kx} S(x) f S (x) f {kx} S(x)+α. Proof: Since f S {kx}(x) =f {kx}(x)+f S\{kx}(x) for any S, it is sufficient to prove the result when k x / S. Then f S (x) as desired. p 4 b j (x) d x b j j k x j k x j k x 4 < d <α d (( j k x 1) ) 1 (j 1) < π d 6 j=1 40

Robust polynomial regression up to the information theoretic limit

Robust polynomial regression up to the information theoretic limit Robust polynomial regression up to the information theoretic limit Daniel Kane Sushrut Karmalkar Eric Price August 16, 017 Abstract We consider the problem of robust polynomial regression, where one receives

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

Lecture 21. Interior Point Methods Setup and Algorithm

Lecture 21. Interior Point Methods Setup and Algorithm Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information

3.8 Three Types of Convergence

3.8 Three Types of Convergence 3.8 Three Types of Convergence 3.8 Three Types of Convergence 93 Suppose that we are given a sequence functions {f k } k N on a set X and another function f on X. What does it ean for f k to converge to

More information

The Weierstrass Approximation Theorem

The Weierstrass Approximation Theorem 36 The Weierstrass Approxiation Theore Recall that the fundaental idea underlying the construction of the real nubers is approxiation by the sipler rational nubers. Firstly, nubers are often deterined

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Tight Information-Theoretic Lower Bounds for Welfare Maximization in Combinatorial Auctions

Tight Information-Theoretic Lower Bounds for Welfare Maximization in Combinatorial Auctions Tight Inforation-Theoretic Lower Bounds for Welfare Maxiization in Cobinatorial Auctions Vahab Mirrokni Jan Vondrák Theory Group, Microsoft Dept of Matheatics Research Princeton University Redond, WA 9805

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

List Scheduling and LPT Oliver Braun (09/05/2017)

List Scheduling and LPT Oliver Braun (09/05/2017) List Scheduling and LPT Oliver Braun (09/05/207) We investigate the classical scheduling proble P ax where a set of n independent jobs has to be processed on 2 parallel and identical processors (achines)

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

Testing Properties of Collections of Distributions

Testing Properties of Collections of Distributions Testing Properties of Collections of Distributions Reut Levi Dana Ron Ronitt Rubinfeld April 9, 0 Abstract We propose a fraework for studying property testing of collections of distributions, where the

More information

Support Vector Machines MIT Course Notes Cynthia Rudin

Support Vector Machines MIT Course Notes Cynthia Rudin Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

1 Rademacher Complexity Bounds

1 Rademacher Complexity Bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #10 Scribe: Max Goer March 07, 2013 1 Radeacher Coplexity Bounds Recall the following theore fro last lecture: Theore 1. With probability

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval Unifor Approxiation and Bernstein Polynoials with Coefficients in the Unit Interval Weiang Qian and Marc D. Riedel Electrical and Coputer Engineering, University of Minnesota 200 Union St. S.E. Minneapolis,

More information

On the Inapproximability of Vertex Cover on k-partite k-uniform Hypergraphs

On the Inapproximability of Vertex Cover on k-partite k-uniform Hypergraphs On the Inapproxiability of Vertex Cover on k-partite k-unifor Hypergraphs Venkatesan Guruswai and Rishi Saket Coputer Science Departent Carnegie Mellon University Pittsburgh, PA 1513. Abstract. Coputing

More information

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Multi-Dimensional Hegselmann-Krause Dynamics

Multi-Dimensional Hegselmann-Krause Dynamics Multi-Diensional Hegselann-Krause Dynaics A. Nedić Industrial and Enterprise Systes Engineering Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu B. Touri Coordinated Science Laboratory

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents

More information

1 Generalization bounds based on Rademacher complexity

1 Generalization bounds based on Rademacher complexity COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #0 Scribe: Suqi Liu March 07, 08 Last tie we started proving this very general result about how quickly the epirical average converges

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

Improved Guarantees for Agnostic Learning of Disjunctions

Improved Guarantees for Agnostic Learning of Disjunctions Iproved Guarantees for Agnostic Learning of Disjunctions Pranjal Awasthi Carnegie Mellon University pawasthi@cs.cu.edu Avri Blu Carnegie Mellon University avri@cs.cu.edu Or Sheffet Carnegie Mellon University

More information

lecture 36: Linear Multistep Mehods: Zero Stability

lecture 36: Linear Multistep Mehods: Zero Stability 95 lecture 36: Linear Multistep Mehods: Zero Stability 5.6 Linear ultistep ethods: zero stability Does consistency iply convergence for linear ultistep ethods? This is always the case for one-step ethods,

More information

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science A Better Algorith For an Ancient Scheduling Proble David R. Karger Steven J. Phillips Eric Torng Departent of Coputer Science Stanford University Stanford, CA 9435-4 Abstract One of the oldest and siplest

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Topic 5a Introduction to Curve Fitting & Linear Regression

Topic 5a Introduction to Curve Fitting & Linear Regression /7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Convex Programming for Scheduling Unrelated Parallel Machines

Convex Programming for Scheduling Unrelated Parallel Machines Convex Prograing for Scheduling Unrelated Parallel Machines Yossi Azar Air Epstein Abstract We consider the classical proble of scheduling parallel unrelated achines. Each job is to be processed by exactly

More information

Chaotic Coupled Map Lattices

Chaotic Coupled Map Lattices Chaotic Coupled Map Lattices Author: Dustin Keys Advisors: Dr. Robert Indik, Dr. Kevin Lin 1 Introduction When a syste of chaotic aps is coupled in a way that allows the to share inforation about each

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory TTIC 31120 Prof. Nati Srebro Lecture 2: PAC Learning and VC Theory I Fro Adversarial Online to Statistical Three reasons to ove fro worst-case deterinistic

More information

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40 On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and This article appeared in a ournal published by Elsevier. The attached copy is furnished to the author for internal non-coercial research and education use, including for instruction at the authors institution

More information

Estimating Entropy and Entropy Norm on Data Streams

Estimating Entropy and Entropy Norm on Data Streams Estiating Entropy and Entropy Nor on Data Streas Ait Chakrabarti 1, Khanh Do Ba 1, and S. Muthukrishnan 2 1 Departent of Coputer Science, Dartouth College, Hanover, NH 03755, USA 2 Departent of Coputer

More information

Shannon Sampling II. Connections to Learning Theory

Shannon Sampling II. Connections to Learning Theory Shannon Sapling II Connections to Learning heory Steve Sale oyota echnological Institute at Chicago 147 East 60th Street, Chicago, IL 60637, USA E-ail: sale@athberkeleyedu Ding-Xuan Zhou Departent of Matheatics,

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

Collision-based Testers are Optimal for Uniformity and Closeness

Collision-based Testers are Optimal for Uniformity and Closeness Electronic Colloquiu on Coputational Coplexity, Report No. 178 (016) Collision-based Testers are Optial for Unifority and Closeness Ilias Diakonikolas Theis Gouleakis John Peebles Eric Price USC MIT MIT

More information

Bipartite subgraphs and the smallest eigenvalue

Bipartite subgraphs and the smallest eigenvalue Bipartite subgraphs and the sallest eigenvalue Noga Alon Benny Sudaov Abstract Two results dealing with the relation between the sallest eigenvalue of a graph and its bipartite subgraphs are obtained.

More information

CSE525: Randomized Algorithms and Probabilistic Analysis May 16, Lecture 13

CSE525: Randomized Algorithms and Probabilistic Analysis May 16, Lecture 13 CSE55: Randoied Algoriths and obabilistic Analysis May 6, Lecture Lecturer: Anna Karlin Scribe: Noah Siegel, Jonathan Shi Rando walks and Markov chains This lecture discusses Markov chains, which capture

More information

On Conditions for Linearity of Optimal Estimation

On Conditions for Linearity of Optimal Estimation On Conditions for Linearity of Optial Estiation Erah Akyol, Kuar Viswanatha and Kenneth Rose {eakyol, kuar, rose}@ece.ucsb.edu Departent of Electrical and Coputer Engineering University of California at

More information

a a a a a a a m a b a b

a a a a a a a m a b a b Algebra / Trig Final Exa Study Guide (Fall Seester) Moncada/Dunphy Inforation About the Final Exa The final exa is cuulative, covering Appendix A (A.1-A.5) and Chapter 1. All probles will be ultiple choice

More information

1 Identical Parallel Machines

1 Identical Parallel Machines FB3: Matheatik/Inforatik Dr. Syaantak Das Winter 2017/18 Optiizing under Uncertainty Lecture Notes 3: Scheduling to Miniize Makespan In any standard scheduling proble, we are given a set of jobs J = {j

More information

Math Real Analysis The Henstock-Kurzweil Integral

Math Real Analysis The Henstock-Kurzweil Integral Math 402 - Real Analysis The Henstock-Kurzweil Integral Steven Kao & Jocelyn Gonzales April 28, 2015 1 Introduction to the Henstock-Kurzweil Integral Although the Rieann integral is the priary integration

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 11 10/15/2008 ABSTRACT INTEGRATION I

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 11 10/15/2008 ABSTRACT INTEGRATION I MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 11 10/15/2008 ABSTRACT INTEGRATION I Contents 1. Preliinaries 2. The ain result 3. The Rieann integral 4. The integral of a nonnegative

More information

A Bernstein-Markov Theorem for Normed Spaces

A Bernstein-Markov Theorem for Normed Spaces A Bernstein-Markov Theore for Nored Spaces Lawrence A. Harris Departent of Matheatics, University of Kentucky Lexington, Kentucky 40506-0027 Abstract Let X and Y be real nored linear spaces and let φ :

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

In this chapter, we consider several graph-theoretic and probabilistic models

In this chapter, we consider several graph-theoretic and probabilistic models THREE ONE GRAPH-THEORETIC AND STATISTICAL MODELS 3.1 INTRODUCTION In this chapter, we consider several graph-theoretic and probabilistic odels for a social network, which we do under different assuptions

More information

Estimating Parameters for a Gaussian pdf

Estimating Parameters for a Gaussian pdf Pattern Recognition and achine Learning Jaes L. Crowley ENSIAG 3 IS First Seester 00/0 Lesson 5 7 Noveber 00 Contents Estiating Paraeters for a Gaussian pdf Notation... The Pattern Recognition Proble...3

More information

Homework 3 Solutions CSE 101 Summer 2017

Homework 3 Solutions CSE 101 Summer 2017 Hoework 3 Solutions CSE 0 Suer 207. Scheduling algoriths The following n = 2 jobs with given processing ties have to be scheduled on = 3 parallel and identical processors with the objective of iniizing

More information

Lecture October 23. Scribes: Ruixin Qiang and Alana Shine

Lecture October 23. Scribes: Ruixin Qiang and Alana Shine CSCI699: Topics in Learning and Gae Theory Lecture October 23 Lecturer: Ilias Scribes: Ruixin Qiang and Alana Shine Today s topic is auction with saples. 1 Introduction to auctions Definition 1. In a single

More information

Understanding Machine Learning Solution Manual

Understanding Machine Learning Solution Manual Understanding Machine Learning Solution Manual Written by Alon Gonen Edited by Dana Rubinstein Noveber 17, 2014 2 Gentle Start 1. Given S = ((x i, y i )), define the ultivariate polynoial p S (x) = i []:y

More information

VC Dimension and Sauer s Lemma

VC Dimension and Sauer s Lemma CMSC 35900 (Spring 2008) Learning Theory Lecture: VC Diension and Sauer s Lea Instructors: Sha Kakade and Abuj Tewari Radeacher Averages and Growth Function Theore Let F be a class of ±-valued functions

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

Learnability and Stability in the General Learning Setting

Learnability and Stability in the General Learning Setting Learnability and Stability in the General Learning Setting Shai Shalev-Shwartz TTI-Chicago shai@tti-c.org Ohad Shair The Hebrew University ohadsh@cs.huji.ac.il Nathan Srebro TTI-Chicago nati@uchicago.edu

More information

Computable Shell Decomposition Bounds

Computable Shell Decomposition Bounds Coputable Shell Decoposition Bounds John Langford TTI-Chicago jcl@cs.cu.edu David McAllester TTI-Chicago dac@autoreason.co Editor: Leslie Pack Kaelbling and David Cohn Abstract Haussler, Kearns, Seung

More information

New upper bound for the B-spline basis condition number II. K. Scherer. Institut fur Angewandte Mathematik, Universitat Bonn, Bonn, Germany.

New upper bound for the B-spline basis condition number II. K. Scherer. Institut fur Angewandte Mathematik, Universitat Bonn, Bonn, Germany. New upper bound for the B-spline basis condition nuber II. A proof of de Boor's 2 -conjecture K. Scherer Institut fur Angewandte Matheati, Universitat Bonn, 535 Bonn, Gerany and A. Yu. Shadrin Coputing

More information

Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression

Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression Sha M Kakade Microsoft Research and Wharton, U Penn skakade@icrosoftco Varun Kanade SEAS, Harvard University vkanade@fasharvardedu

More information

SPECTRUM sensing is a core concept of cognitive radio

SPECTRUM sensing is a core concept of cognitive radio World Acadey of Science, Engineering and Technology International Journal of Electronics and Counication Engineering Vol:6, o:2, 202 Efficient Detection Using Sequential Probability Ratio Test in Mobile

More information

Support recovery in compressed sensing: An estimation theoretic approach

Support recovery in compressed sensing: An estimation theoretic approach Support recovery in copressed sensing: An estiation theoretic approach Ain Karbasi, Ali Horati, Soheil Mohajer, Martin Vetterli School of Coputer and Counication Sciences École Polytechnique Fédérale de

More information

Exact tensor completion with sum-of-squares

Exact tensor completion with sum-of-squares Proceedings of Machine Learning Research vol 65:1 54, 2017 30th Annual Conference on Learning Theory Exact tensor copletion with su-of-squares Aaron Potechin Institute for Advanced Study, Princeton David

More information

. The univariate situation. It is well-known for a long tie that denoinators of Pade approxiants can be considered as orthogonal polynoials with respe

. The univariate situation. It is well-known for a long tie that denoinators of Pade approxiants can be considered as orthogonal polynoials with respe PROPERTIES OF MULTIVARIATE HOMOGENEOUS ORTHOGONAL POLYNOMIALS Brahi Benouahane y Annie Cuyt? Keywords Abstract It is well-known that the denoinators of Pade approxiants can be considered as orthogonal

More information

Probability and Stochastic Processes: A Friendly Introduction for Electrical and Computer Engineers Roy D. Yates and David J.

Probability and Stochastic Processes: A Friendly Introduction for Electrical and Computer Engineers Roy D. Yates and David J. Probability and Stochastic Processes: A Friendly Introduction for Electrical and oputer Engineers Roy D. Yates and David J. Goodan Proble Solutions : Yates and Goodan,1..3 1.3.1 1.4.6 1.4.7 1.4.8 1..6

More information

Solutions of some selected problems of Homework 4

Solutions of some selected problems of Homework 4 Solutions of soe selected probles of Hoework 4 Sangchul Lee May 7, 2018 Proble 1 Let there be light A professor has two light bulbs in his garage. When both are burned out, they are replaced, and the next

More information

Complex Quadratic Optimization and Semidefinite Programming

Complex Quadratic Optimization and Semidefinite Programming Coplex Quadratic Optiization and Seidefinite Prograing Shuzhong Zhang Yongwei Huang August 4 Abstract In this paper we study the approxiation algoriths for a class of discrete quadratic optiization probles

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

The degree of a typical vertex in generalized random intersection graph models

The degree of a typical vertex in generalized random intersection graph models Discrete Matheatics 306 006 15 165 www.elsevier.co/locate/disc The degree of a typical vertex in generalized rando intersection graph odels Jerzy Jaworski a, Michał Karoński a, Dudley Stark b a Departent

More information

Lecture 20 November 7, 2013

Lecture 20 November 7, 2013 CS 229r: Algoriths for Big Data Fall 2013 Prof. Jelani Nelson Lecture 20 Noveber 7, 2013 Scribe: Yun Willia Yu 1 Introduction Today we re going to go through the analysis of atrix copletion. First though,

More information

E0 370 Statistical Learning Theory Lecture 5 (Aug 25, 2011)

E0 370 Statistical Learning Theory Lecture 5 (Aug 25, 2011) E0 370 Statistical Learning Theory Lecture 5 Aug 5, 0 Covering Nubers, Pseudo-Diension, and Fat-Shattering Diension Lecturer: Shivani Agarwal Scribe: Shivani Agarwal Introduction So far we have seen how

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

On the Use of A Priori Information for Sparse Signal Approximations

On the Use of A Priori Information for Sparse Signal Approximations ITS TECHNICAL REPORT NO. 3/4 On the Use of A Priori Inforation for Sparse Signal Approxiations Oscar Divorra Escoda, Lorenzo Granai and Pierre Vandergheynst Signal Processing Institute ITS) Ecole Polytechnique

More information

Asynchronous Gossip Algorithms for Stochastic Optimization

Asynchronous Gossip Algorithms for Stochastic Optimization Asynchronous Gossip Algoriths for Stochastic Optiization S. Sundhar Ra ECE Dept. University of Illinois Urbana, IL 680 ssrini@illinois.edu A. Nedić IESE Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu

More information

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion Suppleentary Material for Fast and Provable Algoriths for Spectrally Sparse Signal Reconstruction via Low-Ran Hanel Matrix Copletion Jian-Feng Cai Tianing Wang Ke Wei March 1, 017 Abstract We establish

More information

OBJECTIVES INTRODUCTION

OBJECTIVES INTRODUCTION M7 Chapter 3 Section 1 OBJECTIVES Suarize data using easures of central tendency, such as the ean, edian, ode, and idrange. Describe data using the easures of variation, such as the range, variance, and

More information

arxiv: v1 [math.nt] 14 Sep 2014

arxiv: v1 [math.nt] 14 Sep 2014 ROTATION REMAINDERS P. JAMESON GRABER, WASHINGTON AND LEE UNIVERSITY 08 arxiv:1409.411v1 [ath.nt] 14 Sep 014 Abstract. We study properties of an array of nubers, called the triangle, in which each row

More information

Mixed Robust/Average Submodular Partitioning

Mixed Robust/Average Submodular Partitioning Mixed Robust/Average Subodular Partitioning Kai Wei 1 Rishabh Iyer 1 Shengjie Wang 2 Wenruo Bai 1 Jeff Biles 1 1 Departent of Electrical Engineering, University of Washington 2 Departent of Coputer Science,

More information

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate The Siplex Method is Strongly Polynoial for the Markov Decision Proble with a Fixed Discount Rate Yinyu Ye April 20, 2010 Abstract In this note we prove that the classic siplex ethod with the ost-negativereduced-cost

More information

Approximation in Stochastic Scheduling: The Power of LP-Based Priority Policies

Approximation in Stochastic Scheduling: The Power of LP-Based Priority Policies Approxiation in Stochastic Scheduling: The Power of -Based Priority Policies Rolf Möhring, Andreas Schulz, Marc Uetz Setting (A P p stoch, r E( w and (B P p stoch E( w We will assue that the processing

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

Generalized eigenfunctions and a Borel Theorem on the Sierpinski Gasket.

Generalized eigenfunctions and a Borel Theorem on the Sierpinski Gasket. Generalized eigenfunctions and a Borel Theore on the Sierpinski Gasket. Kasso A. Okoudjou, Luke G. Rogers, and Robert S. Strichartz May 26, 2006 1 Introduction There is a well developed theory (see [5,

More information

Handout 7. and Pr [M(x) = χ L (x) M(x) =? ] = 1.

Handout 7. and Pr [M(x) = χ L (x) M(x) =? ] = 1. Notes on Coplexity Theory Last updated: October, 2005 Jonathan Katz Handout 7 1 More on Randoized Coplexity Classes Reinder: so far we have seen RP,coRP, and BPP. We introduce two ore tie-bounded randoized

More information

Optimal Jamming Over Additive Noise: Vector Source-Channel Case

Optimal Jamming Over Additive Noise: Vector Source-Channel Case Fifty-first Annual Allerton Conference Allerton House, UIUC, Illinois, USA October 2-3, 2013 Optial Jaing Over Additive Noise: Vector Source-Channel Case Erah Akyol and Kenneth Rose Abstract This paper

More information