1 Large Deviations. Korea Lectures June 2017 Joel Spencer Tuesday Lecture III
|
|
- Jade Ramsey
- 5 years ago
- Views:
Transcription
1 Korea Lectures June 017 Joel Spencer Tuesday Lecture III 1 Large Deviations We analyze how random variables behave asymptotically Initially we consider the sum of random variables that take values of 1 and 1 Then we will consider sums of arbitrary random variables The key is what are called Chernoff Bounds Markov s Inequality: Given a non-negative random variable Z, Pr[Z αe[z]] α 1 Chernoff Bound: For any Z and any a Pr[Z > a] = Pr[E[e λz > e λa ] E[eλZ ] e λa for any nonnegative real λ The strength of the Chernoff Bound is that one selects λ optimally (more usually near optimally as the calculations are too hard for true optimality and generally we are happy with a reasonable bound) Let X 1,,X n be independent random variables defined as ±1 with probability 1/ It is immediate that E[X i ] = 0 and V ar[x i ] = 1 Define also S n = X X n The Central Limit Theorem says that when n is large S n N(0,n) Applying the Chernoff bound, for any λ > 0 Pr[S n an 1/ ] = Pr[e λsn e λan1/ ] E[e λsn ]e λan1/, where the first equality is because λ > 0 and monotonicity of the exponential function Theorem: Given n and a > 0, Pr[S n an 1/ ] e a / Proof: Recall that although the expectation of the multiplication of many RVs is not necessarily the multiplication of the expectation, that is true when the RVs are mutually independent As that is the case here, [ ] E[e λsn ] = E e λx i = E[e λx i ] = ( e λ + e λ ) n = cosh n λ ( e λ / ) n The last inequality can be shown by comparing the terms of both Taylor series Using the Chernoff Bound, we have 1
2 Pr[S n an 1/ ] e λ n/ λan 1/ = e a /, setting λ = an 1/ which minimizes λ n λan 1/ End of Proof Now consider a more general setting Let Y 1,,Y n be a set of mutually independent RVs with E[Y i ] = 0 and V [Y i ] = σi Define the sum of them as Y = Y i, thus E[Y ] = 0 and V [Y ] = σ σi We would like to conclude, as before, that the decay of Y is like e a /, but unfortunately that is not always true Using again the Markov Inequality and independence, But in this case Pr[Y aσ] E[e λy ]e λaσ = E[e λy i ]e λaσ [ ] E[e λy i λ j ] = E j! Y j i j=0 = λ σ i +? e λ σ i / is not always true For the last equality to be true we need that the higher order terms are small with respect to the quadratic one When this happens for all i, multiplying all together we see that Pr[Y aσ] < e λ σ / λaσ = e a /, letting λ be a/σ Most of the times we would be using this result, a will be small compared to σ because otherwise we would be looking very far off the mean Therefore, λ will be small and the high order terms will be negligible, which will make the bound correct Now let s try the same for a normal distribution If N N(0,1), it is known that its transform is E[e λn ] = Doing the same thing one more time, e t 1/ π e λt dt = e λ / Pr[N a] e λ / λa = e a /, letting λ be a The last equation is what it should intuitively be Indeed, as the sum of mutually independent RVs is roughly normal, the Chernoff Bound and this bound should say the same Applications Let s consider tournaments again We would like to see how well we can rank the vertices in a way that reflects the outcome of the tournament Given a ranking σ, we will measure its quality by defining FIT[T n,σ] = #nonupsets #upsets, where {i,j} is a nonupset when σ i < σ j and (σ i,σ j ) T n and an upset otherwise The overall measure of how good we do is given by
3 F(n) = min max FIT(T n,σ) T N σ This can be though of as an adversary giving us a tournament which we must rank as good as possible The adversary will of course choose the tournament that makes us return a bad solution Theorem: F(n) 0 Proof: Take a ranking and if it does not work just consider the opposite ranking Thats it Theorem: (Erdos- Moon) F(n) n 3/ ln n This means that there always exists a tournament T n such that no σ has FIT[T n,σ] > n 3/ ln n No matter how we rank it, it will be bad For a large n we won t be able to rank more than 50% of the games correctly Proof: Take a random T n Define the event B σ as FIT[T n,σ] > α By construction of T n, every game will count as 1 or 1 with equal probability As there are ( n ) games, FIT[T n,σ] S ( n ), where means that the two have the same distribution Thus Then with α = n 3/ ln n Pr[B σ ] = Pr[S ( n α] exp ) Pr[ B σ ] n!exp ( ) α ( n) ( ) α ( n) 3 Large Deviations: Generalizing the Chernoff Bound Theorem: (Generalized Chernoff Bound) Given n independent variables X i, with 1 X i 1, and E[X i ] = 0 for every i, let X = i X i Then Pr[X > a n] < e a / Proof: The only difference between this proof and the one we gave in the simpler case is that this time, we can t simply replace E[e λx i ] by eλ +e λ However, it turns out that, for X i as above, E[e λx i ] eλ + e λ Why would this be true? Let s examine the function e λx (see figure) As one can see, the function is convex, which means that it lies entirely under the linear function h(x) But then it follows that e λx h(x) so E[e λx i ] E[h(X i )] However, since h is linear, E[h(X i )] = h(e[x i ]), so we get E[e λx i ] E[h(X i )] = h(e[x i ]) = h(0) = eλ + e λ, so we have our upper bound, and the proof of the Chernoff Bound continues just as in the simpler case 3 < 1
4 4 More Applications to Large Deviations 41 Balancing Vectors Theorem: Let v 1,, v n be vectors in R m, with v i 1 for every i Then there is a choice of signs ǫ i { 1,1} for every v i, such that p n = i ǫ i v i nṗroof: As it happens, this theorem is easy to proof in non-probabilistic context, in an algorithmical way: introduce the vectors one by one, keeping track of the position vectors p i If the new vector v i+1 makes an acute angle with p i, add it (ǫ i = 1); if not, subtract it (ǫ i = 1); if they are perpedicular, do any of the two This result in a new position vector p i+1, which, by the law of cosines, has p i+1 p i + 1; by induction, this shows that the choice of ǫ s yields p n n The probabilistic proof is somewhat shorter: independently choose ǫ i to be 1 or 1 with equal probabilities Then look at E[ p n ] = E[ i,j ǫ i ǫ j v i v j ] = i,j E[ǫ i ǫ j ] v i v j ; When i j, E[ǫ i ǫ j ] = E[ǫ i ]E[ǫ j ] = 0 while when i = j, E[ǫ i ǫ j ] = E[ǫ i ] = 1 so the sum sieves out to E[ p n ] = i E[ǫ i ] v i i E[ǫ i ] = n ; so by the Probabilistic Method there must be a choice of ǫ s such that p n n End of Proof Theorem: Let v 1,, v n be vectors in R m with v i 1 for every i, and p 1,,p n be numbers in [0,1] Set w = i v ip i Then there is a choice of signs ǫ i { 1,1} for every v i, such that n ǫ i v i w i This has an interesting geometric interpretation When they are linearly independent the v i generate a parallelopiped with the n vertices i ǫ i v i w can lie anywhere inside this parallelopiped and some vertex will be within n from it distance proof Choose ǫ i s independently to be 1 with probability p i, and 0 with probability 1 p i Now E[(ǫ i p i )(ǫ j p j )] is zero when i j and p i (1 p i ) when i = j So E[ i ǫ i v i w ] = i,j E[(ǫ i p i )(ǫ j p j )] v i v j = i E[(ǫ i p i ) ] v i i E[(ǫ i p i ) ] = i p i (1 p i ) n 4, since x(1 x) 1/4, for any x [0,1] Hence, by the Probabilistic Method, there must be a choice of signs which leads to n ǫ i v i w i 4
5 4 Balancing Lights Description Suppose we have an n n array of lights, and n switches (1 for each row and each column) A switch changes all the lights on a particular column/row The purpose of the problem is to switch on as many lights as possible Anecdotal History This problem has been studied by Berlekamp (who is now at Berkeley); not surprisingly, both his undergraduate and graduate degrees are in Electrical Engineering! He actually built an 8 8 circuit, which he kept in his office, and invited people to play with it, to see how many lights they could turn on Helpful Remarks Switch pulling is order-independent, and is does not make sense to pull the same switch twice, since the effects will cancel each other Mathematical Formulation Since the total number of lights is n, maximizing the number of on s is the same with maximizing the difference (discrepancy) between the number of on s and off s Hence the above problem is equivalent to finding F(n) = min a ij = ±1 1 i,j, n max x i = ±1 y j = ±1 1 i,j, n a ij x i y j = xa y; i,j x i = 1 or y j = 1 corresponds to a switch in the ith row or jth column Note that F(n) 0, because if in the initial state of the matrix we have more off lights, we can switch every row and no columns, and thus switch it all around, and get to a state with more on lights Theorem: F(n) β = n 3/ ln Proof: The purpose is to find a matrix A which minimizes the discrepancy in lights Randomly set the a ij s to ±1, and define the bad event then define the union of bad events B x, y : xa y > β ; B = x, y B x, y Since there are n pairs of vectors x, y, and we want to use Union Bound, we want to make Pr[B x, y ] < 1 n As it turns out, Pr[ xa y > β] Pr[S n > β], since the great property of randomization is that multiplication by a random number randomizes, ie if x i, y j are arbitrary, then a ij x i y j is randomly ±1 So we get that Pr[B x, y ] < e β /n = 1 n ; in fact, you can see that this is how β was chosen in the first place 5
6 Hence, by the Probabilistic Method, there must be an A such that the discrepancy of it is no more than n 3/ ln, and the theorem follows End of Proof Remarks The above can be interpreted as follows: if an adversary is to hand us an n n array, we can not hope to turn on more than n /+n 3/ ln of the lights In reality, it s even less than that, because in the above proof, we have not accounted for the fact that antipodal switching (all rows and all columns) creates the exact same array of lights, so instead of having n possibilities, we have only n 1 As it turns out, this makes no difference whatsoever in the asymptotics The moral is, don t sweat the small stuff (assuming you can tell which stuff is small ) 6
Machine Learning Theory Lecture 4
Machine Learning Theory Lecture 4 Nicholas Harvey October 9, 018 1 Basic Probability One of the first concentration bounds that you learn in probability theory is Markov s inequality. It bounds the right-tail
More informationTail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).
Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested
More informationLecture 4: Two-point Sampling, Coupon Collector s problem
Randomized Algorithms Lecture 4: Two-point Sampling, Coupon Collector s problem Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms
More informationQuick Sort Notes , Spring 2010
Quick Sort Notes 18.310, Spring 2010 0.1 Randomized Median Finding In a previous lecture, we discussed the problem of finding the median of a list of m elements, or more generally the element of rank m.
More informationLectures 6: Degree Distributions and Concentration Inequalities
University of Washington Lecturer: Abraham Flaxman/Vahab S Mirrokni CSE599m: Algorithms and Economics of Networks April 13 th, 007 Scribe: Ben Birnbaum Lectures 6: Degree Distributions and Concentration
More informationCS5314 Randomized Algorithms. Lecture 18: Probabilistic Method (De-randomization, Sample-and-Modify)
CS5314 Randomized Algorithms Lecture 18: Probabilistic Method (De-randomization, Sample-and-Modify) 1 Introduce two topics: De-randomize by conditional expectation provides a deterministic way to construct
More informationTheorem (Special Case of Ramsey s Theorem) R(k, l) is finite. Furthermore, it satisfies,
Math 16A Notes, Wee 6 Scribe: Jesse Benavides Disclaimer: These notes are not nearly as polished (and quite possibly not nearly as correct) as a published paper. Please use them at your own ris. 1. Ramsey
More informationRandomized Algorithms Week 2: Tail Inequalities
Randomized Algorithms Week 2: Tail Inequalities Rao Kosaraju In this section, we study three ways to estimate the tail probabilities of random variables. Please note that the more information we know about
More informationLecture 5: January 30
CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 5: January 30 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They
More informationIntroduction to Algorithms
Lecture 1 Introduction to Algorithms 1.1 Overview The purpose of this lecture is to give a brief overview of the topic of Algorithms and the kind of thinking it involves: why we focus on the subjects that
More informationExpectation of geometric distribution
Expectation of geometric distribution What is the probability that X is finite? Can now compute E(X): Σ k=1f X (k) = Σ k=1(1 p) k 1 p = pσ j=0(1 p) j = p 1 1 (1 p) = 1 E(X) = Σ k=1k (1 p) k 1 p = p [ Σ
More informationCS 6820 Fall 2014 Lectures, October 3-20, 2014
Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given
More informationDiscrete Mathematics and Probability Theory Fall 2011 Rao Midterm 2 Solutions
CS 70 Discrete Mathematics and Probability Theory Fall 20 Rao Midterm 2 Solutions True/False. [24 pts] Circle one of the provided answers please! No negative points will be assigned for incorrect answers.
More informationCS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003
CS6999 Probabilistic Methods in Integer Programming Randomized Rounding April 2003 Overview 2 Background Randomized Rounding Handling Feasibility Derandomization Advanced Techniques Integer Programming
More informationNotes on Discrete Probability
Columbia University Handout 3 W4231: Analysis of Algorithms September 21, 1999 Professor Luca Trevisan Notes on Discrete Probability The following notes cover, mostly without proofs, the basic notions
More information1 Review of The Learning Setting
COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #8 Scribe: Changyan Wang February 28, 208 Review of The Learning Setting Last class, we moved beyond the PAC model: in the PAC model we
More information11.1 Set Cover ILP formulation of set cover Deterministic rounding
CS787: Advanced Algorithms Lecture 11: Randomized Rounding, Concentration Bounds In this lecture we will see some more examples of approximation algorithms based on LP relaxations. This time we will use
More informationVector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n.
Vector Spaces Definition: The usual addition and scalar multiplication of n-tuples x = (x 1,..., x n ) R n (also called vectors) are the addition and scalar multiplication operations defined component-wise:
More informationExpectation of geometric distribution. Variance and Standard Deviation. Variance: Examples
Expectation of geometric distribution Variance and Standard Deviation What is the probability that X is finite? Can now compute E(X): Σ k=f X (k) = Σ k=( p) k p = pσ j=0( p) j = p ( p) = E(X) = Σ k=k (
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationIntroduction to Mathematical Programming IE406. Lecture 3. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 3 Dr. Ted Ralphs IE406 Lecture 3 1 Reading for This Lecture Bertsimas 2.1-2.2 IE406 Lecture 3 2 From Last Time Recall the Two Crude Petroleum example.
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationInteger Linear Programs
Lecture 2: Review, Linear Programming Relaxations Today we will talk about expressing combinatorial problems as mathematical programs, specifically Integer Linear Programs (ILPs). We then see what happens
More informationLectures One Way Permutations, Goldreich Levin Theorem, Commitments
Lectures 11 12 - One Way Permutations, Goldreich Levin Theorem, Commitments Boaz Barak March 10, 2010 From time immemorial, humanity has gotten frequent, often cruel, reminders that many things are easier
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationAssignment 4: Solutions
Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each
More informationA RANDOM VARIANT OF THE GAME OF PLATES AND OLIVES
A RANDOM VARIANT OF THE GAME OF PLATES AND OLIVES ANDRZEJ DUDEK, SEAN ENGLISH, AND ALAN FRIEZE Abstract The game of plates and olives was originally formulated by Nicolaescu and encodes the evolution of
More informationRANDOM GRAPHS. Joel Spencer ICTP 12
RANDOM GRAPHS Joel Spencer ICTP 12 Graph Theory Preliminaries A graph G, formally speaking, is a pair (V (G), E(G)) where the elements v V (G) are called vertices and the elements of E(G), called edges,
More informationX = X X n, + X 2
CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk
More informationThe first bound is the strongest, the other two bounds are often easier to state and compute. Proof: Applying Markov's inequality, for any >0 we have
The first bound is the strongest, the other two bounds are often easier to state and compute Proof: Applying Markov's inequality, for any >0 we have Pr (1 + ) = Pr For any >0, we can set = ln 1+ (4.4.1):
More informationLecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality
Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality Discrete Structures II (Summer 2018) Rutgers University Instructor: Abhishek Bhrushundi
More informationExpectation is linear. So far we saw that E(X + Y ) = E(X) + E(Y ). Let α R. Then,
Expectation is linear So far we saw that E(X + Y ) = E(X) + E(Y ). Let α R. Then, E(αX) = ω = ω (αx)(ω) Pr(ω) αx(ω) Pr(ω) = α ω X(ω) Pr(ω) = αe(x). Corollary. For α, β R, E(αX + βy ) = αe(x) + βe(y ).
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationExpected Value II. 1 The Expected Number of Events that Happen
6.042/18.062J Mathematics for Computer Science December 5, 2006 Tom Leighton and Ronitt Rubinfeld Lecture Notes Expected Value II 1 The Expected Number of Events that Happen Last week we concluded by showing
More informationRandomized algorithm
Tutorial 4 Joyce 2009-11-24 Outline Solution to Midterm Question 1 Question 2 Question 1 Question 2 Question 3 Question 4 Question 5 Solution to Midterm Solution to Midterm Solution to Midterm Question
More informationA Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover
A Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover Grant Schoenebeck Luca Trevisan Madhur Tulsiani Abstract We study semidefinite programming relaxations of Vertex Cover arising
More informationMidterm Exam 1 (Solutions)
EECS 6 Probability and Random Processes University of California, Berkeley: Spring 07 Kannan Ramchandran February 3, 07 Midterm Exam (Solutions) Last name First name SID Name of student on your left: Name
More informationLecture 4. P r[x > ce[x]] 1/c. = ap r[x = a] + a>ce[x] P r[x = a]
U.C. Berkeley CS273: Parallel and Distributed Theory Lecture 4 Professor Satish Rao September 7, 2010 Lecturer: Satish Rao Last revised September 13, 2010 Lecture 4 1 Deviation bounds. Deviation bounds
More informationMarch 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.
Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)
More informationLecture 5: Two-point Sampling
Randomized Algorithms Lecture 5: Two-point Sampling Sotiris Nikoletseas Professor CEID - ETY Course 2017-2018 Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 1 / 26 Overview A. Pairwise
More informationLecture 4: Inequalities and Asymptotic Estimates
CSE 713: Random Graphs and Applications SUNY at Buffalo, Fall 003 Lecturer: Hung Q. Ngo Scribe: Hung Q. Ngo Lecture 4: Inequalities and Asymptotic Estimates We draw materials from [, 5, 8 10, 17, 18].
More informationHomework 4 Solutions
CS 174: Combinatorics and Discrete Probability Fall 01 Homework 4 Solutions Problem 1. (Exercise 3.4 from MU 5 points) Recall the randomized algorithm discussed in class for finding the median of a set
More information15-251: Great Theoretical Ideas In Computer Science Recitation 9 : Randomized Algorithms and Communication Complexity Solutions
15-251: Great Theoretical Ideas In Computer Science Recitation 9 : Randomized Algorithms and Communication Complexity Solutions Definitions We say a (deterministic) protocol P computes f if (x, y) {0,
More informationDefinition 2.3. We define addition and multiplication of matrices as follows.
14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row
More informationECEN 689 Special Topics in Data Science for Communications Networks
ECEN 689 Special Topics in Data Science for Communications Networks Nick Duffield Department of Electrical & Computer Engineering Texas A&M University Lecture 3 Estimation and Bounds Estimation from Packet
More informationCSE 525 Randomized Algorithms & Probabilistic Analysis Spring Lecture 3: April 9
CSE 55 Randomized Algorithms & Probabilistic Analysis Spring 01 Lecture : April 9 Lecturer: Anna Karlin Scribe: Tyler Rigsby & John MacKinnon.1 Kinds of randomization in algorithms So far in our discussion
More informationSecond-Order Homogeneous Linear Equations with Constant Coefficients
15 Second-Order Homogeneous Linear Equations with Constant Coefficients A very important class of second-order homogeneous linear equations consists of those with constant coefficients; that is, those
More informationAdvanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12
Advanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12 Lecture 4: Multiarmed Bandit in the Adversarial Model Lecturer: Yishay Mansour Scribe: Shai Vardi 4.1 Lecture Overview
More informationExpectation, inequalities and laws of large numbers
Chapter 3 Expectation, inequalities and laws of large numbers 3. Expectation and Variance Indicator random variable Let us suppose that the event A partitions the sample space S, i.e. A A S. The indicator
More information1 Maintaining a Dictionary
15-451/651: Design & Analysis of Algorithms February 1, 2016 Lecture #7: Hashing last changed: January 29, 2016 Hashing is a great practical tool, with an interesting and subtle theory too. In addition
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More informationLecture 7: Schwartz-Zippel Lemma, Perfect Matching. 1.1 Polynomial Identity Testing and Schwartz-Zippel Lemma
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 7: Schwartz-Zippel Lemma, Perfect Matching Lecturer: Shayan Oveis Gharan 01/30/2017 Scribe: Philip Cho Disclaimer: These notes have not
More informationConcentration Inequalities
Chapter Concentration Inequalities I. Moment generating functions, the Chernoff method, and sub-gaussian and sub-exponential random variables a. Goal for this section: given a random variable X, how does
More information6.080 / Great Ideas in Theoretical Computer Science Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 6.080 / 6.089 Great Ideas in Theoretical Computer Science Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationLecture 5. 1 Review (Pairwise Independence and Derandomization)
6.842 Randomness and Computation September 20, 2017 Lecture 5 Lecturer: Ronitt Rubinfeld Scribe: Tom Kolokotrones 1 Review (Pairwise Independence and Derandomization) As we discussed last time, we can
More informationCS5314 Randomized Algorithms. Lecture 5: Discrete Random Variables and Expectation (Conditional Expectation, Geometric RV)
CS5314 Randomized Algorithms Lecture 5: Discrete Random Variables and Expectation (Conditional Expectation, Geometric RV) Objectives Introduce Geometric RV We then introduce Conditional Expectation Application:
More informationProbability Notes (A) , Fall 2010
Probability Notes (A) 18.310, Fall 2010 We are going to be spending around four lectures on probability theory this year. These notes cover approximately the first three lectures on it. Probability theory
More informationThe Probabilistic Method
Lecture 3: Tail bounds, Probabilistic Method Today we will see what is nown as the probabilistic method for showing the existence of combinatorial objects. We also review basic concentration inequalities.
More informationProjective Spaces. Chapter The Projective Line
Chapter 3 Projective Spaces 3.1 The Projective Line Suppose you want to describe the lines through the origin O = (0, 0) in the Euclidean plane R 2. The first thing you might think of is to write down
More informationDiscrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20
CS 70 Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20 Today we shall discuss a measure of how close a random variable tends to be to its expectation. But first we need to see how to compute
More informationNotes 6 : First and second moment methods
Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More information2. Limits at Infinity
2 Limits at Infinity To understand sequences and series fully, we will need to have a better understanding of its at infinity We begin with a few examples to motivate our discussion EXAMPLE 1 Find SOLUTION
More informationEssentials on the Analysis of Randomized Algorithms
Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random
More informationErrata: First Printing of Mitzenmacher/Upfal Probability and Computing
Errata: First Printing of Mitzenmacher/Upfal Probability and Computing Michael Mitzenmacher and Eli Upfal October 10, 2006 We would like to thank the many of you who have bought our book, and we would
More information1 Proof techniques. CS 224W Linear Algebra, Probability, and Proof Techniques
1 Proof techniques Here we will learn to prove universal mathematical statements, like the square of any odd number is odd. It s easy enough to show that this is true in specific cases for example, 3 2
More informationIntroduction to Probability
Introduction to Probability Math 353, Section 1 Fall 212 Homework 9 Solutions 1. During a season, a basketball team plays 7 home games and 6 away games. The coach estimates at the beginning of the season
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationMath 1b Sequences and series summary
Math b Sequences and series summary December 22, 2005 Sequences (Stewart p. 557) Notations for a sequence: or a, a 2, a 3,..., a n,... {a n }. The numbers a n are called the terms of the sequence.. Limit
More informationDiscrete Mathematics and Probability Theory Fall 2010 Tse/Wagner MT 2 Soln
CS 70 Discrete Mathematics and Probability heory Fall 00 se/wagner M Soln Problem. [Rolling Dice] (5 points) You roll a fair die three times. Consider the following events: A first roll is a 3 B second
More informationLecture 7: More Arithmetic and Fun With Primes
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Advanced Course on Computational Complexity Lecture 7: More Arithmetic and Fun With Primes David Mix Barrington and Alexis Maciel July
More informationFinal Solutions Fri, June 8
EE178: Probabilistic Systems Analysis, Spring 2018 Final Solutions Fri, June 8 1. Small problems (62 points) (a) (8 points) Let X 1, X 2,..., X n be independent random variables uniformly distributed on
More informationTHE SOLOVAY STRASSEN TEST
THE SOLOVAY STRASSEN TEST KEITH CONRAD 1. Introduction The Jacobi symbol satisfies many formulas that the Legendre symbol does, such as these: for a, b Z and odd m, n Z +, (1) a b mod n ( a n ) = ( b n
More informationLecture 5: Probabilistic tools and Applications II
T-79.7003: Graphs and Networks Fall 2013 Lecture 5: Probabilistic tools and Applications II Lecturer: Charalampos E. Tsourakakis Oct. 11, 2013 5.1 Overview In the first part of today s lecture we will
More information1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).
CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 8 Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According to clinical trials,
More information[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]
Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the
More informationCSE548, AMS542: Analysis of Algorithms, Spring 2014 Date: May 12. Final In-Class Exam. ( 2:35 PM 3:50 PM : 75 Minutes )
CSE548, AMS54: Analysis of Algorithms, Spring 014 Date: May 1 Final In-Class Exam ( :35 PM 3:50 PM : 75 Minutes ) This exam will account for either 15% or 30% of your overall grade depending on your relative
More informationCS261: A Second Course in Algorithms Lecture #18: Five Essential Tools for the Analysis of Randomized Algorithms
CS261: A Second Course in Algorithms Lecture #18: Five Essential Tools for the Analysis of Randomized Algorithms Tim Roughgarden March 3, 2016 1 Preamble In CS109 and CS161, you learned some tricks of
More informationTail Inequalities Randomized Algorithms. Sariel Har-Peled. December 20, 2002
Tail Inequalities 497 - Randomized Algorithms Sariel Har-Peled December 0, 00 Wir mssen wissen, wir werden wissen (We must know, we shall know) David Hilbert 1 Tail Inequalities 1.1 The Chernoff Bound
More informationNotes taken by Graham Taylor. January 22, 2005
CSC4 - Linear Programming and Combinatorial Optimization Lecture : Different forms of LP. The algebraic objects behind LP. Basic Feasible Solutions Notes taken by Graham Taylor January, 5 Summary: We first
More informationRandom Graphs. EECS 126 (UC Berkeley) Spring 2019
Random Graphs EECS 126 (UC Bereley) Spring 2019 1 Introduction In this note, we will briefly introduce the subject of random graphs, also nown as Erdös-Rényi random graphs. Given a positive integer n and
More informationLearning Theory. Sridhar Mahadevan. University of Massachusetts. p. 1/38
Learning Theory Sridhar Mahadevan mahadeva@cs.umass.edu University of Massachusetts p. 1/38 Topics Probability theory meet machine learning Concentration inequalities: Chebyshev, Chernoff, Hoeffding, and
More informationProbability inequalities 11
Paninski, Intro. Math. Stats., October 5, 2005 29 Probability inequalities 11 There is an adage in probability that says that behind every limit theorem lies a probability inequality (i.e., a bound on
More informationSolutions to Problem Set 4
UC Berkeley, CS 174: Combinatorics and Discrete Probability (Fall 010 Solutions to Problem Set 4 1. (MU 5.4 In a lecture hall containing 100 people, you consider whether or not there are three people in
More informationVECTOR STUFF MATH 195, SECTION 59 (VIPUL NAIK)
VECTOR STUFF MATH 195, SECTION 59 (VIPUL NAIK) Corresponding material in the book: Sections 12.2, 12.3, 12.4. What students should definitely get: Vector as a n-tuple, special attention to case n = 3,
More information14.1 Finding frequent elements in stream
Chapter 14 Streaming Data Model 14.1 Finding frequent elements in stream A very useful statistics for many applications is to keep track of elements that occur more frequently. It can come in many flavours
More informationDiscrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations
EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of
More informationLecture 4: Probability and Discrete Random Variables
Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationProblem 1: (Chernoff Bounds via Negative Dependence - from MU Ex 5.15)
Problem 1: Chernoff Bounds via Negative Dependence - from MU Ex 5.15) While deriving lower bounds on the load of the maximum loaded bin when n balls are thrown in n bins, we saw the use of negative dependence.
More informationConstrained and Unconstrained Optimization Prof. Adrijit Goswami Department of Mathematics Indian Institute of Technology, Kharagpur
Constrained and Unconstrained Optimization Prof. Adrijit Goswami Department of Mathematics Indian Institute of Technology, Kharagpur Lecture - 01 Introduction to Optimization Today, we will start the constrained
More informationLecture 2: Minimax theorem, Impagliazzo Hard Core Lemma
Lecture 2: Minimax theorem, Impagliazzo Hard Core Lemma Topics in Pseudorandomness and Complexity Theory (Spring 207) Rutgers University Swastik Kopparty Scribe: Cole Franks Zero-sum games are two player
More informationLower Bounds for Testing Bipartiteness in Dense Graphs
Lower Bounds for Testing Bipartiteness in Dense Graphs Andrej Bogdanov Luca Trevisan Abstract We consider the problem of testing bipartiteness in the adjacency matrix model. The best known algorithm, due
More information12 Count-Min Sketch and Apriori Algorithm (and Bloom Filters)
12 Count-Min Sketch and Apriori Algorithm (and Bloom Filters) Many streaming algorithms use random hashing functions to compress data. They basically randomly map some data items on top of each other.
More informationAPPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of
CHAPTER III APPLICATIONS The eigenvalues are λ =, An orthonormal basis of eigenvectors consists of, The eigenvalues are λ =, A basis of eigenvectors consists of, 4 which are not perpendicular However,
More informationSome Review Problems for Exam 2: Solutions
Math 5366 Fall 017 Some Review Problems for Exam : Solutions 1 Find the coefficient of x 15 in each of the following: 1 (a) (1 x) 6 Solution: 1 (1 x) = ( ) k + 5 x k 6 k ( ) ( ) 0 0 so the coefficient
More informationImpagliazzo s Hardcore Lemma
Average Case Complexity February 8, 011 Impagliazzo s Hardcore Lemma ofessor: Valentine Kabanets Scribe: Hoda Akbari 1 Average-Case Hard Boolean Functions w.r.t. Circuits In this lecture, we are going
More information20.1 2SAT. CS125 Lecture 20 Fall 2016
CS125 Lecture 20 Fall 2016 20.1 2SAT We show yet another possible way to solve the 2SAT problem. Recall that the input to 2SAT is a logical expression that is the conunction (AND) of a set of clauses,
More informationGenerating Function Notes , Fall 2005, Prof. Peter Shor
Counting Change Generating Function Notes 80, Fall 00, Prof Peter Shor In this lecture, I m going to talk about generating functions We ve already seen an example of generating functions Recall when we
More informationKousha Etessami. U. of Edinburgh, UK. Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 7) 1 / 13
Discrete Mathematics & Mathematical Reasoning Chapter 7 (continued): Markov and Chebyshev s Inequalities; and Examples in probability: the birthday problem Kousha Etessami U. of Edinburgh, UK Kousha Etessami
More information