Lecture 2 February 8, 2016

Similar documents
Lecture 2: Concentration Bounds

Problem Set 2 Solutions

CS166 Handout 02 Spring 2018 April 3, 2018 Mathematical Terms and Identities

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

An Introduction to Randomized Algorithms

Lecture 5: April 17, 2013

Lecture 2: April 3, 2013

CS 330 Discussion - Probability

CS161 Handout 05 Summer 2013 July 10, 2013 Mathematical Terms and Identities

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

Math 155 (Lecture 3)

Advanced Stochastic Processes.

CS 171 Lecture Outline October 09, 2008

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Lecture 4: Unique-SAT, Parity-SAT, and Approximate Counting

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n,

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 21. Some Important Distributions

Lecture 2. The Lovász Local Lemma

Lecture Chapter 6: Convergence of Random Sequences

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018)

HOMEWORK 2 SOLUTIONS

ST5215: Advanced Statistical Theory

Learning Theory: Lecture Notes

Math 216A Notes, Week 5

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

Lecture 3: August 31

PH 425 Quantum Measurement and Spin Winter SPINS Lab 1

Approximations and more PMFs and PDFs

Lecture 4 February 16, 2016

Empirical Process Theory and Oracle Inequalities

Distribution of Random Samples & Limit theorems

Discrete Mathematics and Probability Theory Fall 2016 Seshia and Walrand Final Solutions

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

Lecture 12: November 13, 2018

Machine Learning Theory (CS 6783)

7.1 Convergence of sequences of random variables

STAT Homework 1 - Solutions

Metric Space Properties

CS / MCS 401 Homework 3 grader solutions

Lecture 7: Properties of Random Samples

Lecture 14: Graph Entropy

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002

Lecture 20: Multivariate convergence and the Central Limit Theorem

Element sampling: Part 2

Skip Lists. Presentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015 S 3 S S 1

Shannon s noiseless coding theorem

On Random Line Segments in the Unit Square

ACO Comprehensive Exam 9 October 2007 Student code A. 1. Graph Theory

Infinite Sequences and Series

Probability and Random Processes

Simulation. Two Rule For Inverting A Distribution Function

CS284A: Representations and Algorithms in Molecular Biology

Disjoint set (Union-Find)

1 Review and Overview

Application to Random Graphs

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3

Agnostic Learning and Concentration Inequalities

Lecture 12: September 27

Lecture 6: Coupon Collector s problem

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 15

The Random Walk For Dummies

Lecture 15: Learning Theory: Concentration Inequalities

Massachusetts Institute of Technology

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

Math 525: Lecture 5. January 18, 2018

Lecture 23 Rearrangement Inequality

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Solutions to HW Assignment 1

Lecture Notes 15 Hypothesis Testing (Chapter 10)

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8)

The Brunn-Minkowski Theorem and Influences of Boolean Variables

7.1 Convergence of sequences of random variables

It is always the case that unions, intersections, complements, and set differences are preserved by the inverse image of a function.

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

Lecture 19: Convergence

SECTION 1.5 : SUMMATION NOTATION + WORK WITH SEQUENCES

Lecture 3 : Random variables and their distributions

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators

Frequentist Inference

Exercise 4.3 Use the Continuity Theorem to prove the Cramér-Wold Theorem, Theorem. (1) φ a X(1).

Lecture 9: Pseudo-random generators against space bounded computation,

The standard deviation of the mean

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

Probability, Random Variables and Random Processes

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

6.895 Essential Coding Theory October 20, Lecture 11. This lecture is focused in comparisons of the following properties/parameters of a code:

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 19

This section is optional.

Ch3. Asymptotic Notation

Solutions to selected exercise of Randomized Algorithms

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

Lecture 7: October 18, 2017

Introduction to Probability. Ariel Yadin. Lecture 7

Basics of Probability Theory (for Theory of Computation courses)

BIRKHOFF ERGODIC THEOREM

page Suppose that S 0, 1 1, 2.

Lecture 6 Simple alternatives and the Neyman-Pearson lemma

Lecture 18: Sampling distributions

Transcription:

MIT 6.854/8.45: Advaced Algorithms Sprig 206 Prof. Akur Moitra Lecture 2 February 8, 206 Scribe: Calvi Huag, Lih V. Nguye I this lecture, we aalyze the problem of schedulig equal size tasks arrivig olie to differet machies. This could easily be solved with a cetralized server that kows which machies are occupied ad assigs all tasks to uique machies. Istead, we wat to see if it s possible to distribute the tasks without without such coordiatio amog the tasks. I this problem, the metric we care about is the maximum load o ay give machie. We will cosider two such approaches: aive radom assigmet, ad the power of two choice. Radom Assigmet Oe approach might be that each task chooses a machie idepedetly at radom. This is the classic balls-i-bis ball = task, bi = machie. Let z i be the umber of balls tasks assiged to bi machie i. There are a couple ways we might try to put a boud o the maximum load.. Markov s iequality Theorem. Markov s iequality: If X is a oegative radom variable, the for ay a > 0 P[X a] E[X] a Usig Markov s iequality, sice E[z i ] = o average, each machie will receive task we get, for some k, Pz i k k, ad usig the uio boud P i : z i k k, which is ot very useful for all practical values of k, it tells us that the probability is at most..2 Chebyshev s iequality Theorem 2. Chebyshev s iequality: Let X be a radom variable with expectatio µ ad variace σ 2. The P[ X µ kσ] k 2 Let y i,j be radom idicator variable, defied so that y i,j = if the j-th ball is throw ito the i-th bi, ad y i,j = 0 otherwise. Note that y i,j = with probability of, ad Vary i,j = 2.

The for z i = j= y i,j, we get E[z i ] = ad Varz i = Vary i, j= = / / 2 = <. Usig Chebyshev s iequality ad a uio boud: Pz i k Pz i µ kσ k 2 P i : z i k + k 2 ad so we ca place a costat boud o the probability that ay bi has more tha k balls, if k >..3 Direct aalysis I a more direct aalysis, we take a uio boud over all possible subsets of k balls that might be set to machie i: e k Pz i k Pall balls i S set to machie i = k k 2 k S [], S =k Here, we used the fact that k e k. k The boud e k k is o the order of whe k = Θ log log log, so we ca put a costat boud o the probability that o bi will have more tha Θ balls. log log log.4 Cheroff boud Theorem 3. Upper tail of Cheroff boud: Let X i be idepedet Beroulli radom variables with E[ X i ] = µ. The e δ P Xi + δµ + δ +δ µ exp δ 2 µ 2 + δ I this case the X i are whether or ot each ball falls ito a particular bi, so µ =, k = +δ, ad so we get P X i k ek, which gives the same result as i direct aalysis P X k k i k = O whe k = Θ. log log log 2

2 Power of Two Choices This is a simple variatio o the fully radom method: istead of pickig oe machie at radom, each task picks two radom machies, ad assigs itself to the machie with the lower load. This simple chage will drastically lower the expected max load from Θ to Θlog log. log log log We ll start with a heuristic justificatio of this improvemet, before givig a more rigorous proof. 2. Heuristic Aalysis Cosider this origorous aalysis to help uderstad where the Θlog log boud arises. Let β i be a upper boud o the umber of bis with at least i balls at the ed of the process. Now, cosider the umber of bis which will cotai at least i + balls. Whe a ball arrives, i order for it to icrease a bi to i +, both radom bis picked must have at least i balls. The probability of a give radom choice edig up that way would be at most β i ; sice that must 2. happe twice idepedetly of such a icrease would be at most We d the expect, over 2 all arrivals, at most βi β = i 2 bis cotaiig at least i + balls, ad thus set β i+ = β2 i. Let β 6 = 2e < 6 must be true by pigeohole priciple, the by iductio β i the lowest i c such that β ic < will be Olog log. βi 2e 2i 6, therefore Note that this heuristic is ot a formal proof, sice it assumes that the β i bouds always hold while i reality there must be some small failure probability ad sice we bouded the expectatio of β i+ but we will wat a boud that holds with high probability. However, the full proof is coceptually essetially the same. 2.2 Rigorous Proof Theorem 4. Suppose that balls are distributed to bis, accordig to the power of two choices method. The i the ed, with high probability the most loaded bi cotais at most Olog log balls. If d 2 choices are used istead, the the load is at most O log log balls. Proof: We shall give the proof for geeral d. First we defie a few otatios. Let t be the time right after the t-th ball is placed. Defie ht, the height of the t-th ball be the umber of balls i the same bi as the tth ball immediately after it is placed icludig the tth ball itself. Defie β 6 = 2e ad β i+ = eβd i for i 6. Note that we ca boud β d i+ c β i d for some costat c, so with j = Olog log we have β j <. Let B, p be a Beroulli radom variable deotig the total umber of heads resultig from flippig cois, each flip with probability of heads p. 3

Let ν i t be the umber of bis with load of at least i, ad µ i t be the umber of balls of height at least i. Obviously ν i t µ i t. Let E i be the evet that ν i β i. Sice β 6 = 2e < 6, the evet E 6 holds with certaity. Fix a iteger i. Let Y i be a radom biary variable such that Y t = iff ht i + ad ν i t β i That is, Y t = iff the height of the t-th ball is at least i +, ad at time t there are at most β i bis of load at least i. Let ω j be the bi selected by the j-th ball. The P[Y t = ω,.., ω t ] βd i d := p i The proof will also use the cocept of first-order stochastic domiace. A radom variable x is said to stochastically domiate aother radom variable y if for all k, Px k Py k We use the followig theorem o stochastic domiace: Theorem 5. Stochastic Domiace: If X, X 2,... X are i.i.d. radom variables Y i = f i X, X 2,... X i, ad PY i = X, X 2,... X i p, the i Y i is stochastically domiated by B, p. Usig the Stochastic Domiace theorem P[ Y t k] P[B, p i k] t= With respect to E i it follows µ i+ = t= Y t. Thus P[ν i+ k E i ] P[µ i+ k E i ] = P[ By the Cheroff boud, with k = β i+ t= Y t k E i ] P[ t= Y t k] P[B, p i k] P[ν i+ β i+ E i ] P[B, p i ep i ] e p i If p i 2 log, we ca rewrite the above expressio as P[ E i+ E i ] 2 P[E i. That implies with ] high probability ν i β i for large eough i. However, if p i 2 log, we show with high probability that there is o ball at height i + 2. Let i be the smallest i such that βd i d We have P[ν i + 6 log E i ] 2 log. Sice β i+6 2 di, by iductio i P[B, 2 log / 6 log ] P[E i ] 4 2 P[E i ] log log + O.

also Apply the iequality we obtai P[µ i +2 µ i + 6 log ] P[B, 6 log /d ] P[µ i + 6 log ] P[ E i+ ] P[ E i+ E i ] + P[ E i ] P[µ i +2 ] 6 d + i + 2 = O That is, with high probability there s o ball at height i + 2 = O proof. 6 log /d P[µ i + 6 log ] log log. That cocludes the 5