Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds

Similar documents
With high probability

Lecture 4: Two-point Sampling, Coupon Collector s problem

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).

Lecture 2: A Las Vegas Algorithm for finding the closest pair of points in the plane

Lecture 5: Two-point Sampling

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Matchings in Graphs. Definition 3 A matching N in G is said to be stable if it does not contain a blocking pair.

Ma/CS 6b Class 3: Stable Matchings

Homework 4 Solutions

Randomized Algorithms

CSE 548: Analysis of Algorithms. Lectures 18, 19, 20 & 21 ( Randomized Algorithms & High Probability Bounds )

Bipartite Matchings and Stable Marriage

Ma/CS 6b Class 3: Stable Matchings

Stable Matching Existence, Computation, Convergence Correlated Preferences. Stable Matching. Algorithmic Game Theory.

Chernoff Bounds. Theme: try to show that it is unlikely a random variable X is far away from its expectation.

Essentials on the Analysis of Randomized Algorithms

Game Theory: Lecture #5

STABLE MARRIAGE PROBLEM WITH TIES AND INCOMPLETE BOUNDED LENGTH PREFERENCE LIST UNDER SOCIAL STABILITY

Analysis of Algorithms Fall Some Representative Problems Stable Matching

Lecture 4: Sampling, Tail Inequalities

Lecture 5: January 30

The key is that there are two disjoint populations, and everyone in the market is on either one side or the other

CS 6901 (Applied Algorithms) Lecture 2

CSE 312 Final Review: Section AA

Matching Theory and the Allocation of Kidney Transplantations

Algorithm Design and Analysis

1. REPRESENTATIVE PROBLEMS

Expectation, inequalities and laws of large numbers

CSC2556. Lecture 5. Matching - Stable Matching - Kidney Exchange [Slides : Ariel D. Procaccia]

Consider a complete bipartite graph with sets A and B, each with n vertices.

Lecture 2 Sep 5, 2017

Notes 6 : First and second moment methods

THE METHOD OF CONDITIONAL PROBABILITIES: DERANDOMIZING THE PROBABILISTIC METHOD

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017

Random Variable. Pr(X = a) = Pr(s)

Tail Inequalities Randomized Algorithms. Sariel Har-Peled. December 20, 2002

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by:

PROBLEMS OF MARRIAGE Eugene Mukhin

APPROXIMATION ALGORITHMS FOR THE STABLE MATCHING PROBLEM

Lecture 4. P r[x > ce[x]] 1/c. = ap r[x = a] + a>ce[x] P r[x = a]

CPSC 320 Sample Final Examination December 2013

Stable Marriage with Ties and Bounded Length Preference Lists

An improved approximation algorithm for the stable marriage problem with one-sided ties

CSE525: Randomized Algorithms and Probabilistic Analysis April 2, Lecture 1

Cheating to Get Better Roommates in a Random Stable Matching

A (ln 4)-Approximation Algorithm for Maximum Stable Matching with One-Sided Ties and Incomplete Lists

Omega notation. Transitivity etc.

Lectures on Elementary Probability. William G. Faris

variance of independent variables: sum of variances So chebyshev predicts won t stray beyond stdev.

Matching: The Theory. Muriel Niederle Stanford and NBER. September 26, 2011

CS5314 Randomized Algorithms. Lecture 18: Probabilistic Method (De-randomization, Sample-and-Modify)

Lecture 1 : Probabilistic Method

ACO Comprehensive Exam October 14 and 15, 2013

Notes for Lecture Notes 2

Final Examination. Adrian Georgi Josh Karen Lee Min Nikos Tina. There are 12 problems totaling 150 points. Total time is 170 minutes.

1 Randomized complexity

Algorithm Design and Analysis

The Fitness Level Method with Tail Bounds

CS261: A Second Course in Algorithms Lecture #18: Five Essential Tools for the Analysis of Randomized Algorithms

Randomized algorithm

CS 598RM: Algorithmic Game Theory, Spring Practice Exam Solutions

Model Counting for Logical Theories

Randomized Complexity Classes; RP

Efficient Approximation for Restricted Biclique Cover Problems

Incentives in Large Random Two-Sided Markets

In a five-minute period, you get a certain number m of requests. Each needs to be served from one of your n servers.

Notes on Discrete Probability

1 Computational Problems

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

We set up the basic model of two-sided, one-to-one matching

Problem 1: (Chernoff Bounds via Negative Dependence - from MU Ex 5.15)

CS 125 Section #12 (More) Probability and Randomized Algorithms 11/24/14. For random numbers X which only take on nonnegative integer values, E(X) =

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

Approximation and Randomized Algorithms (ARA) Lecture 1, September 3, 2012

COOPERATIVE GAME THEORY: CORE AND SHAPLEY VALUE

Lecture 8: February 8

CPSC 320 Sample Solution, Reductions and Resident Matching: A Residentectomy

Twelfth Problem Assignment

CSE 525 Randomized Algorithms & Probabilistic Analysis Spring Lecture 3: April 9

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 2 Luca Trevisan August 29, 2017

Coupling. 2/3/2010 and 2/5/2010

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22

CS280, Spring 2004: Final

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003

Lecture 5: Probabilistic tools and Applications II

2. AXIOMATIC PROBABILITY

1 Definitions and Things You Know

Umans Complexity Theory Lectures

1 Exercises for lecture 1

CPSC 536N: Randomized Algorithms Term 2. Lecture 2

CS320 Algorithms: Theory and Practice. Course Introduction

25.1 Markov Chain Monte Carlo (MCMC)

1. REPRESENTATIVE PROBLEMS

The first bound is the strongest, the other two bounds are often easier to state and compute. Proof: Applying Markov's inequality, for any >0 we have

Solutions to Problem Set 4

Expectation of geometric distribution

4. Suppose that we roll two die and let X be equal to the maximum of the two rolls. Find P (X {1, 3, 5}) and draw the PMF for X.

Midterm 1. Your Exam Room: Name of Person Sitting on Your Left: Name of Person Sitting on Your Right: Name of Person Sitting in Front of You:

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Transcription:

Randomized Algorithms Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 1 / 38

Overview A1. The Principle of Deferred Decisions A2. The Proposal Algorithm for the Stable Marriage Problem B1. Chernoff Bounds B2. A Randomized Algorithm for Dominating Sets Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 2 / 38

A1. The Principle of Deferred Decisions The Clock Solitaire game: randomly shuffle a standard pack of 52 cards split the cards into 13 piles of 4 cards each; label piles as A, 2,..., 10, J, Q, K take first card from K pile take next card from pile X, where X is the value of the previous card taken repeat until: - either all cards removed ( win ) - or you get stuck ( lose ) We want to evaluate the probability of win. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 3 / 38

Key features - game termination Remark 1. The last card we take before the game ends (either winning or loosing) is a K. Proof: Let us assume that at iteration j we draw card X but pile X is empty (thus the game terminates). Let X K (i.e. we lose). Because pile X is empty and X K, we must have already drawn (prior to draw j) 4 X cards. But then we can not draw an X card at the jth iteration, a contradiction. Note: There is no contradiction if the last card is a K and all other cards have been already removed (in that case the game terminates with win). Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 4 / 38

Key features - win Remark 2. We win if the fourth K card is drawn at the 52 iteration. Proof: whenever we draw for the 1st, 2nd or 3rd time a K card, the game does not terminate because the K pile is not empty so we can continue (see remark 1). when the fourth K is drawn at the 52nd iteration then all cards are removed and the game s result is win. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 5 / 38

The probability of win Because of remark 2, it is: Pr{win} = Pr{4th K at the 52nd iteration} = = #game evolutions: 52nd card = 4th K #all game evolutions Note: Considering all possible game evolutions is a rather naive approach since we have to count all ways to partition the 52 cards into 13 distinct piles, with an ordering on the 4 cards in each pile. This complicates the probability evaluation because of the dependence introduced by each random draw of a card. we define another probability space that better captures the random dynamics of the game evolution. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 6 / 38

The principle of deferred decisions Basic idea: rather than fix (end enumerate) the entire set of potential random choices in advance, instead let the random choices unfold with the progress of the random experiment. In this particular game at each draw any card not drawn yet is equally likely to be drawn. A winning game corresponds to a dynamics where the first 51 random draws include 3 K cards exactly. This is equivalent to draw the 4th K at the 52nd iteration. So we forget how the first 51 draws came out and focus on the 52nd draw, which must be a K. But the latter probability is 1 13 because of symmetry (e.g. the type of the 52nd card is random uniform among all 13 types). Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 7 / 38

The probability of win Thus we have proved the following: Theorem: The probability of win at the clock solitaire is 1 13. An alternative approach: - we actually have 13x4=52 distinct positions (13 piles, 4 positions each) where 52 distinct cards are placed. This gives a total of 52! different placements. - each game evolution actually corresponds to an ordered permutation of the 52 cards. - The winning permutations are those where the 52nd card is a K (4 ways) and the 51 preceding cards are arbitrarily chosen (51!). Thus: Pr{win} = 4 51! 52! = 4 52 = 1 13 (the idea was to defer, i.e. first consider the last choice and then conditionally the previous ones!) In other words, the principle does not assume that the entire set of random choices is made in advance. Rather, at each step of the process we fix only the random choices that must be revealed. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 8 / 38

A2. The Proposal Algorithm for the Stable Matching Problem The Stable Matching Problem. Consider n women (w 1,..., w n ) and n men (m 1,..., m n ). A matching is a 1-1 correspondence between the men and the women (i.e. we assume monogamous, heterosexual matchings) Each person has a strict preference list of the members of the other sex. A matching is unstable iff there exist w i and m j such that: - w i and m j are not matched together - w i prefers m j to her match - m j prefers w i to his match a matching which is not unstable is stable Many applications (e.g. assigning teachers to schools they want to serve at, doctors to hospitals, etc.) Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 9 / 38

Questions does a stable matching always exists? (i.e. for all choices of preference lists?) can we find one efficiently? Answers: yes, there is at least one stable matching for every choice of preference lists we will prove this by providing an algorithm that finds a stable matching this algorithm is randomized (Las Vegas) and needs O(n ln n) time w.h.p. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 10 / 38

The Gale-Shapley Proposal Algorithm (I) Basic idea: man proposes, woman disposes. Each currently unattached man proposes to the woman he most desires and has not rejected him already. The woman accepts him if she is currently unattached or she prefers him to her current match. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 11 / 38

The Gale-Shapley Proposal Algorithm (II) Features: - Once a woman gets matched, she remains matched forever (though her mates may change) - The desirability of her mates (from her perspective) can only increase with time, i.e. at each step either a woman matches for the first time, or she matches to a more desired (to her) mate - Unmatched men always have (at least one) an unmatched woman to make proposals to. - Unmatched men can propose to currently matched women - Men can change status from unmatched to matched and then to unmatched (rejected) and so on, based on the proposals of other men and the womens choice. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 12 / 38

A more formal description let us assume some arbitrary ordering of the men let i the smallest value such that man m i is unmatched m i proposes to the most desirable woman (according to his own preference list) that has not already rejected him. she accepts him if either a) she is currently unmatched or b) she prefers him to her current match (in that case, her current match becomes unmatched). this is repeated until there are no unmatched men left. Questions: does the algorithm terminate? is the resulting matching stable? how much time it takes? Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 13 / 38

Is the algorithm well-defined? Lemma. Whenever there is an unmatched man m i, there is some woman he has not proposed to (so she cannot have rejected him in the past). Proof: Once a woman becomes matched, she never becomes unmatched in the future Since m i is unmatched currently, all women he has proposed to (if any) so far are matched. Thus, if m i has proposed to all women, then all women are matched, hence all men are matched too - a contradiction. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 14 / 38

Worst case time complexity Theorem. The algorithm terminates after O(n 2 ) iterations (proposals). Proof: For man m i, let t i the number of women m i could still potentially propose. n At each step (proposal), the sum t i decreases by 1 (three cases actually: a) get accepted by a matched woman so her current mate gets rejected and cannot propose her again b) get accepted by an unmatched woman so he cannot propose her again c) get rejected) n Initially t i = n 2, so the number of proposals is at most n 2. i=1 Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 15 / 38 i=1

Correctness Theorem: The matching found by the algorithm is stable. Proof: Let us assume the matching is unstable, so there is at least two pairs m i w j and m k w l, however with m i and w l preferring to be matched together. Since m i prefers w l to w j, he must have proposed to w l before he proposed to w j. But she rejected him, so she must prefer her current match m k to m i : a) either she already had a better match at the time m i proposed to her or b) she matched m i initially and then got a more desirable proposal. A contradiction. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 16 / 38

Average case analysis of the Proposal Algorithm Note: In randomized algorithms random choices are made when processing a fixed input. In average case analysis, the input is random and we analyze the time complexity (a random variable) of a deterministic algorithm In the matching problem, the input s randomness is introduced by assuming that the preference lists are random uniform. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 17 / 38

The Amnesiac version of the Proposal algorithm Note: A simplified modification of the Gale-Shapley algorithm: - At each step, m i proposes to a woman chosen uniformly at a random among all n women (including those he has been rejected by) This does not affect the output of the algorithm, since if m i was rejected by a woman, he will be rejected again if he proposes her again The Amnesiac algorithm thus performs more proposals since it includes some wasted rejects, so his expected running time is an upper bound on the time of the original algorithm Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 18 / 38

The expected time complexity Theorem If the preference lists are chosen in a uniform random manner, the expected number of proposals in the Gale-Shapley algorithm is at most O(n log n). Proof: Clearly the algorithm terminates once all women have received at least one proposal. So the matching random process is actually a coupon collectors problem, for which we have proved the following bound: Coupon Collectors: If m = n ln n + cn (for any constant c R) then the time T for collecting all n coupons obeys the following: Pr{T > m} = 1 e e c Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 19 / 38

The use of the principle of deferred decisions We have actually used the principle in the sense that we do not assume that the (random) preference lists are chosen in advance. In fact we somehow (in the average case analysis of the Gale-Shapley algorithm) assume that men do not know their preference lists and each time a man makes a proposal he picks a random woman. The only dependency (from the past proposals) left is eliminated in the Amnesiac algorithm by the wasted proposals to women having already rejected a man (i.e. we forget the past) Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 20 / 38

B1. The Chernoffs Bounds Tail Bounds on the probability of deviation from the expected value of a random variable. Focus on sums of independent, indicator (Bernoulli) random variables Such sums abstract quantitative results of random choices in randomized algorithms (e.g. the number of comparisons in the randomized quicksort algorithm, where for each two elements i, j the r.v. X ij is 1 if S (i), S (j) are compared (and 0 otherwise). The Chernoff bounds are exponential i.e. much sharper than the Markov, Chebyshev bounds. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 21 / 38

Poisson and Bernoulli trials Poisson trials: - repeated, independent trials - two possible outcomes in each trial ( success, failure ) - potentially different success probability p i in each trial i. - we take the sum of the corresponding indicator variables X i for each trial i (it measures the total number of successes). i.e. for { 1 i n : 1 (success), with probability p i X i = 0 (failure), with probability q i = 1 p i n X = X 1 + + X n = i=1 Bernoulli trials: Poisson trials when p i = p, i. (in that case X B(n, p) i.e. it follows the binomial distribution) Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 22 / 38 X i

The tail probability Clearly, the expected value (or mean) of X is: n µ = E(X) = Two important questions: i=1 (1) For a real number β > 0 what is the probability that X exceeds (1 + β)µ? (i.e. we seek a bound on the tail probability) this is useful in the analysis of randomized algorithms (i.e., to show that the probability of failure to achieve a certain expected performance is small). (2) How large must β be so that the tail probability is less than a certain value ɛ? (this is relevant in the design of the algorithm). p i Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 23 / 38

The Chernoff bound (for exceeding the mean) (I) Theorem 1. Let X i be n independent Poisson(p i ) trials, n n X = X i, µ = E(X) = p i. i=1 Then, for any β > 0, it is : [ Pr{X > (1 + β)µ} < i=1 ] µ e β (1+β) (1+β) Proof: For any positive t: Pr{X > (1 + β)µ} = Pr{e tx > e t(1+β)µ } By the Markov inequality: Pr{e tx e t(1+β)µ } < E[etX ] [ n ] n But E[e tx ] = E [ e t(x 1+ +X n ) ] = E i=1 e tx i = e 1. t(1+β)µ i=1 E[e tx i ]. 2 Now E[e tx i ] = e t p i + 1 (1 p i ) = 1 + p i (e t 1) e p i(e t 1) 3. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 24 / 38

The Chernoff bound (for exceeding the mean) (II) 2. 3. E[e tx ] = n i=1. 1 Pr{e tx e t(1+β)µ } < e(et 1)µ e t(1+β)µ e p i(e t 1) = e n i=1 p i(e t 1) = e (et 1)µ The right part is minimized when t = ln(1 + β) (note that t > 0 β > 0). t = ln(1 + β) Pr{X > (1 + β)µ} < e(eln(1+β) 1)µ [ ] µ e = βµ e = β (1+β) (1+β)µ (1+β) (1+β) e ln(1+β)(1+β)µ = Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 25 / 38

The various Chernoff bounds (I) Theorem 2. (Chernoff bound for exceeding the mean) For X i, X, µ as in Theorem 1 it is: β [0, 1] : Pr{X (1 + β)µ} e β2 µ 3 ( ) µ e Proof: β [0, 1] : β (1+β) e β 2 µ (1+β) 3 Theorem 3. (Chernoff bound for the left tail) For X i, X, µ as in Theorem 1 it is: β [0, 1] : Pr{X (1 β)µ} e β2 µ 2 Proof: See the MR book. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 26 / 38

The various Chernoff bounds (II) Theorem 4. (The combined Chernoff bound) For X i, X, µ as in Theorem 1 it is: β [0, 1] : Pr{X (1 ± β)µ} 1 2e β2 µ 3 Proof: Follows easily from Theorems 2, 3. Important remark: - µ (at an arbitrary slow rate) Pr{X (1 ± β)µ} 1 - µ = Ω(log n) Pr{X (1 ± β)µ} 1 2n γ, where γ > 1 i.e. a logarithmic mean guarantees a polynomially fast convergence to 1 of the concentration probability. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 27 / 38

B2. A Randomized Algorithm for Dominating Sets The dominating set problem: Find a subset of the vertices of a graph such that every vertex not in this set is adjacent to at least one vertex in it. Formally: V V (G) dominating set in G(V, E) iff u / V, v V : (u, v) E(G) The problem is important and well motivated from real networks, since the dominating set plays a central role in the graph. Obviously we want to find a dominating set which is as small as possible. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 28 / 38

The complexity of the problem / using randomness Finding a minimum dominating set is an NP-hard problem. Randomness can be used to attack the problem: - It has been shown that in G n,p (p = 1 2 ) random graphs, dominating set of size < log n w.h.p. (technique: Linearity of expectation + Markov inequality) - Also, w.h.p. dominating set of size log n (technique: the second moment method) - We will here present a randomized algorithm that w.h.p. finds a d.s. of size (1 + ɛ) log n (ɛ > 0 arbitrarily small), in polynomial time (thus, this algorithm is near-optimal). Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 29 / 38

The GREEDY-DS algorithm - basic idea choose a random vertex and put it in a dominating set under construction remove from graph all vertices adjacent to that vertex (they are covered by the vertex) repeat until the number of vertices left becomes small explicitly add those vertices to the dominating set under construction (obviously, the resulting set is a dominating set) Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 30 / 38

The algorithm - pseudo code ALGORITHM GREEDY-DS Input: a random instance G(V, E) of G n,p (p = 1 2 ) (1) i 0; V i V ; D (2) until V i ɛ log n do begin choose a random vertex u i V i V V i N i (N i : neighor vertices of u i ) D D {u i } (add the vertex to the evolving d.s.) i i + 1 V i V end (3) D D V i (add the vertices left) (4) return D Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 31 / 38

Analysis - informal in each repetition about half (because of p = 1 2 ) of the vertices left are covered by the random vertex and are removed from the graph thus, after log n repetitions the entire graph is almost covered and the number of vertices left drops to less than ɛ log n Then, these ɛ log n vertices are explicity added to the dominating set under construction The resulting set has size log n + ɛ log n = (1 + ɛ) log n and is obviously a dominating set. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 32 / 38

Two comments (1) when we choose a vertex, we expose its neighbours, so they can not be assumed random anymore (e.g. we know exactly their number, which is not a random variable anymore) however these exposed vertices are anyway removed from the graph, so the rest graph (the vertices left) remains random and the randomized analysis remains valid. (2) the Chernoff bound is used to show concentration of the number of covered vertices in each repetition around the expected number (which is easy to compute) as said, the bound is polynomially fast approaching 1 as long as the mean (of the vertices left) is logarithmic! This is why when this number gets logarithmic, the vertices left are explicitly added to the dominating set. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 33 / 38

Analysis - Chernoff bound Lemma 1 Let β [0, 1]. Then for any constant ɛ > 0 and given that V i ɛ log n it is: { } Pr N i (1 β) V i 2 1 n β2 4 ɛ Proof: The vertices covered at repetition i (set N i ) follow a Binomial distribution with parameters V i, 1 2. In other words, we have V i Bernoulli trials each one with success probability 1 2. Thus the mean µ i of their sum ( N i ) is bounded by the Chernoff bound as follows: Pr{ N i (1 β)µ i } 1 e β2 2 µ i = 1 e β2 V i 2 2 1 e β2 2 ɛ log n 2 = 1 n β2 4 ɛ Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 34 / 38

Analysis - time complexity Let the event E i = at the i-th random repetition it is N i (1 β) V i 2 Let the event E = E 1 E 2 E t until V t < ɛ log n Lemma 2: Assuming (probabilistically) event E the number t of repetitions of the random loop of the algorithm is at most (1 + ɛ ) log n, where ɛ > 0 constant. Proof: After repetition i: V i+1 = V i N i V i (1 β) 2 V i = (1 γ) V i where γ = 1 1 β. Recursively: V t (1 γ) t n, so for V t ɛ log n we need t log n log( 1 1 γ ) + Θ(log log n). We note that 1 1 γ = 2 1+β so by choosing (for any ɛ > 0) β = 2 ɛ 1+ɛ 1 we finally get t (1 + ɛ ) log n Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 35 / 38

Analysis - continued Lemma 3: For any constants β [0, 1], ɛ > 0 it is: Pr{E} 1 n β2 8 ɛ Proof: From previous lemmata, it is Pr{E} i Pr{E j} t n β2 4 ɛ (1 + ɛ ) log n n β2 4 ɛ n β2 8 ɛ Time complexity: Clearly, the algorithm needs time O((1 + ɛ )n log n) with probability at least 1 n β2 8 ɛ. The constructed dominating set has size at least (1 + ɛ + ɛ) log n Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 36 / 38

Summary (I) The algorithm constructs a near optimal dominating set (i.e. the approximation ratio to the optimal log n size is 1 + ɛ, where ɛ > 0 arbitrarily small). The time complexity is polynomial (both in expectation and w.h.p.) However the probability of good performance, although tending to 1, is not polynomially large since in the 1 n β2 8 ɛ bound the β2 8 ɛ constant is just positive, (not necessarily larger than 1) Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 37 / 38

Summary (II) In the full paper (S. Nikoletseas and P. Spirakis, Near-Optimal Dominating Sets in Dense Random Graphs in Polynomial Expected Time, in the Proceedings of the 19th International Workshop on Graph-Theoretic Concepts in Computer Science (WG)) includes an enhanced algorithm (of repetitive trials) boosting this probability to 1 n a, where a > 1. Sotiris Nikoletseas, Associate Professor Randomized Algorithms - Lecture 5 38 / 38