Essentials on the Analysis of Randomized Algorithms

Size: px
Start display at page:

Download "Essentials on the Analysis of Randomized Algorithms"

Transcription

1 Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random variables, techniques on approximating and bounding combinations and probabilities (emphasis on Hoeffding bound), the central limit theorem, the weak and strong law of large numbers, and fundamental problems that show these techniques in action. Basic definitions on Markov Chains are also presented. In the appendix we find the basic randomized algorithmic schemes, as well as an overview of the complexity classes where these algorithms fall. Most definitions and results are drawn from [BT02]. Basics Definition. (Probability Mass Function (PMF)). The PMF p X of a discrete random variable X is a function that describes the probability mass of each (discrete) value x that X can take; i.e. p X (x) = Pr[X = x].. (Discrete) Random Variables Definition.2 (Bernoulli Random Variable). X is a Bernoulli random variable that takes two values 0 and depending on the outcome of a random process (e.g. tossing a coin once). Its PMF is: { p, if x =, p X (x) = p, if x = 0. The expected value of X is E[X] = p, while the variance is Var[X] = p( p). Definition.3 (Binomial Random Variable). Y is a Binomial random variable with parameters and p that is constructed by Bernoulli random variables X,...,X, each of which is with probability p. It is defined as the sum Y = i= X i. Its PMF is: p Y (k) = Pr[Y = k] = ( k ) p k ( p) k, k = 0,,...,. The expected value of Y is E[Y] = p, while the variance is Var[Y] = p( p). Definition.4 (Geometric Random Variable). Given a sequence of Bernoulli random variables X,X 2,..., each of which is with probability p, Z is a Geometric random variable expressing the minimum i such that X i =. Its PMF is: p Z (k) = ( p) k p, k =,...,. The expected value of Z is E[Z] = p p, while the variance is Var[Z] = p. 2 Definition.5 (Poisson Random Variable). S is a Poisson random variable with parameter λ and PMF given by: λ λk p S (k) = e k!, k = 0,,...,. The expected value of S is E[S] = λ, and the variance is also Var[S] = λ. p Y (k) = k=0 p Z (k) = k= p S (k) = k=0

2 .2 Bernoulli process Informally it is a sequence of independent coin tosses. Definition.6 (Bernoulli process). It is a sequence X,X 2,... of independent Bernoulli random variables X i such that for every i it holds: { Pr[Xi = ] = Pr[success at the ith trial] = p Pr[X i = 0] = Pr[failure at the ith trial] = p 2 Approximating and Bounding In this section important tools on approximating and bounding probabilities will be explored. 2. The Cauchy-Schwartz Inequality ( n ) 2 ( n ) ( n x i y i x 2 i y 2 i i= i= i= ) () 2.2 Bounding Combinations Let < k < n, with k,n. Then: ( n ) k ( n ) < < k k ( ne ) k and k ( n e ) n < n! (2) 2.3 Common Approximations Exponential: ( x) e x. Poisson: The Poisson PMF with parameter λ is a good approximation for a binomial PMF with parameters and p, provided that λ = p, is very large, and p is very small. 2.4 Bounding Probabilities Union Bound: Let A,A 2,...,A be events in a probability space. Then [ ] Pr A i Pr[A i ] max {Pr[A i ]} (3) i i= i= The first inequality is equality for disjoint events A i. Markov s Inequality: Any non-negative random variable X satisfies: Pr[X α] E[X], α > 0. (4) α Chebyshev s Inequality: Let X be a r. v. with expected value µ and variance σ 2. Then: Pr[ X µ α] σ2, α > 0. (5) α2 Remark 2. (Chebyshev vs. Markov). The Chebyshev inequality tends to give better bounds than the Markov inequality, because it also uses information on the variance of X. 2

3 Theorem 2.2 (Weak Law of Large umbers). Let X,...,X be a sequence of independent identically distributed random variables, with expected value µ. For every ǫ > 0: ] Pr[ X i µ ǫ 0, as (6) i= Proof. Let X,...,X be a sequence of independent identically distributed random variables, with expected value µ and variance σ 2. Define the random variable Y = i= X i. By linearity of expectation we get E[Y] = i= E[X i] = µ. Since all the X i are independent, the variance is Var[Y] = 2 i= Var[X i] = σ2. We now apply Chebyshev s inequality and obtain Pr[ Y µ ǫ] σ2 ǫ, for any ǫ > Concentration and Tail Inequalities Proposition 2.3 (Chernoff - Hoeffding Bound [Hoe63]). Let X,...,X be independent random variables, each taking values in the range I = [α,β], and let µ denote the mean of their expectations. Then: [ ] Pr X i µ ǫ e 2ǫ2 /(β α) 2. (7) i= Assuming we want to bound the quantity above by δ, it is enough (β α) 2 2ǫ ln 2 δ. Some typical bounds obtained by the inequality are shown below: ǫ = Ω(/ ) δ , , 04 38, 56 (a) ǫ = = δ , 605, 7 9, 20, 34 3, 85, 5 (b) ǫ = = Figure : Typical lower bounds on when I = β α =. δ , , , 62 (a) ǫ = = δ , 420, 68 36, 84, , 262, 043 (b) ǫ = = Figure 2: Typical lower bounds on when I = β α = 2. Definition 2.4 (Martingale [AS08]). A martingale is a sequence X 0,...,X of random variables so that for 0 i < it holds: E[X i+ X i,x i,...,x 0 ] = X i. fair gambling Proposition 2.5 (Azuma s Inequality [Mau79, AS08]). Let c = X 0,...,X be a martingale Azuma: c = 0 with X i+ X i 0 i <. Then: Pr[ X c > λ ] < 2e λ2 /2 For λ = 2ln e λ2 /2 =. Check tbl. for some typical approximate values. (8) 3

4 λ e λ2 / Table : Typical bounds obtained for some λ by Azuma s inequality Lower bounds on Tails [AS08, appendix] In section 2.4. we obtained upper bounds on Pr[X > α] which were of the form e cǫ2. We can also obtain lower bounds on Pr[X > α]; typically we get Pr[X > α] = Ω(e cǫ2 e dǫ ). 3 Fundamental Problems 3. Coins A coin has two sides; H and T. Set Pr[H] = p and Pr[T] = p, where p is a fixed number. 3.. Games with coins How many heads H: Tossing a coin times and recording the number of times H appeared is a Binomial random variable. First H: Tossing a coin until H comes up is a Geometric random variable. Both H and T: The weighted sum p( + /( p)) + ( p)( + /p) = /p + p/( p) expresses the expected amount of coin tosses in order to observe both H and T. Fair coin: After coin tosses, we observe H T + λ or H T λ with probability bounded by Azuma s Inequality; eq. (8). 3.2 The Coupon Collector s Problem Given coupons what is the expected amount of trials in order to observe all of them? i.e. we are drawing coupons with replacement. Let T be the total time to observe all coupons, and let t i denote the time needed to collect coupon i after i coupons have been collected; i.e. T = i= t i. ote that p i = ( (i ))/, and each t i is a geometric r.v. By linearity of expectation we get: E[T] = E[t i ] = = p i i= i= i= (i ) = i= where H is the th harmonic number. For large we get: + i = i= i = H (9) E[T] = H = ln + γ o ( / 4), (0) where γ = lim (H ln) = is the Euler-Mascheroni constant [Wei]. Hence, E[T] = O(ln ). All the t i are independent, so the variance is: Var[T] = Var[t i ] = i= < 2 i= i= p i p 2 i This problem is also known as Balls in Bins Problem. < p 2 i= i = i= 2 ( + i) 2 = 2 i 2 = 2 π2 6 < 2 2 () i= i 2 4

5 Applying Chebyshev s inequality (5) to equations (9) and () we get Pr[ T H λ] < 2 λ 2. (2) 3.2. Generalized Coupon Collector s Problem Again we have coupons, but this time we are interested in a subset K of them ( K = k < ). What we want is the minimum number of trials so that the probability some member of K is missed is less than some predefined value η. Let T be the number of trials. Then a member of K is missed in all T trials with probability ( /) T. By the Union Bound (eq. (3)) this probability is upper bounded by k( /) T. By the exponential approximation this probability is upper bounded by ke T/. By requiring this probability being less than η we get: 4 Central Limit Theorem T > ln(k/η). (3) Theorem 4. (Central Limit Theorem). Let X,X 2,...,X be a sequence of independent identically distributed random variables with common µ and variance σ 2, and define i= Z = X i µ σ. (4) Then, the CDF of Z converges to the standard normal CDF E[Z ] = 0 Φ(z) = z e x2 /2 dx, 2π (5) Var[Z ] = in the sense that lim Pr[Z z] = Φ(z), for all z. (6) Proposition 4.2 (De Moivre - Laplace Approximation to the Binomial). If S is a binomial random variable with parameters p and, is large, and κ,λ, then ( ) ( ) λ + /2 p κ /2 p Pr[κ S λ] Φ Φ (7) p( p) p( p) Remark 4.3 (Quality). The closer p is to 0 or, the larger the so that the approximation is good. When p 0.5 around 40 to 50 already gives very good results. Theorem 4.4 (Strong Law of Large umbers). Let X,X 2,...,X be a sequence of independent identically distributed random variables with mean µ. Then, the sequence of sample means M = i= X i converges to µ, with probability, in the sense that 4. Applications Pr [ lim ] X i = µ =. (8) i= Example (Coins Revisited). We toss a coin 25 times and 20 times we observe H. What is the probability of this event, given that the coin is fair? A direct computation yields ( 25 20) Chebyshev yields Pr[ X ] = /9 = 0.. The Hoeffding bound gives Pr[ M ] e = 5

6 e 4.5 = Azuma s inequality gives double of what ) the Hoeffding bound( gives. The ) Central Limit Theorem gives Pr[S c] Φ so, Pr[S 25 9] Φ 6.25 = ( c p σ Φ ( ) = Φ(2.6) = So, the requested probability is less than = Using the De Moivre approximation we can compute directly Pr(S 25 = 20) Φ ( ) Φ ( ) = Φ(3.2) Φ(2.8) = = Example 2 (Lower Bound on Iterations). Assume we have a biased coin which gives rise to H with probability p and we want to estimate this value within 0.0 with probability at least 0.9. The Chebyshev inequality gives Pr[ M p 0.0] p( p) (0.0) and we want that bounded by 0.9 = 0.. This gives = 25,000. The Hoeffding bound gives =,53. In the case of the Central Limit theorem we observe that the variance of M p σ 2 is p( p)/ /(4). Hence, z = ǫ/(/(2 )) = 2ǫ. So we get: Pr[ M p deviation 0.0] 2Pr[M p 0.0] 2( Φ(2 0.0 )) 0.. This implies Φ(0.02 ) 0.95 = Φ(.645) (82.25) 2 = 6, Markov Chains [MR95] Definition 5. (Markov Chain:). A Markov Chain M is a discrete-time stochastic process defined over a set of states S in terms of a S S matrix P of transition probabilities. The set S is either finite or countably infinite. M is in only one state at a time. State transitions occur at time-steps t =,2,... The entry P ij denotes the probability that the next state will be j, given that M is currently at state i. ote that P ij [0,], i,j S and j P ij =. Remark 5.2 (Memorylessness Property:). The next state of M depends only on its current state. 5. otation and Conventions Definition 5.3 (t-step transition probability:). We denote as P (t) ij = Pr[X t = j X 0 = i]. Definition 5.4 (First transition into state j at time t:). is denoted by r t ij and is given by r (t) ij = Pr[X t = j, and for s t,x s j X 0 = i]. Definition 5.5 (Transition into state j at some time t > 0:). is denoted by f ij and is given by f ij = t>0 r(t) ij. Definition 5.6 (Expected # of steps to reach j starting from i:). is denoted by h ij and is given by { h ij = t>0 tr(t) ij, if f ij =,, otherwise. 5.2 Definitions and a theorem The states of M can be classified as: Transient: f ii < ( h ij = ). Persistent: f ii =. These can be further classified as: ull persistent: h ii =. on-null persistent: h ii <. 6

7 Definition 5.7 (Strong component C:). of a directed graph G, is a maximal subgraph of G such that there exists a path from i to j and back to i for every pair of vertices i,j C. Final: There is no edge so that we can leave G. Definition 5.8 (Irreducible Markov chain:). G consists of a single strong component. Definition 5.9 (State probability vector). q (t) = (q (t),q(t) 2,...,q(t) n ) probability that M is in state i at time t. Definition 5.0 (Stationary distribution). for M with transition matrix P is a probability distribution π such that π = πp. Definition 5. (Periodicity T of a state i). guarantees another visit to state i after a+ti steps for some i 0. A state is periodic if T >, and aperiodic otherwise. A Markov Chain M is aperiodic, if every state is aperiodic. Definition 5.2 (Ergodic state). is one that is aperiodic and non-null persistent. Definition 5.3 (Ergodic Markov Chain). is one in which all states are ergodic. Theorem 5.4 (Fundamental Theorem of Markov Chains). Any irreducible, finite, and aperiodic Markov Chain M has the following properties:. All states are ergodic. 2. There is a unique stationary distribution π, such that π i > 0, i {,...,n}. 3. For all i {,...,n} : f ii = and h ii = π i. 4. Let (i,t) be the number of times M visits state i in t steps. Then, lim t (i,t) t = π i. row vector 6 A glimpse beyond Random walks on graphs and expanders, machine learning, random number generators, parallel computation, probabilistic method. References [AS08] oga Alon and Joel Spencer. The Probabilistic Method. John Wiley, [BT02] Dimitri P. Bertsekas and John. Tsitsiklis. Introduction to Probability. Athena Scientific, [Hoe63] Wassily Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58(30):3 30, March 963. [Mau79] B. Maurey. Construction de suites symétriques. C. R. Acad. Sci. Paris, 288:679 68, 979. [MR95] Rajeev Motwani and Prabhakar Raghavan. Randomized Algorithms. Cambridge University Press, th printing, [Wei] Eric W. Weisstein. Euler-Mascheroni constant. Wolfram MathWorld. 7

8 A Basic Randomized Algorithmic Schemes Definition A. (Monte Carlo:). Monte Carlo (MC) algorithms exploit randomness in order to solve problems. The idea is that successive iterations of the core loop of the algorithm give result(s) which are independent of the previous runs. They can be classified as having one-sided error, or a two-sided error. For example, assume you have an algorithm A that decides whether x belongs in a language L, so that the answer we get is: x L Pr[A(x) accepts] p, x L Pr[A(x) accepts] = 0. This is an example of a one-sided error algorithm. A two-sided error algorithm arises if the probability of accepting an input x, when in fact x L, is non-zero. Definition A.2 (Las Vegas:). Las Vegas algorithms are Monte Carlo algorithms which never make a mistake on the result. An example of such an algorithm is quicksort (RandQS). ote that the running time of the Las Vegas algorithms depends on the input. B Complexity Classes on Randomized Algorithms Definition B. (Class RP:). RP (Randomized Polynomial time) algorithms are one-sided error Monte Carlo algorithms, that can err only when x L. Usually p = 2, but choice is arbitrary. Definition B.2 (Class ZPP:). ZPP (Zero-error Probabilistic Polynomial time) algorithms are algorithms that belong in RP co-rp. ote that Las Vegas algorithms belong in this class. Definition B.3 (Class BPP:). BPP (Bounded-error Probabilistic Polynomial time) algorithms are two-sided error Monte Carlo algorithms of the following form: x L Pr[A(x) accepts] 3 4, x L Pr[A(x) accepts] 4. Definition B.4 (Class RLP:). Class RLP is a subclass of RP, in which the algorithms use O(lg n) workspace in worst case. RandQS ZPP B. Categorizing Randomized Complexity Classes. P RP P. 2. RP BPP PP. 3. P PP PSPACE. 8

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

Disjointness and Additivity

Disjointness and Additivity Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

Approximate Counting and Markov Chain Monte Carlo

Approximate Counting and Markov Chain Monte Carlo Approximate Counting and Markov Chain Monte Carlo A Randomized Approach Arindam Pal Department of Computer Science and Engineering Indian Institute of Technology Delhi March 18, 2011 April 8, 2011 Arindam

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Lecture 2 Sep 5, 2017

Lecture 2 Sep 5, 2017 CS 388R: Randomized Algorithms Fall 2017 Lecture 2 Sep 5, 2017 Prof. Eric Price Scribe: V. Orestis Papadigenopoulos and Patrick Rall NOTE: THESE NOTES HAVE NOT BEEN EDITED OR CHECKED FOR CORRECTNESS 1

More information

Random Variable. Pr(X = a) = Pr(s)

Random Variable. Pr(X = a) = Pr(s) Random Variable Definition A random variable X on a sample space Ω is a real-valued function on Ω; that is, X : Ω R. A discrete random variable is a random variable that takes on only a finite or countably

More information

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Random Variable Discrete Random

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

Eleventh Problem Assignment

Eleventh Problem Assignment EECS April, 27 PROBLEM (2 points) The outcomes of successive flips of a particular coin are dependent and are found to be described fully by the conditional probabilities P(H n+ H n ) = P(T n+ T n ) =

More information

Probability & Computing

Probability & Computing Probability & Computing Stochastic Process time t {X t t 2 T } state space Ω X t 2 state x 2 discrete time: T is countable T = {0,, 2,...} discrete space: Ω is finite or countably infinite X 0,X,X 2,...

More information

Twelfth Problem Assignment

Twelfth Problem Assignment EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X

More information

Lecture 4: Two-point Sampling, Coupon Collector s problem

Lecture 4: Two-point Sampling, Coupon Collector s problem Randomized Algorithms Lecture 4: Two-point Sampling, Coupon Collector s problem Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms

More information

Model Counting for Logical Theories

Model Counting for Logical Theories Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) = Math 5. Rumbos Fall 07 Solutions to Review Problems for Exam. A bowl contains 5 chips of the same size and shape. Two chips are red and the other three are blue. Draw three chips from the bowl at random,

More information

Discrete Random Variables

Discrete Random Variables Chapter 5 Discrete Random Variables Suppose that an experiment and a sample space are given. A random variable is a real-valued function of the outcome of the experiment. In other words, the random variable

More information

Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds

Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Randomized Algorithms Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized

More information

1 Ways to Describe a Stochastic Process

1 Ways to Describe a Stochastic Process purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased

More information

Lecture 20 : Markov Chains

Lecture 20 : Markov Chains CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called

More information

CS155: Probability and Computing: Randomized Algorithms and Probabilistic Analysis

CS155: Probability and Computing: Randomized Algorithms and Probabilistic Analysis CS155: Probability and Computing: Randomized Algorithms and Probabilistic Analysis Eli Upfal Eli Upfal@brown.edu Office: 319 TA s: Lorenzo De Stefani and Sorin Vatasoiu cs155tas@cs.brown.edu It is remarkable

More information

Markov chains. Randomness and Computation. Markov chains. Markov processes

Markov chains. Randomness and Computation. Markov chains. Markov processes Markov chains Randomness and Computation or, Randomized Algorithms Mary Cryan School of Informatics University of Edinburgh Definition (Definition 7) A discrete-time stochastic process on the state space

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

The Markov Chain Monte Carlo Method

The Markov Chain Monte Carlo Method The Markov Chain Monte Carlo Method Idea: define an ergodic Markov chain whose stationary distribution is the desired probability distribution. Let X 0, X 1, X 2,..., X n be the run of the chain. The Markov

More information

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials). Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested

More information

Northwestern University Department of Electrical Engineering and Computer Science

Northwestern University Department of Electrical Engineering and Computer Science Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability

More information

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( ) Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr Pr = Pr Pr Pr() Pr Pr. We are given three coins and are told that two of the coins are fair and the

More information

1 Stat 605. Homework I. Due Feb. 1, 2011

1 Stat 605. Homework I. Due Feb. 1, 2011 The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due

More information

Bandits, Experts, and Games

Bandits, Experts, and Games Bandits, Experts, and Games CMSC 858G Fall 2016 University of Maryland Intro to Probability* Alex Slivkins Microsoft Research NYC * Many of the slides adopted from Ron Jin and Mohammad Hajiaghayi Outline

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Markov Processes Hamid R. Rabiee

Markov Processes Hamid R. Rabiee Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete

More information

MARKOV PROCESSES. Valerio Di Valerio

MARKOV PROCESSES. Valerio Di Valerio MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Probability. Lecture Notes. Adolfo J. Rumbos

Probability. Lecture Notes. Adolfo J. Rumbos Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................

More information

Stochastic Models in Computer Science A Tutorial

Stochastic Models in Computer Science A Tutorial Stochastic Models in Computer Science A Tutorial Dr. Snehanshu Saha Department of Computer Science PESIT BSC, Bengaluru WCI 2015 - August 10 to August 13 1 Introduction 2 Random Variable 3 Introduction

More information

Randomized algorithm

Randomized algorithm Tutorial 4 Joyce 2009-11-24 Outline Solution to Midterm Question 1 Question 2 Question 1 Question 2 Question 3 Question 4 Question 5 Solution to Midterm Solution to Midterm Solution to Midterm Question

More information

Notes on Discrete Probability

Notes on Discrete Probability Columbia University Handout 3 W4231: Analysis of Algorithms September 21, 1999 Professor Luca Trevisan Notes on Discrete Probability The following notes cover, mostly without proofs, the basic notions

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

Problem Points S C O R E Total: 120

Problem Points S C O R E Total: 120 PSTAT 160 A Final Exam Solution December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain

More information

Lectures on Markov Chains

Lectures on Markov Chains Lectures on Markov Chains David M. McClendon Department of Mathematics Ferris State University 2016 edition 1 Contents Contents 2 1 Markov chains 4 1.1 The definition of a Markov chain.....................

More information

Markov Chains (Part 3)

Markov Chains (Part 3) Markov Chains (Part 3) State Classification Markov Chains - State Classification Accessibility State j is accessible from state i if p ij (n) > for some n>=, meaning that starting at state i, there is

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms 南京大学 尹一通 Martingales Definition: A sequence of random variables X 0, X 1,... is a martingale if for all i > 0, E[X i X 0,...,X i1 ] = X i1 x 0, x 1,...,x i1, E[X i X 0 = x 0, X 1

More information

STA 711: Probability & Measure Theory Robert L. Wolpert

STA 711: Probability & Measure Theory Robert L. Wolpert STA 711: Probability & Measure Theory Robert L. Wolpert 6 Independence 6.1 Independent Events A collection of events {A i } F in a probability space (Ω,F,P) is called independent if P[ i I A i ] = P[A

More information

Topic 3: The Expectation of a Random Variable

Topic 3: The Expectation of a Random Variable Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

Mathematical Methods for Computer Science

Mathematical Methods for Computer Science Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge

More information

Name: Firas Rassoul-Agha

Name: Firas Rassoul-Agha Midterm 1 - Math 5010 - Spring 016 Name: Firas Rassoul-Agha Solve the following 4 problems. You have to clearly explain your solution. The answer carries no points. Only the work does. CALCULATORS ARE

More information

MATH Solutions to Probability Exercises

MATH Solutions to Probability Exercises MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe

More information

X = X X n, + X 2

X = X X n, + X 2 CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk

More information

Sample Spaces, Random Variables

Sample Spaces, Random Variables Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted

More information

1 Random Variable: Topics

1 Random Variable: Topics Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?

More information

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line Random Variable Random variable is a mapping that maps each outcome s in the sample space to a unique real number,. s s : outcome Sample Space Real Line Eamples Toss a coin. Define the random variable

More information

CS145: Probability & Computing

CS145: Probability & Computing CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,

More information

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by:

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by: Chapter 8 Probability 8. Preliminaries Definition (Sample Space). A Sample Space, Ω, is the set of all possible outcomes of an experiment. Such a sample space is considered discrete if Ω has finite cardinality.

More information

CSE 312 Final Review: Section AA

CSE 312 Final Review: Section AA CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material

More information

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution

More information

Some Definition and Example of Markov Chain

Some Definition and Example of Markov Chain Some Definition and Example of Markov Chain Bowen Dai The Ohio State University April 5 th 2016 Introduction Definition and Notation Simple example of Markov Chain Aim Have some taste of Markov Chain and

More information

CSE 525 Randomized Algorithms & Probabilistic Analysis Spring Lecture 3: April 9

CSE 525 Randomized Algorithms & Probabilistic Analysis Spring Lecture 3: April 9 CSE 55 Randomized Algorithms & Probabilistic Analysis Spring 01 Lecture : April 9 Lecturer: Anna Karlin Scribe: Tyler Rigsby & John MacKinnon.1 Kinds of randomization in algorithms So far in our discussion

More information

An Introduction to Randomized algorithms

An Introduction to Randomized algorithms An Introduction to Randomized algorithms C.R. Subramanian The Institute of Mathematical Sciences, Chennai. Expository talk presented at the Research Promotion Workshop on Introduction to Geometric and

More information

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO QUESTION BOOKLET EE 26 Spring 2006 Final Exam Wednesday, May 7, 8am am DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the final. The final consists of five

More information

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

Lecture 4: Inequalities and Asymptotic Estimates

Lecture 4: Inequalities and Asymptotic Estimates CSE 713: Random Graphs and Applications SUNY at Buffalo, Fall 003 Lecturer: Hung Q. Ngo Scribe: Hung Q. Ngo Lecture 4: Inequalities and Asymptotic Estimates We draw materials from [, 5, 8 10, 17, 18].

More information

Expectation of Random Variables

Expectation of Random Variables 1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete

More information

Problem Points S C O R E Total: 120

Problem Points S C O R E Total: 120 PSTAT 160 A Final Exam December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain with the

More information

Chernoff Bounds. Theme: try to show that it is unlikely a random variable X is far away from its expectation.

Chernoff Bounds. Theme: try to show that it is unlikely a random variable X is far away from its expectation. Chernoff Bounds Theme: try to show that it is unlikely a random variable X is far away from its expectation. The more you know about X, the better the bound you obtain. Markov s inequality: use E[X ] Chebyshev

More information

Things to remember when learning probability distributions:

Things to remember when learning probability distributions: SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions

More information

MATH Notebook 5 Fall 2018/2019

MATH Notebook 5 Fall 2018/2019 MATH442601 2 Notebook 5 Fall 2018/2019 prepared by Professor Jenny Baglivo c Copyright 2004-2019 by Jenny A. Baglivo. All Rights Reserved. 5 MATH442601 2 Notebook 5 3 5.1 Sequences of IID Random Variables.............................

More information

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t 2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition

More information

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 Justify your answers; show your work. 1. A sequence of Events. (10 points) Let {B n : n 1} be a sequence of events in

More information

Assignment 4: Solutions

Assignment 4: Solutions Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each

More information

Expected Value 7/7/2006

Expected Value 7/7/2006 Expected Value 7/7/2006 Definition Let X be a numerically-valued discrete random variable with sample space Ω and distribution function m(x). The expected value E(X) is defined by E(X) = x Ω x m(x), provided

More information

25.1 Markov Chain Monte Carlo (MCMC)

25.1 Markov Chain Monte Carlo (MCMC) CS880: Approximations Algorithms Scribe: Dave Andrzejewski Lecturer: Shuchi Chawla Topic: Approx counting/sampling, MCMC methods Date: 4/4/07 The previous lecture showed that, for self-reducible problems,

More information

6.262: Discrete Stochastic Processes 2/2/11. Lecture 1: Introduction and Probability review

6.262: Discrete Stochastic Processes 2/2/11. Lecture 1: Introduction and Probability review 6.262: Discrete Stochastic Processes 2/2/11 Lecture 1: Introduction and Probability review Outline: Probability in the real world Probability as a branch of mathematics Discrete stochastic processes Processes

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Andreas Klappenecker Texas A&M University Lecture notes of a course given in Fall 2003. Preliminary draft. c 2003 by Andreas Klappenecker. All rights reserved. Chapter 2 Probability

More information

Notes 1 : Measure-theoretic foundations I

Notes 1 : Measure-theoretic foundations I Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Randomized Computation

Randomized Computation Randomized Computation Slides based on S.Aurora, B.Barak. Complexity Theory: A Modern Approach. Ahto Buldas Ahto.Buldas@ut.ee We do not assume anything about the distribution of the instances of the problem

More information

p. 4-1 Random Variables

p. 4-1 Random Variables Random Variables A Motivating Example Experiment: Sample k students without replacement from the population of all n students (labeled as 1, 2,, n, respectively) in our class. = {all combinations} = {{i

More information

Handout 5. α a1 a n. }, where. xi if a i = 1 1 if a i = 0.

Handout 5. α a1 a n. }, where. xi if a i = 1 1 if a i = 0. Notes on Complexity Theory Last updated: October, 2005 Jonathan Katz Handout 5 1 An Improved Upper-Bound on Circuit Size Here we show the result promised in the previous lecture regarding an upper-bound

More information

Lecture 4: Sampling, Tail Inequalities

Lecture 4: Sampling, Tail Inequalities Lecture 4: Sampling, Tail Inequalities Variance and Covariance Moment and Deviation Concentration and Tail Inequalities Sampling and Estimation c Hung Q. Ngo (SUNY at Buffalo) CSE 694 A Fun Course 1 /

More information

Lectures 6: Degree Distributions and Concentration Inequalities

Lectures 6: Degree Distributions and Concentration Inequalities University of Washington Lecturer: Abraham Flaxman/Vahab S Mirrokni CSE599m: Algorithms and Economics of Networks April 13 th, 007 Scribe: Ben Birnbaum Lectures 6: Degree Distributions and Concentration

More information

U.C. Berkeley CS278: Computational Complexity Professor Luca Trevisan 1/29/2002. Notes for Lecture 3

U.C. Berkeley CS278: Computational Complexity Professor Luca Trevisan 1/29/2002. Notes for Lecture 3 U.C. Bereley CS278: Computational Complexity Handout N3 Professor Luca Trevisan 1/29/2002 Notes for Lecture 3 In this lecture we will define the probabilistic complexity classes BPP, RP, ZPP and we will

More information

S n = x + X 1 + X X n.

S n = x + X 1 + X X n. 0 Lecture 0 0. Gambler Ruin Problem Let X be a payoff if a coin toss game such that P(X = ) = P(X = ) = /2. Suppose you start with x dollars and play the game n times. Let X,X 2,...,X n be payoffs in each

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Randomized Complexity Classes; RP

Randomized Complexity Classes; RP Randomized Complexity Classes; RP Let N be a polynomial-time precise NTM that runs in time p(n) and has 2 nondeterministic choices at each step. N is a polynomial Monte Carlo Turing machine for a language

More information

Math 180B Problem Set 3

Math 180B Problem Set 3 Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018 ECE534, Spring 2018: s for Problem Set #4 Due Friday April 6, 2018 1. MMSE Estimation, Data Processing and Innovations The random variables X, Y, Z on a common probability space (Ω, F, P ) are said to

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

6.842 Randomness and Computation March 3, Lecture 8

6.842 Randomness and Computation March 3, Lecture 8 6.84 Randomness and Computation March 3, 04 Lecture 8 Lecturer: Ronitt Rubinfeld Scribe: Daniel Grier Useful Linear Algebra Let v = (v, v,..., v n ) be a non-zero n-dimensional row vector and P an n n

More information

Introduction to probability theory

Introduction to probability theory Introduction to probability theory Fátima Sánchez Cabo Institute for Genomics and Bioinformatics, TUGraz f.sanchezcabo@tugraz.at 07/03/2007 - p. 1/35 Outline Random and conditional probability (7 March)

More information

COUPON COLLECTING. MARK BROWN Department of Mathematics City College CUNY New York, NY

COUPON COLLECTING. MARK BROWN Department of Mathematics City College CUNY New York, NY Probability in the Engineering and Informational Sciences, 22, 2008, 221 229. Printed in the U.S.A. doi: 10.1017/S0269964808000132 COUPON COLLECTING MARK ROWN Department of Mathematics City College CUNY

More information

EE514A Information Theory I Fall 2013

EE514A Information Theory I Fall 2013 EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/

More information

Pr[X = s Y = t] = Pr[X = s] Pr[Y = t]

Pr[X = s Y = t] = Pr[X = s] Pr[Y = t] Homework 4 By: John Steinberger Problem 1. Recall that a real n n matrix A is positive semidefinite if A is symmetric and x T Ax 0 for all x R n. Assume A is a real n n matrix. Show TFAE 1 : (a) A is positive

More information

Probability Theory Overview and Analysis of Randomized Algorithms

Probability Theory Overview and Analysis of Randomized Algorithms Probability Theory Overview and Analysis of Randomized Algorithms Analysis of Algorithms Prepared by John Reif, Ph.D. Probability Theory Topics a) Random Variables: Binomial and Geometric b) Useful Probabilistic

More information

11.1 Set Cover ILP formulation of set cover Deterministic rounding

11.1 Set Cover ILP formulation of set cover Deterministic rounding CS787: Advanced Algorithms Lecture 11: Randomized Rounding, Concentration Bounds In this lecture we will see some more examples of approximation algorithms based on LP relaxations. This time we will use

More information

ECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding.

ECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. NAME: ECE 302 Division MWF 0:30-:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. If you are not in Prof. Pollak s section, you may not take this

More information

EECS 495: Randomized Algorithms Lecture 14 Random Walks. j p ij = 1. Pr[X t+1 = j X 0 = i 0,..., X t 1 = i t 1, X t = i] = Pr[X t+

EECS 495: Randomized Algorithms Lecture 14 Random Walks. j p ij = 1. Pr[X t+1 = j X 0 = i 0,..., X t 1 = i t 1, X t = i] = Pr[X t+ EECS 495: Randomized Algorithms Lecture 14 Random Walks Reading: Motwani-Raghavan Chapter 6 Powerful tool for sampling complicated distributions since use only local moves Given: to explore state space.

More information

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 4: Discrete-Time Markov Chains September 1, 215 Dr. Itamar Arel College of Engineering Department of Electrical Engineering & Computer

More information