Essentials on the Analysis of Randomized Algorithms
|
|
- Sharleen Fitzgerald
- 5 years ago
- Views:
Transcription
1 Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random variables, techniques on approximating and bounding combinations and probabilities (emphasis on Hoeffding bound), the central limit theorem, the weak and strong law of large numbers, and fundamental problems that show these techniques in action. Basic definitions on Markov Chains are also presented. In the appendix we find the basic randomized algorithmic schemes, as well as an overview of the complexity classes where these algorithms fall. Most definitions and results are drawn from [BT02]. Basics Definition. (Probability Mass Function (PMF)). The PMF p X of a discrete random variable X is a function that describes the probability mass of each (discrete) value x that X can take; i.e. p X (x) = Pr[X = x].. (Discrete) Random Variables Definition.2 (Bernoulli Random Variable). X is a Bernoulli random variable that takes two values 0 and depending on the outcome of a random process (e.g. tossing a coin once). Its PMF is: { p, if x =, p X (x) = p, if x = 0. The expected value of X is E[X] = p, while the variance is Var[X] = p( p). Definition.3 (Binomial Random Variable). Y is a Binomial random variable with parameters and p that is constructed by Bernoulli random variables X,...,X, each of which is with probability p. It is defined as the sum Y = i= X i. Its PMF is: p Y (k) = Pr[Y = k] = ( k ) p k ( p) k, k = 0,,...,. The expected value of Y is E[Y] = p, while the variance is Var[Y] = p( p). Definition.4 (Geometric Random Variable). Given a sequence of Bernoulli random variables X,X 2,..., each of which is with probability p, Z is a Geometric random variable expressing the minimum i such that X i =. Its PMF is: p Z (k) = ( p) k p, k =,...,. The expected value of Z is E[Z] = p p, while the variance is Var[Z] = p. 2 Definition.5 (Poisson Random Variable). S is a Poisson random variable with parameter λ and PMF given by: λ λk p S (k) = e k!, k = 0,,...,. The expected value of S is E[S] = λ, and the variance is also Var[S] = λ. p Y (k) = k=0 p Z (k) = k= p S (k) = k=0
2 .2 Bernoulli process Informally it is a sequence of independent coin tosses. Definition.6 (Bernoulli process). It is a sequence X,X 2,... of independent Bernoulli random variables X i such that for every i it holds: { Pr[Xi = ] = Pr[success at the ith trial] = p Pr[X i = 0] = Pr[failure at the ith trial] = p 2 Approximating and Bounding In this section important tools on approximating and bounding probabilities will be explored. 2. The Cauchy-Schwartz Inequality ( n ) 2 ( n ) ( n x i y i x 2 i y 2 i i= i= i= ) () 2.2 Bounding Combinations Let < k < n, with k,n. Then: ( n ) k ( n ) < < k k ( ne ) k and k ( n e ) n < n! (2) 2.3 Common Approximations Exponential: ( x) e x. Poisson: The Poisson PMF with parameter λ is a good approximation for a binomial PMF with parameters and p, provided that λ = p, is very large, and p is very small. 2.4 Bounding Probabilities Union Bound: Let A,A 2,...,A be events in a probability space. Then [ ] Pr A i Pr[A i ] max {Pr[A i ]} (3) i i= i= The first inequality is equality for disjoint events A i. Markov s Inequality: Any non-negative random variable X satisfies: Pr[X α] E[X], α > 0. (4) α Chebyshev s Inequality: Let X be a r. v. with expected value µ and variance σ 2. Then: Pr[ X µ α] σ2, α > 0. (5) α2 Remark 2. (Chebyshev vs. Markov). The Chebyshev inequality tends to give better bounds than the Markov inequality, because it also uses information on the variance of X. 2
3 Theorem 2.2 (Weak Law of Large umbers). Let X,...,X be a sequence of independent identically distributed random variables, with expected value µ. For every ǫ > 0: ] Pr[ X i µ ǫ 0, as (6) i= Proof. Let X,...,X be a sequence of independent identically distributed random variables, with expected value µ and variance σ 2. Define the random variable Y = i= X i. By linearity of expectation we get E[Y] = i= E[X i] = µ. Since all the X i are independent, the variance is Var[Y] = 2 i= Var[X i] = σ2. We now apply Chebyshev s inequality and obtain Pr[ Y µ ǫ] σ2 ǫ, for any ǫ > Concentration and Tail Inequalities Proposition 2.3 (Chernoff - Hoeffding Bound [Hoe63]). Let X,...,X be independent random variables, each taking values in the range I = [α,β], and let µ denote the mean of their expectations. Then: [ ] Pr X i µ ǫ e 2ǫ2 /(β α) 2. (7) i= Assuming we want to bound the quantity above by δ, it is enough (β α) 2 2ǫ ln 2 δ. Some typical bounds obtained by the inequality are shown below: ǫ = Ω(/ ) δ , , 04 38, 56 (a) ǫ = = δ , 605, 7 9, 20, 34 3, 85, 5 (b) ǫ = = Figure : Typical lower bounds on when I = β α =. δ , , , 62 (a) ǫ = = δ , 420, 68 36, 84, , 262, 043 (b) ǫ = = Figure 2: Typical lower bounds on when I = β α = 2. Definition 2.4 (Martingale [AS08]). A martingale is a sequence X 0,...,X of random variables so that for 0 i < it holds: E[X i+ X i,x i,...,x 0 ] = X i. fair gambling Proposition 2.5 (Azuma s Inequality [Mau79, AS08]). Let c = X 0,...,X be a martingale Azuma: c = 0 with X i+ X i 0 i <. Then: Pr[ X c > λ ] < 2e λ2 /2 For λ = 2ln e λ2 /2 =. Check tbl. for some typical approximate values. (8) 3
4 λ e λ2 / Table : Typical bounds obtained for some λ by Azuma s inequality Lower bounds on Tails [AS08, appendix] In section 2.4. we obtained upper bounds on Pr[X > α] which were of the form e cǫ2. We can also obtain lower bounds on Pr[X > α]; typically we get Pr[X > α] = Ω(e cǫ2 e dǫ ). 3 Fundamental Problems 3. Coins A coin has two sides; H and T. Set Pr[H] = p and Pr[T] = p, where p is a fixed number. 3.. Games with coins How many heads H: Tossing a coin times and recording the number of times H appeared is a Binomial random variable. First H: Tossing a coin until H comes up is a Geometric random variable. Both H and T: The weighted sum p( + /( p)) + ( p)( + /p) = /p + p/( p) expresses the expected amount of coin tosses in order to observe both H and T. Fair coin: After coin tosses, we observe H T + λ or H T λ with probability bounded by Azuma s Inequality; eq. (8). 3.2 The Coupon Collector s Problem Given coupons what is the expected amount of trials in order to observe all of them? i.e. we are drawing coupons with replacement. Let T be the total time to observe all coupons, and let t i denote the time needed to collect coupon i after i coupons have been collected; i.e. T = i= t i. ote that p i = ( (i ))/, and each t i is a geometric r.v. By linearity of expectation we get: E[T] = E[t i ] = = p i i= i= i= (i ) = i= where H is the th harmonic number. For large we get: + i = i= i = H (9) E[T] = H = ln + γ o ( / 4), (0) where γ = lim (H ln) = is the Euler-Mascheroni constant [Wei]. Hence, E[T] = O(ln ). All the t i are independent, so the variance is: Var[T] = Var[t i ] = i= < 2 i= i= p i p 2 i This problem is also known as Balls in Bins Problem. < p 2 i= i = i= 2 ( + i) 2 = 2 i 2 = 2 π2 6 < 2 2 () i= i 2 4
5 Applying Chebyshev s inequality (5) to equations (9) and () we get Pr[ T H λ] < 2 λ 2. (2) 3.2. Generalized Coupon Collector s Problem Again we have coupons, but this time we are interested in a subset K of them ( K = k < ). What we want is the minimum number of trials so that the probability some member of K is missed is less than some predefined value η. Let T be the number of trials. Then a member of K is missed in all T trials with probability ( /) T. By the Union Bound (eq. (3)) this probability is upper bounded by k( /) T. By the exponential approximation this probability is upper bounded by ke T/. By requiring this probability being less than η we get: 4 Central Limit Theorem T > ln(k/η). (3) Theorem 4. (Central Limit Theorem). Let X,X 2,...,X be a sequence of independent identically distributed random variables with common µ and variance σ 2, and define i= Z = X i µ σ. (4) Then, the CDF of Z converges to the standard normal CDF E[Z ] = 0 Φ(z) = z e x2 /2 dx, 2π (5) Var[Z ] = in the sense that lim Pr[Z z] = Φ(z), for all z. (6) Proposition 4.2 (De Moivre - Laplace Approximation to the Binomial). If S is a binomial random variable with parameters p and, is large, and κ,λ, then ( ) ( ) λ + /2 p κ /2 p Pr[κ S λ] Φ Φ (7) p( p) p( p) Remark 4.3 (Quality). The closer p is to 0 or, the larger the so that the approximation is good. When p 0.5 around 40 to 50 already gives very good results. Theorem 4.4 (Strong Law of Large umbers). Let X,X 2,...,X be a sequence of independent identically distributed random variables with mean µ. Then, the sequence of sample means M = i= X i converges to µ, with probability, in the sense that 4. Applications Pr [ lim ] X i = µ =. (8) i= Example (Coins Revisited). We toss a coin 25 times and 20 times we observe H. What is the probability of this event, given that the coin is fair? A direct computation yields ( 25 20) Chebyshev yields Pr[ X ] = /9 = 0.. The Hoeffding bound gives Pr[ M ] e = 5
6 e 4.5 = Azuma s inequality gives double of what ) the Hoeffding bound( gives. The ) Central Limit Theorem gives Pr[S c] Φ so, Pr[S 25 9] Φ 6.25 = ( c p σ Φ ( ) = Φ(2.6) = So, the requested probability is less than = Using the De Moivre approximation we can compute directly Pr(S 25 = 20) Φ ( ) Φ ( ) = Φ(3.2) Φ(2.8) = = Example 2 (Lower Bound on Iterations). Assume we have a biased coin which gives rise to H with probability p and we want to estimate this value within 0.0 with probability at least 0.9. The Chebyshev inequality gives Pr[ M p 0.0] p( p) (0.0) and we want that bounded by 0.9 = 0.. This gives = 25,000. The Hoeffding bound gives =,53. In the case of the Central Limit theorem we observe that the variance of M p σ 2 is p( p)/ /(4). Hence, z = ǫ/(/(2 )) = 2ǫ. So we get: Pr[ M p deviation 0.0] 2Pr[M p 0.0] 2( Φ(2 0.0 )) 0.. This implies Φ(0.02 ) 0.95 = Φ(.645) (82.25) 2 = 6, Markov Chains [MR95] Definition 5. (Markov Chain:). A Markov Chain M is a discrete-time stochastic process defined over a set of states S in terms of a S S matrix P of transition probabilities. The set S is either finite or countably infinite. M is in only one state at a time. State transitions occur at time-steps t =,2,... The entry P ij denotes the probability that the next state will be j, given that M is currently at state i. ote that P ij [0,], i,j S and j P ij =. Remark 5.2 (Memorylessness Property:). The next state of M depends only on its current state. 5. otation and Conventions Definition 5.3 (t-step transition probability:). We denote as P (t) ij = Pr[X t = j X 0 = i]. Definition 5.4 (First transition into state j at time t:). is denoted by r t ij and is given by r (t) ij = Pr[X t = j, and for s t,x s j X 0 = i]. Definition 5.5 (Transition into state j at some time t > 0:). is denoted by f ij and is given by f ij = t>0 r(t) ij. Definition 5.6 (Expected # of steps to reach j starting from i:). is denoted by h ij and is given by { h ij = t>0 tr(t) ij, if f ij =,, otherwise. 5.2 Definitions and a theorem The states of M can be classified as: Transient: f ii < ( h ij = ). Persistent: f ii =. These can be further classified as: ull persistent: h ii =. on-null persistent: h ii <. 6
7 Definition 5.7 (Strong component C:). of a directed graph G, is a maximal subgraph of G such that there exists a path from i to j and back to i for every pair of vertices i,j C. Final: There is no edge so that we can leave G. Definition 5.8 (Irreducible Markov chain:). G consists of a single strong component. Definition 5.9 (State probability vector). q (t) = (q (t),q(t) 2,...,q(t) n ) probability that M is in state i at time t. Definition 5.0 (Stationary distribution). for M with transition matrix P is a probability distribution π such that π = πp. Definition 5. (Periodicity T of a state i). guarantees another visit to state i after a+ti steps for some i 0. A state is periodic if T >, and aperiodic otherwise. A Markov Chain M is aperiodic, if every state is aperiodic. Definition 5.2 (Ergodic state). is one that is aperiodic and non-null persistent. Definition 5.3 (Ergodic Markov Chain). is one in which all states are ergodic. Theorem 5.4 (Fundamental Theorem of Markov Chains). Any irreducible, finite, and aperiodic Markov Chain M has the following properties:. All states are ergodic. 2. There is a unique stationary distribution π, such that π i > 0, i {,...,n}. 3. For all i {,...,n} : f ii = and h ii = π i. 4. Let (i,t) be the number of times M visits state i in t steps. Then, lim t (i,t) t = π i. row vector 6 A glimpse beyond Random walks on graphs and expanders, machine learning, random number generators, parallel computation, probabilistic method. References [AS08] oga Alon and Joel Spencer. The Probabilistic Method. John Wiley, [BT02] Dimitri P. Bertsekas and John. Tsitsiklis. Introduction to Probability. Athena Scientific, [Hoe63] Wassily Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58(30):3 30, March 963. [Mau79] B. Maurey. Construction de suites symétriques. C. R. Acad. Sci. Paris, 288:679 68, 979. [MR95] Rajeev Motwani and Prabhakar Raghavan. Randomized Algorithms. Cambridge University Press, th printing, [Wei] Eric W. Weisstein. Euler-Mascheroni constant. Wolfram MathWorld. 7
8 A Basic Randomized Algorithmic Schemes Definition A. (Monte Carlo:). Monte Carlo (MC) algorithms exploit randomness in order to solve problems. The idea is that successive iterations of the core loop of the algorithm give result(s) which are independent of the previous runs. They can be classified as having one-sided error, or a two-sided error. For example, assume you have an algorithm A that decides whether x belongs in a language L, so that the answer we get is: x L Pr[A(x) accepts] p, x L Pr[A(x) accepts] = 0. This is an example of a one-sided error algorithm. A two-sided error algorithm arises if the probability of accepting an input x, when in fact x L, is non-zero. Definition A.2 (Las Vegas:). Las Vegas algorithms are Monte Carlo algorithms which never make a mistake on the result. An example of such an algorithm is quicksort (RandQS). ote that the running time of the Las Vegas algorithms depends on the input. B Complexity Classes on Randomized Algorithms Definition B. (Class RP:). RP (Randomized Polynomial time) algorithms are one-sided error Monte Carlo algorithms, that can err only when x L. Usually p = 2, but choice is arbitrary. Definition B.2 (Class ZPP:). ZPP (Zero-error Probabilistic Polynomial time) algorithms are algorithms that belong in RP co-rp. ote that Las Vegas algorithms belong in this class. Definition B.3 (Class BPP:). BPP (Bounded-error Probabilistic Polynomial time) algorithms are two-sided error Monte Carlo algorithms of the following form: x L Pr[A(x) accepts] 3 4, x L Pr[A(x) accepts] 4. Definition B.4 (Class RLP:). Class RLP is a subclass of RP, in which the algorithms use O(lg n) workspace in worst case. RandQS ZPP B. Categorizing Randomized Complexity Classes. P RP P. 2. RP BPP PP. 3. P PP PSPACE. 8
Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationApproximate Counting and Markov Chain Monte Carlo
Approximate Counting and Markov Chain Monte Carlo A Randomized Approach Arindam Pal Department of Computer Science and Engineering Indian Institute of Technology Delhi March 18, 2011 April 8, 2011 Arindam
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationLecture 2 Sep 5, 2017
CS 388R: Randomized Algorithms Fall 2017 Lecture 2 Sep 5, 2017 Prof. Eric Price Scribe: V. Orestis Papadigenopoulos and Patrick Rall NOTE: THESE NOTES HAVE NOT BEEN EDITED OR CHECKED FOR CORRECTNESS 1
More informationRandom Variable. Pr(X = a) = Pr(s)
Random Variable Definition A random variable X on a sample space Ω is a real-valued function on Ω; that is, X : Ω R. A discrete random variable is a random variable that takes on only a finite or countably
More informationELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Random Variable Discrete Random
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationRandomized Algorithms
Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours
More informationEleventh Problem Assignment
EECS April, 27 PROBLEM (2 points) The outcomes of successive flips of a particular coin are dependent and are found to be described fully by the conditional probabilities P(H n+ H n ) = P(T n+ T n ) =
More informationProbability & Computing
Probability & Computing Stochastic Process time t {X t t 2 T } state space Ω X t 2 state x 2 discrete time: T is countable T = {0,, 2,...} discrete space: Ω is finite or countably infinite X 0,X,X 2,...
More informationTwelfth Problem Assignment
EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X
More informationLecture 4: Two-point Sampling, Coupon Collector s problem
Randomized Algorithms Lecture 4: Two-point Sampling, Coupon Collector s problem Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms
More informationModel Counting for Logical Theories
Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern
More informationIntroduction to Probability
LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute
More informationMath 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =
Math 5. Rumbos Fall 07 Solutions to Review Problems for Exam. A bowl contains 5 chips of the same size and shape. Two chips are red and the other three are blue. Draw three chips from the bowl at random,
More informationDiscrete Random Variables
Chapter 5 Discrete Random Variables Suppose that an experiment and a sample space are given. A random variable is a real-valued function of the outcome of the experiment. In other words, the random variable
More informationLecture 5: The Principle of Deferred Decisions. Chernoff Bounds
Randomized Algorithms Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized
More information1 Ways to Describe a Stochastic Process
purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased
More informationLecture 20 : Markov Chains
CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called
More informationCS155: Probability and Computing: Randomized Algorithms and Probabilistic Analysis
CS155: Probability and Computing: Randomized Algorithms and Probabilistic Analysis Eli Upfal Eli Upfal@brown.edu Office: 319 TA s: Lorenzo De Stefani and Sorin Vatasoiu cs155tas@cs.brown.edu It is remarkable
More informationMarkov chains. Randomness and Computation. Markov chains. Markov processes
Markov chains Randomness and Computation or, Randomized Algorithms Mary Cryan School of Informatics University of Edinburgh Definition (Definition 7) A discrete-time stochastic process on the state space
More informationMarkov Chains and Stochastic Sampling
Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,
More informationThe Markov Chain Monte Carlo Method
The Markov Chain Monte Carlo Method Idea: define an ergodic Markov chain whose stationary distribution is the desired probability distribution. Let X 0, X 1, X 2,..., X n be the run of the chain. The Markov
More informationTail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).
Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationTheorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )
Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr Pr = Pr Pr Pr() Pr Pr. We are given three coins and are told that two of the coins are fair and the
More information1 Stat 605. Homework I. Due Feb. 1, 2011
The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due
More informationBandits, Experts, and Games
Bandits, Experts, and Games CMSC 858G Fall 2016 University of Maryland Intro to Probability* Alex Slivkins Microsoft Research NYC * Many of the slides adopted from Ron Jin and Mohammad Hajiaghayi Outline
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More informationMarkov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued
Markov Chains X(t) is a Markov Process if, for arbitrary times t 1 < t 2
More informationMARKOV PROCESSES. Valerio Di Valerio
MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationProbability. Lecture Notes. Adolfo J. Rumbos
Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................
More informationStochastic Models in Computer Science A Tutorial
Stochastic Models in Computer Science A Tutorial Dr. Snehanshu Saha Department of Computer Science PESIT BSC, Bengaluru WCI 2015 - August 10 to August 13 1 Introduction 2 Random Variable 3 Introduction
More informationRandomized algorithm
Tutorial 4 Joyce 2009-11-24 Outline Solution to Midterm Question 1 Question 2 Question 1 Question 2 Question 3 Question 4 Question 5 Solution to Midterm Solution to Midterm Solution to Midterm Question
More informationNotes on Discrete Probability
Columbia University Handout 3 W4231: Analysis of Algorithms September 21, 1999 Professor Luca Trevisan Notes on Discrete Probability The following notes cover, mostly without proofs, the basic notions
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationProblem Points S C O R E Total: 120
PSTAT 160 A Final Exam Solution December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain
More informationLectures on Markov Chains
Lectures on Markov Chains David M. McClendon Department of Mathematics Ferris State University 2016 edition 1 Contents Contents 2 1 Markov chains 4 1.1 The definition of a Markov chain.....................
More informationMarkov Chains (Part 3)
Markov Chains (Part 3) State Classification Markov Chains - State Classification Accessibility State j is accessible from state i if p ij (n) > for some n>=, meaning that starting at state i, there is
More informationRandomized Algorithms
Randomized Algorithms 南京大学 尹一通 Martingales Definition: A sequence of random variables X 0, X 1,... is a martingale if for all i > 0, E[X i X 0,...,X i1 ] = X i1 x 0, x 1,...,x i1, E[X i X 0 = x 0, X 1
More informationSTA 711: Probability & Measure Theory Robert L. Wolpert
STA 711: Probability & Measure Theory Robert L. Wolpert 6 Independence 6.1 Independent Events A collection of events {A i } F in a probability space (Ω,F,P) is called independent if P[ i I A i ] = P[A
More informationTopic 3: The Expectation of a Random Variable
Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationName: Firas Rassoul-Agha
Midterm 1 - Math 5010 - Spring 016 Name: Firas Rassoul-Agha Solve the following 4 problems. You have to clearly explain your solution. The answer carries no points. Only the work does. CALCULATORS ARE
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationX = X X n, + X 2
CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk
More informationSample Spaces, Random Variables
Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationRandom variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line
Random Variable Random variable is a mapping that maps each outcome s in the sample space to a unique real number,. s s : outcome Sample Space Real Line Eamples Toss a coin. Define the random variable
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationExample 1. The sample space of an experiment where we flip a pair of coins is denoted by:
Chapter 8 Probability 8. Preliminaries Definition (Sample Space). A Sample Space, Ω, is the set of all possible outcomes of an experiment. Such a sample space is considered discrete if Ω has finite cardinality.
More informationCSE 312 Final Review: Section AA
CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material
More informationRecap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks
Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution
More informationSome Definition and Example of Markov Chain
Some Definition and Example of Markov Chain Bowen Dai The Ohio State University April 5 th 2016 Introduction Definition and Notation Simple example of Markov Chain Aim Have some taste of Markov Chain and
More informationCSE 525 Randomized Algorithms & Probabilistic Analysis Spring Lecture 3: April 9
CSE 55 Randomized Algorithms & Probabilistic Analysis Spring 01 Lecture : April 9 Lecturer: Anna Karlin Scribe: Tyler Rigsby & John MacKinnon.1 Kinds of randomization in algorithms So far in our discussion
More informationAn Introduction to Randomized algorithms
An Introduction to Randomized algorithms C.R. Subramanian The Institute of Mathematical Sciences, Chennai. Expository talk presented at the Research Promotion Workshop on Introduction to Geometric and
More informationDO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO
QUESTION BOOKLET EE 26 Spring 2006 Final Exam Wednesday, May 7, 8am am DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the final. The final consists of five
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationLecture 4: Inequalities and Asymptotic Estimates
CSE 713: Random Graphs and Applications SUNY at Buffalo, Fall 003 Lecturer: Hung Q. Ngo Scribe: Hung Q. Ngo Lecture 4: Inequalities and Asymptotic Estimates We draw materials from [, 5, 8 10, 17, 18].
More informationExpectation of Random Variables
1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete
More informationProblem Points S C O R E Total: 120
PSTAT 160 A Final Exam December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain with the
More informationChernoff Bounds. Theme: try to show that it is unlikely a random variable X is far away from its expectation.
Chernoff Bounds Theme: try to show that it is unlikely a random variable X is far away from its expectation. The more you know about X, the better the bound you obtain. Markov s inequality: use E[X ] Chebyshev
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationMATH Notebook 5 Fall 2018/2019
MATH442601 2 Notebook 5 Fall 2018/2019 prepared by Professor Jenny Baglivo c Copyright 2004-2019 by Jenny A. Baglivo. All Rights Reserved. 5 MATH442601 2 Notebook 5 3 5.1 Sequences of IID Random Variables.............................
More informationLet (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t
2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition
More informationIEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008
IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 Justify your answers; show your work. 1. A sequence of Events. (10 points) Let {B n : n 1} be a sequence of events in
More informationAssignment 4: Solutions
Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each
More informationExpected Value 7/7/2006
Expected Value 7/7/2006 Definition Let X be a numerically-valued discrete random variable with sample space Ω and distribution function m(x). The expected value E(X) is defined by E(X) = x Ω x m(x), provided
More information25.1 Markov Chain Monte Carlo (MCMC)
CS880: Approximations Algorithms Scribe: Dave Andrzejewski Lecturer: Shuchi Chawla Topic: Approx counting/sampling, MCMC methods Date: 4/4/07 The previous lecture showed that, for self-reducible problems,
More information6.262: Discrete Stochastic Processes 2/2/11. Lecture 1: Introduction and Probability review
6.262: Discrete Stochastic Processes 2/2/11 Lecture 1: Introduction and Probability review Outline: Probability in the real world Probability as a branch of mathematics Discrete stochastic processes Processes
More informationRandomized Algorithms
Randomized Algorithms Andreas Klappenecker Texas A&M University Lecture notes of a course given in Fall 2003. Preliminary draft. c 2003 by Andreas Klappenecker. All rights reserved. Chapter 2 Probability
More informationNotes 1 : Measure-theoretic foundations I
Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationRandomized Computation
Randomized Computation Slides based on S.Aurora, B.Barak. Complexity Theory: A Modern Approach. Ahto Buldas Ahto.Buldas@ut.ee We do not assume anything about the distribution of the instances of the problem
More informationp. 4-1 Random Variables
Random Variables A Motivating Example Experiment: Sample k students without replacement from the population of all n students (labeled as 1, 2,, n, respectively) in our class. = {all combinations} = {{i
More informationHandout 5. α a1 a n. }, where. xi if a i = 1 1 if a i = 0.
Notes on Complexity Theory Last updated: October, 2005 Jonathan Katz Handout 5 1 An Improved Upper-Bound on Circuit Size Here we show the result promised in the previous lecture regarding an upper-bound
More informationLecture 4: Sampling, Tail Inequalities
Lecture 4: Sampling, Tail Inequalities Variance and Covariance Moment and Deviation Concentration and Tail Inequalities Sampling and Estimation c Hung Q. Ngo (SUNY at Buffalo) CSE 694 A Fun Course 1 /
More informationLectures 6: Degree Distributions and Concentration Inequalities
University of Washington Lecturer: Abraham Flaxman/Vahab S Mirrokni CSE599m: Algorithms and Economics of Networks April 13 th, 007 Scribe: Ben Birnbaum Lectures 6: Degree Distributions and Concentration
More informationU.C. Berkeley CS278: Computational Complexity Professor Luca Trevisan 1/29/2002. Notes for Lecture 3
U.C. Bereley CS278: Computational Complexity Handout N3 Professor Luca Trevisan 1/29/2002 Notes for Lecture 3 In this lecture we will define the probabilistic complexity classes BPP, RP, ZPP and we will
More informationS n = x + X 1 + X X n.
0 Lecture 0 0. Gambler Ruin Problem Let X be a payoff if a coin toss game such that P(X = ) = P(X = ) = /2. Suppose you start with x dollars and play the game n times. Let X,X 2,...,X n be payoffs in each
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationRandomized Complexity Classes; RP
Randomized Complexity Classes; RP Let N be a polynomial-time precise NTM that runs in time p(n) and has 2 nondeterministic choices at each step. N is a polynomial Monte Carlo Turing machine for a language
More informationMath 180B Problem Set 3
Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P
More informationMarch 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.
Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)
More informationECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018
ECE534, Spring 2018: s for Problem Set #4 Due Friday April 6, 2018 1. MMSE Estimation, Data Processing and Innovations The random variables X, Y, Z on a common probability space (Ω, F, P ) are said to
More informationLecture 2: Review of Probability
Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................
More information6.842 Randomness and Computation March 3, Lecture 8
6.84 Randomness and Computation March 3, 04 Lecture 8 Lecturer: Ronitt Rubinfeld Scribe: Daniel Grier Useful Linear Algebra Let v = (v, v,..., v n ) be a non-zero n-dimensional row vector and P an n n
More informationIntroduction to probability theory
Introduction to probability theory Fátima Sánchez Cabo Institute for Genomics and Bioinformatics, TUGraz f.sanchezcabo@tugraz.at 07/03/2007 - p. 1/35 Outline Random and conditional probability (7 March)
More informationCOUPON COLLECTING. MARK BROWN Department of Mathematics City College CUNY New York, NY
Probability in the Engineering and Informational Sciences, 22, 2008, 221 229. Printed in the U.S.A. doi: 10.1017/S0269964808000132 COUPON COLLECTING MARK ROWN Department of Mathematics City College CUNY
More informationEE514A Information Theory I Fall 2013
EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/
More informationPr[X = s Y = t] = Pr[X = s] Pr[Y = t]
Homework 4 By: John Steinberger Problem 1. Recall that a real n n matrix A is positive semidefinite if A is symmetric and x T Ax 0 for all x R n. Assume A is a real n n matrix. Show TFAE 1 : (a) A is positive
More informationProbability Theory Overview and Analysis of Randomized Algorithms
Probability Theory Overview and Analysis of Randomized Algorithms Analysis of Algorithms Prepared by John Reif, Ph.D. Probability Theory Topics a) Random Variables: Binomial and Geometric b) Useful Probabilistic
More information11.1 Set Cover ILP formulation of set cover Deterministic rounding
CS787: Advanced Algorithms Lecture 11: Randomized Rounding, Concentration Bounds In this lecture we will see some more examples of approximation algorithms based on LP relaxations. This time we will use
More informationECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding.
NAME: ECE 302 Division MWF 0:30-:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. If you are not in Prof. Pollak s section, you may not take this
More informationEECS 495: Randomized Algorithms Lecture 14 Random Walks. j p ij = 1. Pr[X t+1 = j X 0 = i 0,..., X t 1 = i t 1, X t = i] = Pr[X t+
EECS 495: Randomized Algorithms Lecture 14 Random Walks Reading: Motwani-Raghavan Chapter 6 Powerful tool for sampling complicated distributions since use only local moves Given: to explore state space.
More informationECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains
ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 4: Discrete-Time Markov Chains September 1, 215 Dr. Itamar Arel College of Engineering Department of Electrical Engineering & Computer
More information