Weak convergence in Probability Theory A summer excursion! Day 3
|
|
- Della Lucas
- 5 years ago
- Views:
Transcription
1 BCAM June Weak convergence in Probability Theory A summer excursion! Day 3 Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu
2 BCAM June Day 1: Basic definitions of convergence for random variables will be reviewed, together with criteria and counter-examples. Day 2: Skorokhod s Theorem and coupling Examples in queueing theory, in the theory of Markov chains and time series analysis. Day 3: Poisson convergence: The Stein-Chen method with applications to problems in the theory of random graphs. Day 4: Weak convergence in function spaces Prohorov s Theorem and sequential compactness Day 5: An illustration: From random walks to Brownian motion
3 BCAM June A LITTLE DETOUR WITH N-VALUED RVS
4 BCAM June Probability mass functions Consider a rv X defined on the probability triple (Ω, F, P). It is said to be an N-valued rv if P [X N] = 1. Its probability mass function (pmf) p X (p X (x), x = 0, 1,...) is given by p X (x) = P [X = x], x = 0, 1,...
5 BCAM June Since F X (x) = p X (k), x R k N: k x we have F X p X Conversely, any [0, 1]-valued sequence p = (p(x), x = 0, 1,...) is called a pmf on N if p(x) = 1 x N Any pmf sequence uniquely defines a probability distribution function F : R [0, 1] through F(x) = p(k), x R k N: k x
6 BCAM June Fact: The N-valued rvs {X n, n = 1, 2,...} converge weakly to the N-valued rv X if and only if We also write p Xn = n p X lim p X n n (x) = p X (x), x = 0, 1,... A natural idea: Look for the limiting behavior of the pmfs, namely lim p X n n (x), x N Beware: While this limit may exist, it is not always a pmf! X n = st n + Poi(1)
7 BCAM June Theorem 1 Consider a collection of N-valued rvs {X n, n = 1, 2,...} such that the limits p(x) = lim n p X n (x), x N all exist. Then, the sequence p = (p(x), x N) is a pmf on N if and only the collection {X n, n = 1, 2,...} is tight, in which case X n = n X where X is an N-valued rv distributed according to the pmf p If p(k) = lim n p Xn (k) for each k = 0, 1,..., then lim F X n (x) = F X (x) n for all x in R which are non-integer. Thus, F Xn = n F X, and the collection {X n, n = 1, 2,...} is tight.
8 BCAM June Assume the collection {X n, n = 1, 2,...} to be tight. For every n = 1, 2,..., 1 = P [X n = x] = for arbitrary N 1.Thus, N P [X n = x] + P [X n > N] 1 = lim n = N P [X n = x] p(x) + limsup n P [X n > N] By tightness, for every ε > 0 there exists N(ε) such that sup P [X n > N(ε)] ε, n=1,2,...
9 BCAM June in which case 1 N(ε) Take ε = r 1 with r = 1, 2,... so that 1 N(r 1 ) p(x) + ε p(x) + r 1 The integers {N(r 1 ), r = 1, 2,...} can be selected to be strictly increasing, hence lim r N(r 1 ) =. Thus, 1 p(x) and the conclusion follows. p(x) = 1
10 BCAM June POISSON CONVERGENCE
11 BCAM June (Classical) Poisson convergence For each p [0, 1], let {B n (p), n = 1, 2,...} denote a collection of i.i.d. {0, 1}-valued (Bernoulli) rvs with and define P [B n (p) = 1] = 1 P [B n (p) = 0] = p, n = 1, 2,... S n (p) := B 1 (p) B n (p), n = 1, 2,... S n (p) = st Bin(n; p)
12 BCAM June Theorem 2 Consider a [0, 1]-valued sequence {p n, n = 1, 2,...} with lim np n = λ n for some λ > 0. Then, it holds that S n (p n ) = n Poi(λ) where Poi(λ) denotes a Poisson rv with parameter λ. For n large, p n λ n and S n(p n ) st Poi(np n )
13 BCAM June The Poisson paradigm The setting: For each r = 1, 2,..., let {B r,k (p r,k ), k = 1,...,k r } denote a collection of {0, 1}-valued rvs, which are not necessarily independent, and write S r (p r,1,...,p r,kr ) = B r,1 (p r,1 ) B r,kr (p r,kr )
14 BCAM June A typical result: With lim r k r =, if ( ) lim max p r,k = 0 r k=1,...,k r and lim (p r, p r,kr ) = λ r for some λ > 0, then under additional conditions of vanishingly weak correlations, S r (p r,1,...,p r,kr ) = r Poi(λ) Thus, and E [S r (p r,1,...,p r,kr )] = p r, p r,kr λ S r (p r,1,...,p r,kr ) st Poi(λ)
15 BCAM June Obvious ideas Via pmfs: lim P [S r(p r,1,...,p r,kr ) = x] = λx r x! e λ, x N Via pgfs: [ ] lim E z S r(p r,1,...,p r,k r ) = e λ(1 z), z R r
16 BCAM June Via the method of moments: For each p = 0, 1,..., lim E [S r(p r,1,...,p r,kr ) p ] = E [Poi(λ) p ] r Via the method of factorial moments: For each p = 0, 1,..., [ p ] lim E (S r (p r,1,...,p r,kr ) l) = λ p+1 r l=0
17 BCAM June Not obvious how to do this under general conditions Find new approaches Rates of convergence?
18 BCAM June TOTAL VARIATION DISTANCE, COUPLING AND APPROXIMATIONS
19 BCAM June Total variation between two pmfs For pmfs µ and ν on N, with X µ and Y ν, d TV (µ; ν) := 1 2 µ(x) ν(x) = d TV (X; Y ) This defines a distance on the space of all pmfs on N!
20 BCAM June Easy facts For any subset A of N, µ(a) ν(a) x A µ(x) ν(x) so that 1 2 ( µ(a) ν(a) + µ(ac ) ν(a c ) ) d TV (µ; ν) where A c denotes the complement of A in N. Since we conclude that µ(a c ) ν(a c ) = µ(a) ν(a), µ(a) ν(a) d TV (µ; ν).
21 BCAM June It follows that sup µ(a) ν(a) d TV (µ; ν). A N From the discussion around the maximal coupling (see later) we see that equality is in fact achieved, i.e., d TV (µ; ν) = sup µ(a) ν(a) A N This justifies the terminology maximal coupling.
22 BCAM June Weak convergence via total variation Fact: For N-valued rvs {X, X n, n = 1, 2,...}, we have X n = n X if and only if lim n d TV(X n ; X) = 0 If lim n d TV (X n ; X) = 0, then lim P [X n = x] P [X = x] = 0, n x N If X n = n X, then for each n = 1, 2,..., we have P [X n = x] P [X = x]
23 BCAM June N P [X n = x] P [X = x] + (P [X n = x] + P [X = x]) x=n+1 N = P [X n = x] P [X = x] + P [X n > N] + P [X > N] for any integer N 1. By tightness, for every ε > 0 there exists N(ε) such that sup P [X n > N(ε)] ε n=1,2,... and P [X > N(ε)] ε,
24 BCAM June whence Therefore, d TV (X n ; X) N(ε) lim sup n P [X n = x] P [X = x] + 2ε d TV (X n ; X) 2ε
25 BCAM June The coupling inequality Lemma 1 For pmfs µ and ν on N, we have d TV (µ; ν) P [X Y ] for any pair of N-valued rvs X and Y, with X µ and Y ν, which are defined on a common probability space (Ω, F, P). A pair (X, Y ) of N-valued rvs X and Y, which are defined on the same probability space (Ω, F, P), is called a coupling for the pair of pmfs µ and ν if X µ and Y ν
26 BCAM June Indeed, for any coupling (X, Y ), we have d TV (µ; ν) = 1 P [X = x] P [Y = x] 2 = P [X Y, X = x] P [X Y, Y = x] (P [X Y, X = x] + P [X Y, Y = x]) P [X Y, X = x] = P [X Y ] P [X Y, Y = x] Conclusion: d TV (X; Y ) small if coupling selected so that P [X Y ] small
27 BCAM June Independent coupling A coupling (X, Y ) for the pair of pmfs µ and ν is an independent coupling if the rvs X and Y are independent rvs, i.e., P [X = x, Y = y] = P [X = x] P [Y = y], x, y N. In that case P [X Y ] = = P [X = x, Y x] P [X = x] P [Y x]
28 BCAM June = P [X = x](1 P [Y = x]) = 1 P [X = x] P [Y = x]
29 BCAM June Maximal coupling (Dobrushin) Theorem 3 For pmfs µ and ν on N, we have d TV (µ; ν) = inf (P [X Y ] : (X, Y ) C(µ, ν)) where C(µ, ν) denotes the collection of all couplings for the pair µ and ν. Corollary 1 For pmfs µ and ν on N, there exists a coupling (X, Y ) in C(µ, ν) such that d TV (µ; ν) = P [X Y ] Such a coupling is called a maximal coupling for the pair µ and ν.
30 BCAM June Here is the one! Set γ(x) = min(ν(x), µ(x)), x N Define the pmf {η(x, y), (x, y) N 2 } on N 2 as γ(x) η(x, y) = (ν(x) γ(x))(µ(y) γ(y)) d TV (ν;µ) if x = y if x y Take Ω = N 2, F = 2 Ω and define P through the atoms, i.e., P [(x, y)] = η(x, y), (x, y) N 2
31 BCAM June Need to check Marginal conditions: ν(x) = η(x, y), x N y=0 and µ(y) = η(x, y), y N Achievability: P [X Y ] = d TV (µ; ν).
32 BCAM June Key observation 1: With A = {x N : ν(x) µ(x)}, we have ν(a) µ(a) = x A(ν(x) µ(x)) = (ν(x) µ(x)) + Similarly, µ(a c ) ν(a c ) = (µ(x) ν(x)) + Since we get ν(a) µ(a) = µ(a c ) ν(a c ), (ν(x) µ(x)) + = (µ(x) ν(x)) +
33 BCAM June Key observation 2: The obvious relations µ(x) ν(x) = (µ(x) ν(x)) + + (ν(x) µ(x)) +, x N now give (ν(x) µ(x)) + = (µ(x) ν(x)) + = d TV (µ; ν) Key observation 3: and (ν(x) γ(x)) = (µ(x) γ(x)) = (ν(x) µ(x)) + (µ(x) ν(x)) +
34 BCAM June Bernoulli rvs Pick 0 < p < p < 1. It is easy to verify that d TV (B(p), B(p )) = p p The independent coupling is not maximal The maximal coupling is achieved by taking B (p) = 1 [U p] and B (p ) = 1 [U p ] with U uniform on (0, 1). Indeed, P [1 [U p] 1 [U p ]] = P [p < U p ] = p p
35 BCAM June Poisson rvs Fact: For every λ > 0 and µ, we have d TV (Poi(λ), Poi(µ)) λ µ Assume 0 < λ < µ. Pick X Poi(λ) and Z Poi(µ λ) with X and Z independent. Thus, Y = X + Z Poi(µ). Obviously P [X Y ] = P [X X + Z] = P [Z 0] = 1 e (µ λ) and use the fact that 1 x e x for all x 0.
36 BCAM June A useful fact via coupling Proposition 1 For arbitrary pmfs µ 1,...,µ n, ν 1,...,ν n on N, it holds d TV (µ 1... µ n ; ν 1... ν n ) d TV (µ i ; ν i ) i=1
37 BCAM June An equivalent form Proposition 2 Consider mutually independent N-valued rvs X 1,...,X n defined on a common probability space with X i µ i for all i = 1,..., n. Similarly, consider mutually independent N-valued rvs Y 1,...,Y n defined on a common (possibly different) probability space with Y i ν i for all i = 1,...,n. Then, it holds d TV (X X n ; Y Y n ) d TV (X i ; Y i ) i=1
38 BCAM June For each i = 1,...,n, consider any coupling (X i, Y i ) in C(µ i, ν i ) such that the N 2 -valued rvs (X 1, Y 1 ),...,(X n, Y n ) are mutually independent pairs defined on a common probability space. By construction, X X n µ 1... µ n and Y Y n ν 1... ν n. since the rvs X 1,...,X n (resp. Y 1,...,Y n ) are mutually independent.
39 BCAM June By the coupling inequality, d TV (µ 1... µ n ; ν 1... ν n ) = d TV (X X n ; Y Y n ) P [X X n Y Y n ] P [ n i=1[x i Y i ]] P [X i Y i ]. i=1 Now use the maximal coupling for each i = 1,...,n so that d TV (µ i ; ν i ) = P [X i Y i ] and therefore, d TV (µ 1... µ n ; ν 1... ν n ) d TV (µ i ; ν i ). i=1
40 BCAM June An easy Poisson approximation result Consider a collection {B k (p k ), k = 1, 2,...,n} of mutually independent {0, 1}-valued (Bernoulli) rvs with P [B k (p k ) = 1] = 1 P [B k (p k ) = 0] = p k, k = 1,...,n and define S n := B 1 (p 1 ) B n (p n ). Also write λ n = p p n.
41 BCAM June Question: How well is S n approximated by a Poisson rv, say with parameter λ n? In particular, what can we say about the distance d TV (S n ; Poi(λ n ))? Answer: With mutually independent Poisson rvs Poi(p 1 ),...,Poi(p n ), we get d TV (S n ; Poi(λ n )) = d TV (B 1 (p 1 ) B n (p n ); Poi(p 1 ) Poi(p n )) d TV (B i (p i ); Poi(p i )) i=1 and an approximation is just within grasp!
42 BCAM June Computing d TV (B(p); Poi(p)) (0 < p < 1) The maximal coupling (B (p), Poi (p)) is given by P [B (p) = x, Poi (p) = y] 1 p if x = y = 0 = e p (1 p) if x = 1, y = 0 0 if x = 0, y = 1, 2,... p y y! e p if x = 1, y = 1, 2,... See Dobrushin s Theorem
43 BCAM June It is easy to see that P [B (p) Poi (p)] = ( e p (1 p) ) + y=2 p y y! e p = ( e p (1 p) ) + ( 1 e p pe p) = ( 1 e p) p Thus, d TV (B(p); Poi(p)) = ( 1 e p) p p 2 for all 0 < p < 1.
44 BCAM June A Poisson approximation is born! Thus, d TV (S n ; Poi(λ n )) d TV (B i (p i ); Poi(p i )) i=1 i=1 p 2 i Prohorov (1956) [Homogeneous case], LeCam (1960) [Heterogeneous case]
45 BCAM June THE STEIN-CHEN METHOD
46 BCAM June Handling lack of independence! Stein (1971) introduced a revolutionary method to build approximations to the distribution of sums of rvs as they appear in the CLT Chen (1975) extended the method to the situation encountered in the Poisson paradigm
47 BCAM June The basic ideas Fact 1: If Z = st Poi(λ), then E [λg(z + 1) Zg(Z)] = 0 for every bounded mapping g : N R. Fact 2: If Z = st Poi(λ), then any mapping f : N R satisfying can be put in the form E [f(z)] = 0 f(x) = λg(x + 1) xg(x), x N for some bounded mapping g : N R.
48 BCAM June Fact 3: The N-valued rv Z is Poisson distributed with parameter λ if and only if for every bounded mapping g : N R. E [λg(z + 1) Zg(Z)] = 0 (1) Pick k = 0, 1,... and take g k (x) = δ(x; k), x N Then, the condition (1) reads E [λg k (Z + 1) Zg k (Z)] = 0 and this is equivalent to λp [Z + 1 = k] = kp [Z = k].
49 BCAM June The leap Consider an N-valued rv W. If the distributional approximation W st Poi(λ) holds in some appropriate technical sense, then we should expect E [λg(w + 1) Wg(W)] 0 for every bounded mapping g : N R. Conversely, if the N-valued rv W has the property that E [λg(w + 1) Wg(W)] 0 for every bounded mapping g : N R, then it might be possible to tease out a formal statement for the approximation W st Poi(λ).
50 BCAM June More formally Step 1: For A N, find the mapping g : N R which solves the functional equation λg(x + 1) xg(x) = 1 [x A] Poi(λ; A), x N with g(0) = 0. Denote this solution by g λ,a. Why is this useful? Because for any N-valued rv W, we have E [λg(w + 1) Wg(W)] = P [W A] Poi(λ; A) and bounds on the total variation distance beween W and Poi(λ) are possibly within reach!
51 BCAM June Step 2: The bounds and g λ,a = sup g λ,a (x) min(1, λ 1 ) x N g λ,a = sup g λ,a (x + 1) g λ,a (x) λ 1 (1 e λ ) x N both hold uniformly in A! Question: How do we use these facts?
52 BCAM June Back to the independent case Consider a collection {B k (p k ), k = 1, 2,...,n} of mutually independent {0, 1}-valued (Bernoulli) rvs with P [B k (p k ) = 1] = 1 P [B k (p k ) = 0] = p k, k = 1,...,n and define W = B 1 (p 1 ) B n (p n ). Also, for each k = 1,...,n, write W k = B k (p k ) l=1,l k Set λ = p p n.
53 BCAM June Consider a bounded mapping g : N R. First, E [Wg(W)] = E [(B 1 (p 1 ) B n (p n )) g(w)] = E [B k (p k )g(w)] = = = k=1 E [B k (p k )g(w k + 1)] k=1 E [B k (p k )] E [g(w k + 1)] k=1 p k E [g(w k + 1)] k=1 as we note the independence of the rvs B k (p k ) and W k.
54 BCAM June On the other hand, E [λg(w + 1)] = p k E [g(w + 1)] k=1 Consequently, E [λg(w + 1) Wg(W)] = λe [g(w + 1)] E [Wg(W)] = p k E [g(w + 1) g(w k + 1)] = k=1 p k E [B k (p k ) (g(w k + 2) g(w k + 1))] k=1
55 BCAM June Without further invoking independence, we find two bounds: E [λg(w + 1) Wg(W)] p k E [B k (p k ) g(w k + 2) g(w k + 1) ] k=1 g g p k E [B k (p k )] k=1 k=1 p 2 k Also E [λg(w + 1) Wg(W)] 2 g k=1 p 2 k
56 BCAM June Now pick A N and apply this last bound with g = g λ,a : We get P [W A] Poi(λ; A) = E [λg λ,a (W + 1) Wg λ,a (W)] ( g λ,a or g λ,a ) k=1 p 2 k Conclusion: with d TV (W; Poi(λ)) C(λ) p 2 k. k=1 C(λ) = λ 1 ( 1 e λ)
57 BCAM June And the dependent case? For each k = 1, 2,...,n, we have E [B k (p k )g(w)] = E [B k (p k )g(w k + 1)] = E [B k (p k )E [g(w k + 1) B k (p k )]] = p k E [g(w k + 1) B k (p k ) = 1] = p k E [g(w) B k (p k ) = 1] Consequently, = E [λg(w + 1) Wg(W)] p k E [g(w + 1) E [g(w) B k (p k ) = 1]] k=1
58 BCAM June What if? Coupling assumption: Assume that for each k = 1,..., n, there exist a probability triple (Ω k, F k, P k ) and N-valued rvs V k and U k both defined on it such that U k = st W and V k + 1 = st [W B k (p k ) = 1] Therefore, E [g(w + 1) E [g(w) B k (p k ) = 1]] = E [g(u k + 1) g(v k + 1)] g E [ U k V k ]
59 BCAM June It is now plain that E [λg(w + 1) Wg(W)] p k E [g(u k + 1) g(v k + 1)] k=1 g p k E [ U k V k ] k=1
60 BCAM June Now pick A N and apply this last bound with g = g λ,a : We get P [W A] Poi(λ; A) = E [λg λ,a (W + 1) Wg λ,a (W)] g λ,a p k E [ U k V k ] k=1 Conclusion: d TV (W; Poi(λ)) C(λ) p k E [ U k V k ] k=1
61 BCAM June Strategy A good approximation can be obtained if we are able to construct the couplings (U 1, V 1 ),...,(U n, V n ) such that E [ U k V k ] is small since then d TV (W; Poi(λ)) C(λ) p k E [ U k V k ] k=1 U k = st W and V k + 1 = st [W B k (p k ) = 1], k = 1,...,K Many strategies have been developed to do just that!
62 BCAM June Sometimes two moments suffice An interesting situation occurs when the construction is such that V k U k, k = 1, 2,... then = = p k E [ U k V k ] k=1 p k E [U k V k ] k=1 p k E [U k ] k=1 k=1 p k E [V k ]
63 BCAM June = p k E [W] p k (E [W B k (p k ) = 1] 1) k=1 k=1 k=1 [ n ] = E [W] 2 E B k (p k )E [W B k (p k ) = 1] + λ = λ + E [W] 2 E [ W 2] = λ Var[W]
On the convergence of sequences of random variables: A primer
BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :
More informationChapter 1. Measure Spaces. 1.1 Algebras and σ algebras of sets Notation and preliminaries
Chapter 1 Measure Spaces 1.1 Algebras and σ algebras of sets 1.1.1 Notation and preliminaries We shall denote by X a nonempty set, by P(X) the set of all parts (i.e., subsets) of X, and by the empty set.
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationFundamental Tools - Probability Theory IV
Fundamental Tools - Probability Theory IV MSc Financial Mathematics The University of Warwick October 1, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory IV 1 / 14 Model-independent
More informationSusceptible-Infective-Removed Epidemics and Erdős-Rényi random
Susceptible-Infective-Removed Epidemics and Erdős-Rényi random graphs MSR-Inria Joint Centre October 13, 2015 SIR epidemics: the Reed-Frost model Individuals i [n] when infected, attempt to infect all
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More informationWeak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria
Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationSUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES
SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,
More informationThe Kadec-Pe lczynski theorem in L p, 1 p < 2
The Kadec-Pe lczynski theorem in L p, 1 p < 2 I. Berkes and R. Tichy Abstract By a classical result of Kadec and Pe lczynski (1962), every normalized weakly null sequence in L p, p > 2 contains a subsequence
More informationInvariance Principle for Variable Speed Random Walks on Trees
Invariance Principle for Variable Speed Random Walks on Trees Wolfgang Löhr, University of Duisburg-Essen joint work with Siva Athreya and Anita Winter Stochastic Analysis and Applications Thoku University,
More informationCoupling AMS Short Course
Coupling AMS Short Course January 2010 Distance If µ and ν are two probability distributions on a set Ω, then the total variation distance between µ and ν is Example. Let Ω = {0, 1}, and set Then d TV
More informationMATH5011 Real Analysis I. Exercise 1 Suggested Solution
MATH5011 Real Analysis I Exercise 1 Suggested Solution Notations in the notes are used. (1) Show that every open set in R can be written as a countable union of mutually disjoint open intervals. Hint:
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationThe coupling method - Simons Counting Complexity Bootcamp, 2016
The coupling method - Simons Counting Complexity Bootcamp, 2016 Nayantara Bhatnagar (University of Delaware) Ivona Bezáková (Rochester Institute of Technology) January 26, 2016 Techniques for bounding
More information3 Measurable Functions
3 Measurable Functions Notation A pair (X, F) where F is a σ-field of subsets of X is a measurable space. If µ is a measure on F then (X, F, µ) is a measure space. If µ(x) < then (X, F, µ) is a probability
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationChapter 4. Measure Theory. 1. Measure Spaces
Chapter 4. Measure Theory 1. Measure Spaces Let X be a nonempty set. A collection S of subsets of X is said to be an algebra on X if S has the following properties: 1. X S; 2. if A S, then A c S; 3. if
More informationMeasure and Integration: Concepts, Examples and Exercises. INDER K. RANA Indian Institute of Technology Bombay India
Measure and Integration: Concepts, Examples and Exercises INDER K. RANA Indian Institute of Technology Bombay India Department of Mathematics, Indian Institute of Technology, Bombay, Powai, Mumbai 400076,
More informationProbability and Measure
Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability
More information1.4 Outer measures 10 CHAPTER 1. MEASURE
10 CHAPTER 1. MEASURE 1.3.6. ( Almost everywhere and null sets If (X, A, µ is a measure space, then a set in A is called a null set (or µ-null if its measure is 0. Clearly a countable union of null sets
More informationIntegration on Measure Spaces
Chapter 3 Integration on Measure Spaces In this chapter we introduce the general notion of a measure on a space X, define the class of measurable functions, and define the integral, first on a class of
More informationII - REAL ANALYSIS. This property gives us a way to extend the notion of content to finite unions of rectangles: we define
1 Measures 1.1 Jordan content in R N II - REAL ANALYSIS Let I be an interval in R. Then its 1-content is defined as c 1 (I) := b a if I is bounded with endpoints a, b. If I is unbounded, we define c 1
More informationPROBABILITY THEORY II
Ruprecht-Karls-Universität Heidelberg Institut für Angewandte Mathematik Prof. Dr. Jan JOHANNES Outline of the lecture course PROBABILITY THEORY II Summer semester 2016 Preliminary version: April 21, 2016
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More information3 Integration and Expectation
3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ
More informationApproximation of the conditional number of exceedances
Approximation of the conditional number of exceedances Han Liang Gan University of Melbourne July 9, 2013 Joint work with Aihua Xia Han Liang Gan Approximation of the conditional number of exceedances
More informationTHEOREMS, ETC., FOR MATH 516
THEOREMS, ETC., FOR MATH 516 Results labeled Theorem Ea.b.c (or Proposition Ea.b.c, etc.) refer to Theorem c from section a.b of Evans book (Partial Differential Equations). Proposition 1 (=Proposition
More informationOn the Converse Law of Large Numbers
On the Converse Law of Large Numbers H. Jerome Keisler Yeneng Sun This version: March 15, 2018 Abstract Given a triangular array of random variables and a growth rate without a full upper asymptotic density,
More informationStable Lévy motion with values in the Skorokhod space: construction and approximation
Stable Lévy motion with values in the Skorokhod space: construction and approximation arxiv:1809.02103v1 [math.pr] 6 Sep 2018 Raluca M. Balan Becem Saidani September 5, 2018 Abstract In this article, we
More informationLECTURE 3. Last time:
LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate
More informationSTAT/MATH 395 PROBABILITY II
STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions
More informationAnnalee Gomm Math 714: Assignment #2
Annalee Gomm Math 714: Assignment #2 3.32. Verify that if A M, λ(a = 0, and B A, then B M and λ(b = 0. Suppose that A M with λ(a = 0, and let B be any subset of A. By the nonnegativity and monotonicity
More informationHomework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018
Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Topics: consistent estimators; sub-σ-fields and partial observations; Doob s theorem about sub-σ-field measurability;
More informationHahn-Banach theorems. 1. Continuous Linear Functionals
(April 22, 2014) Hahn-Banach theorems Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/fun/notes 2012-13/07c hahn banach.pdf] 1.
More informationSUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS
Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,
More informationCourse: ESO-209 Home Work: 1 Instructor: Debasis Kundu
Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationMARKOV CHAINS AND HIDDEN MARKOV MODELS
MARKOV CHAINS AND HIDDEN MARKOV MODELS MERYL SEAH Abstract. This is an expository paper outlining the basics of Markov chains. We start the paper by explaining what a finite Markov chain is. Then we describe
More informationNotes on Poisson Approximation
Notes on Poisson Approximation A. D. Barbour* Universität Zürich Progress in Stein s Method, Singapore, January 2009 These notes are a supplement to the article Topics in Poisson Approximation, which appeared
More information4 Sums of Independent Random Variables
4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables
More informationAdmin and Lecture 1: Recap of Measure Theory
Admin and Lecture 1: Recap of Measure Theory David Aldous January 16, 2018 I don t use bcourses: Read web page (search Aldous 205B) Web page rather unorganized some topics done by Nike in 205A will post
More informationTools from Lebesgue integration
Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationI. ANALYSIS; PROBABILITY
ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so
More informationFunctional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals
Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico
More informationWeak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij
Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationGeometrical Constants and. Norm Inequalities in Banach Spaces
Geometrical Constants and Norm Inequalities in Banach Spaces Hiroyasu Mizuguchi Doctoral Program in Fundamental Sciences Graduate School of Science and Technology Niigata University March 013 Contents
More informationTHE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON
GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More informationDiscrete Distributions
Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More informationConvergence of Random Variables
1 / 15 Convergence of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay March 19, 2014 2 / 15 Motivation Theorem (Weak
More informationGeneral Theory of Large Deviations
Chapter 30 General Theory of Large Deviations A family of random variables follows the large deviations principle if the probability of the variables falling into bad sets, representing large deviations
More informationThe main results about probability measures are the following two facts:
Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a
More informationThe Caratheodory Construction of Measures
Chapter 5 The Caratheodory Construction of Measures Recall how our construction of Lebesgue measure in Chapter 2 proceeded from an initial notion of the size of a very restricted class of subsets of R,
More informationContents. Index... 15
Contents Filter Bases and Nets................................................................................ 5 Filter Bases and Ultrafilters: A Brief Overview.........................................................
More informationNotes on Ordered Sets
Notes on Ordered Sets Mariusz Wodzicki September 10, 2013 1 Vocabulary 1.1 Definitions Definition 1.1 A binary relation on a set S is said to be a partial order if it is reflexive, x x, weakly antisymmetric,
More informationAnnals of Mathematics, 168 (2008), Maharam s problem. Dedicated to J. W. Roberts
Annals of Mathematics, 168 (2008), 981 1009 Maharam s problem By Michel Talagrand Dedicated to J. W. Roberts Abstract We construct an exhaustive submeasure that is not equivalent to a measure. This solves
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationCONVERGENCE THEOREM FOR FINITE MARKOV CHAINS. Contents
CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS ARI FREEDMAN Abstract. In this expository paper, I will give an overview of the necessary conditions for convergence in Markov chains on finite state spaces.
More information2. The Concept of Convergence: Ultrafilters and Nets
2. The Concept of Convergence: Ultrafilters and Nets NOTE: AS OF 2008, SOME OF THIS STUFF IS A BIT OUT- DATED AND HAS A FEW TYPOS. I WILL REVISE THIS MATE- RIAL SOMETIME. In this lecture we discuss two
More informationSigned Measures. Chapter Basic Properties of Signed Measures. 4.2 Jordan and Hahn Decompositions
Chapter 4 Signed Measures Up until now our measures have always assumed values that were greater than or equal to 0. In this chapter we will extend our definition to allow for both positive negative values.
More informationChapter 8. General Countably Additive Set Functions. 8.1 Hahn Decomposition Theorem
Chapter 8 General Countably dditive Set Functions In Theorem 5.2.2 the reader saw that if f : X R is integrable on the measure space (X,, µ) then we can define a countably additive set function ν on by
More informationChapter 2. Markov Chains. Introduction
Chapter 2 Markov Chains Introduction A Markov chain is a sequence of random variables {X n ; n = 0, 1, 2,...}, defined on some probability space (Ω, F, IP), taking its values in a set E which could be
More information7 Convergence in R d and in Metric Spaces
STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a
More informationA note on adiabatic theorem for Markov chains and adiabatic quantum computation. Yevgeniy Kovchegov Oregon State University
A note on adiabatic theorem for Markov chains and adiabatic quantum computation Yevgeniy Kovchegov Oregon State University Introduction Max Born and Vladimir Fock in 1928: a physical system remains in
More informationP i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=
2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]
More informationLaplace s Equation. Chapter Mean Value Formulas
Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic
More informationWhy study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables
ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section
More informationEmpirical Processes: General Weak Convergence Theory
Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated
More informationUseful Probability Theorems
Useful Probability Theorems Shiu-Tang Li Finished: March 23, 2013 Last updated: November 2, 2013 1 Convergence in distribution Theorem 1.1. TFAE: (i) µ n µ, µ n, µ are probability measures. (ii) F n (x)
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More informationDedicated to Michel Théra in honor of his 70th birthday
VARIATIONAL GEOMETRIC APPROACH TO GENERALIZED DIFFERENTIAL AND CONJUGATE CALCULI IN CONVEX ANALYSIS B. S. MORDUKHOVICH 1, N. M. NAM 2, R. B. RECTOR 3 and T. TRAN 4. Dedicated to Michel Théra in honor of
More informationConstraint qualifications for convex inequality systems with applications in constrained optimization
Constraint qualifications for convex inequality systems with applications in constrained optimization Chong Li, K. F. Ng and T. K. Pong Abstract. For an inequality system defined by an infinite family
More informationBanach Spaces II: Elementary Banach Space Theory
BS II c Gabriel Nagy Banach Spaces II: Elementary Banach Space Theory Notes from the Functional Analysis Course (Fall 07 - Spring 08) In this section we introduce Banach spaces and examine some of their
More informationConvergence Rate of Markov Chains
Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution
More informationn [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)
1.4. CONSTRUCTION OF LEBESGUE-STIELTJES MEASURES In this section we shall put to use the Carathéodory-Hahn theory, in order to construct measures with certain desirable properties first on the real line
More informationTHEOREMS, ETC., FOR MATH 515
THEOREMS, ETC., FOR MATH 515 Proposition 1 (=comment on page 17). If A is an algebra, then any finite union or finite intersection of sets in A is also in A. Proposition 2 (=Proposition 1.1). For every
More informationLECTURE 13. Last time: Lecture outline
LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to
More informationNormed and Banach Spaces
(August 30, 2005) Normed and Banach Spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ We have seen that many interesting spaces of functions have natural structures of Banach spaces:
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationJump Processes. Richard F. Bass
Jump Processes Richard F. Bass ii c Copyright 214 Richard F. Bass Contents 1 Poisson processes 1 1.1 Definitions............................. 1 1.2 Stopping times.......................... 3 1.3 Markov
More informationMAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9
MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended
More informationFUNDAMENTALS OF REAL ANALYSIS by. II.1. Prelude. Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as
FUNDAMENTALS OF REAL ANALYSIS by Doğan Çömez II. MEASURES AND MEASURE SPACES II.1. Prelude Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as b n f(xdx :=
More informationOptimization Theory. A Concise Introduction. Jiongmin Yong
October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization
More informationDifferentiation of Measures and Functions
Chapter 6 Differentiation of Measures and Functions This chapter is concerned with the differentiation theory of Radon measures. In the first two sections we introduce the Radon measures and discuss two
More information1 Generating functions
1 Generating functions Even quite straightforward counting problems can lead to laborious and lengthy calculations. These are greatly simplified by using generating functions. 2 Definition 1.1. Given a
More informationChapter 7. Basic Probability Theory
Chapter 7. Basic Probability Theory I-Liang Chern October 20, 2016 1 / 49 What s kind of matrices satisfying RIP Random matrices with iid Gaussian entries iid Bernoulli entries (+/ 1) iid subgaussian entries
More information2.2 Some Consequences of the Completeness Axiom
60 CHAPTER 2. IMPORTANT PROPERTIES OF R 2.2 Some Consequences of the Completeness Axiom In this section, we use the fact that R is complete to establish some important results. First, we will prove that
More informationLecture 3: Probability Measures - 2
Lecture 3: Probability Measures - 2 1. Continuation of measures 1.1 Problem of continuation of a probability measure 1.2 Outer measure 1.3 Lebesgue outer measure 1.4 Lebesgue continuation of an elementary
More informationDivision of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45
Division of the Humanities and Social Sciences Supergradients KC Border Fall 2001 1 The supergradient of a concave function There is a useful way to characterize the concavity of differentiable functions.
More informationEcon Slides from Lecture 1
Econ 205 Sobel Econ 205 - Slides from Lecture 1 Joel Sobel August 23, 2010 Warning I can t start without assuming that something is common knowledge. You can find basic definitions of Sets and Set Operations
More informationLIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974
LIMITS FOR QUEUES AS THE WAITING ROOM GROWS by Daniel P. Heyman Ward Whitt Bell Communications Research AT&T Bell Laboratories Red Bank, NJ 07701 Murray Hill, NJ 07974 May 11, 1988 ABSTRACT We study the
More informationPreliminaries. Probability space
Preliminaries This section revises some parts of Core A Probability, which are essential for this course, and lists some other mathematical facts to be used (without proof) in the following. Probability
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationFUZZY LIE IDEALS OVER A FUZZY FIELD. M. Akram. K.P. Shum. 1. Introduction
italian journal of pure and applied mathematics n. 27 2010 (281 292) 281 FUZZY LIE IDEALS OVER A FUZZY FIELD M. Akram Punjab University College of Information Technology University of the Punjab Old Campus,
More information