Weak convergence in Probability Theory A summer excursion! Day 3

Size: px
Start display at page:

Download "Weak convergence in Probability Theory A summer excursion! Day 3"

Transcription

1 BCAM June Weak convergence in Probability Theory A summer excursion! Day 3 Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu

2 BCAM June Day 1: Basic definitions of convergence for random variables will be reviewed, together with criteria and counter-examples. Day 2: Skorokhod s Theorem and coupling Examples in queueing theory, in the theory of Markov chains and time series analysis. Day 3: Poisson convergence: The Stein-Chen method with applications to problems in the theory of random graphs. Day 4: Weak convergence in function spaces Prohorov s Theorem and sequential compactness Day 5: An illustration: From random walks to Brownian motion

3 BCAM June A LITTLE DETOUR WITH N-VALUED RVS

4 BCAM June Probability mass functions Consider a rv X defined on the probability triple (Ω, F, P). It is said to be an N-valued rv if P [X N] = 1. Its probability mass function (pmf) p X (p X (x), x = 0, 1,...) is given by p X (x) = P [X = x], x = 0, 1,...

5 BCAM June Since F X (x) = p X (k), x R k N: k x we have F X p X Conversely, any [0, 1]-valued sequence p = (p(x), x = 0, 1,...) is called a pmf on N if p(x) = 1 x N Any pmf sequence uniquely defines a probability distribution function F : R [0, 1] through F(x) = p(k), x R k N: k x

6 BCAM June Fact: The N-valued rvs {X n, n = 1, 2,...} converge weakly to the N-valued rv X if and only if We also write p Xn = n p X lim p X n n (x) = p X (x), x = 0, 1,... A natural idea: Look for the limiting behavior of the pmfs, namely lim p X n n (x), x N Beware: While this limit may exist, it is not always a pmf! X n = st n + Poi(1)

7 BCAM June Theorem 1 Consider a collection of N-valued rvs {X n, n = 1, 2,...} such that the limits p(x) = lim n p X n (x), x N all exist. Then, the sequence p = (p(x), x N) is a pmf on N if and only the collection {X n, n = 1, 2,...} is tight, in which case X n = n X where X is an N-valued rv distributed according to the pmf p If p(k) = lim n p Xn (k) for each k = 0, 1,..., then lim F X n (x) = F X (x) n for all x in R which are non-integer. Thus, F Xn = n F X, and the collection {X n, n = 1, 2,...} is tight.

8 BCAM June Assume the collection {X n, n = 1, 2,...} to be tight. For every n = 1, 2,..., 1 = P [X n = x] = for arbitrary N 1.Thus, N P [X n = x] + P [X n > N] 1 = lim n = N P [X n = x] p(x) + limsup n P [X n > N] By tightness, for every ε > 0 there exists N(ε) such that sup P [X n > N(ε)] ε, n=1,2,...

9 BCAM June in which case 1 N(ε) Take ε = r 1 with r = 1, 2,... so that 1 N(r 1 ) p(x) + ε p(x) + r 1 The integers {N(r 1 ), r = 1, 2,...} can be selected to be strictly increasing, hence lim r N(r 1 ) =. Thus, 1 p(x) and the conclusion follows. p(x) = 1

10 BCAM June POISSON CONVERGENCE

11 BCAM June (Classical) Poisson convergence For each p [0, 1], let {B n (p), n = 1, 2,...} denote a collection of i.i.d. {0, 1}-valued (Bernoulli) rvs with and define P [B n (p) = 1] = 1 P [B n (p) = 0] = p, n = 1, 2,... S n (p) := B 1 (p) B n (p), n = 1, 2,... S n (p) = st Bin(n; p)

12 BCAM June Theorem 2 Consider a [0, 1]-valued sequence {p n, n = 1, 2,...} with lim np n = λ n for some λ > 0. Then, it holds that S n (p n ) = n Poi(λ) where Poi(λ) denotes a Poisson rv with parameter λ. For n large, p n λ n and S n(p n ) st Poi(np n )

13 BCAM June The Poisson paradigm The setting: For each r = 1, 2,..., let {B r,k (p r,k ), k = 1,...,k r } denote a collection of {0, 1}-valued rvs, which are not necessarily independent, and write S r (p r,1,...,p r,kr ) = B r,1 (p r,1 ) B r,kr (p r,kr )

14 BCAM June A typical result: With lim r k r =, if ( ) lim max p r,k = 0 r k=1,...,k r and lim (p r, p r,kr ) = λ r for some λ > 0, then under additional conditions of vanishingly weak correlations, S r (p r,1,...,p r,kr ) = r Poi(λ) Thus, and E [S r (p r,1,...,p r,kr )] = p r, p r,kr λ S r (p r,1,...,p r,kr ) st Poi(λ)

15 BCAM June Obvious ideas Via pmfs: lim P [S r(p r,1,...,p r,kr ) = x] = λx r x! e λ, x N Via pgfs: [ ] lim E z S r(p r,1,...,p r,k r ) = e λ(1 z), z R r

16 BCAM June Via the method of moments: For each p = 0, 1,..., lim E [S r(p r,1,...,p r,kr ) p ] = E [Poi(λ) p ] r Via the method of factorial moments: For each p = 0, 1,..., [ p ] lim E (S r (p r,1,...,p r,kr ) l) = λ p+1 r l=0

17 BCAM June Not obvious how to do this under general conditions Find new approaches Rates of convergence?

18 BCAM June TOTAL VARIATION DISTANCE, COUPLING AND APPROXIMATIONS

19 BCAM June Total variation between two pmfs For pmfs µ and ν on N, with X µ and Y ν, d TV (µ; ν) := 1 2 µ(x) ν(x) = d TV (X; Y ) This defines a distance on the space of all pmfs on N!

20 BCAM June Easy facts For any subset A of N, µ(a) ν(a) x A µ(x) ν(x) so that 1 2 ( µ(a) ν(a) + µ(ac ) ν(a c ) ) d TV (µ; ν) where A c denotes the complement of A in N. Since we conclude that µ(a c ) ν(a c ) = µ(a) ν(a), µ(a) ν(a) d TV (µ; ν).

21 BCAM June It follows that sup µ(a) ν(a) d TV (µ; ν). A N From the discussion around the maximal coupling (see later) we see that equality is in fact achieved, i.e., d TV (µ; ν) = sup µ(a) ν(a) A N This justifies the terminology maximal coupling.

22 BCAM June Weak convergence via total variation Fact: For N-valued rvs {X, X n, n = 1, 2,...}, we have X n = n X if and only if lim n d TV(X n ; X) = 0 If lim n d TV (X n ; X) = 0, then lim P [X n = x] P [X = x] = 0, n x N If X n = n X, then for each n = 1, 2,..., we have P [X n = x] P [X = x]

23 BCAM June N P [X n = x] P [X = x] + (P [X n = x] + P [X = x]) x=n+1 N = P [X n = x] P [X = x] + P [X n > N] + P [X > N] for any integer N 1. By tightness, for every ε > 0 there exists N(ε) such that sup P [X n > N(ε)] ε n=1,2,... and P [X > N(ε)] ε,

24 BCAM June whence Therefore, d TV (X n ; X) N(ε) lim sup n P [X n = x] P [X = x] + 2ε d TV (X n ; X) 2ε

25 BCAM June The coupling inequality Lemma 1 For pmfs µ and ν on N, we have d TV (µ; ν) P [X Y ] for any pair of N-valued rvs X and Y, with X µ and Y ν, which are defined on a common probability space (Ω, F, P). A pair (X, Y ) of N-valued rvs X and Y, which are defined on the same probability space (Ω, F, P), is called a coupling for the pair of pmfs µ and ν if X µ and Y ν

26 BCAM June Indeed, for any coupling (X, Y ), we have d TV (µ; ν) = 1 P [X = x] P [Y = x] 2 = P [X Y, X = x] P [X Y, Y = x] (P [X Y, X = x] + P [X Y, Y = x]) P [X Y, X = x] = P [X Y ] P [X Y, Y = x] Conclusion: d TV (X; Y ) small if coupling selected so that P [X Y ] small

27 BCAM June Independent coupling A coupling (X, Y ) for the pair of pmfs µ and ν is an independent coupling if the rvs X and Y are independent rvs, i.e., P [X = x, Y = y] = P [X = x] P [Y = y], x, y N. In that case P [X Y ] = = P [X = x, Y x] P [X = x] P [Y x]

28 BCAM June = P [X = x](1 P [Y = x]) = 1 P [X = x] P [Y = x]

29 BCAM June Maximal coupling (Dobrushin) Theorem 3 For pmfs µ and ν on N, we have d TV (µ; ν) = inf (P [X Y ] : (X, Y ) C(µ, ν)) where C(µ, ν) denotes the collection of all couplings for the pair µ and ν. Corollary 1 For pmfs µ and ν on N, there exists a coupling (X, Y ) in C(µ, ν) such that d TV (µ; ν) = P [X Y ] Such a coupling is called a maximal coupling for the pair µ and ν.

30 BCAM June Here is the one! Set γ(x) = min(ν(x), µ(x)), x N Define the pmf {η(x, y), (x, y) N 2 } on N 2 as γ(x) η(x, y) = (ν(x) γ(x))(µ(y) γ(y)) d TV (ν;µ) if x = y if x y Take Ω = N 2, F = 2 Ω and define P through the atoms, i.e., P [(x, y)] = η(x, y), (x, y) N 2

31 BCAM June Need to check Marginal conditions: ν(x) = η(x, y), x N y=0 and µ(y) = η(x, y), y N Achievability: P [X Y ] = d TV (µ; ν).

32 BCAM June Key observation 1: With A = {x N : ν(x) µ(x)}, we have ν(a) µ(a) = x A(ν(x) µ(x)) = (ν(x) µ(x)) + Similarly, µ(a c ) ν(a c ) = (µ(x) ν(x)) + Since we get ν(a) µ(a) = µ(a c ) ν(a c ), (ν(x) µ(x)) + = (µ(x) ν(x)) +

33 BCAM June Key observation 2: The obvious relations µ(x) ν(x) = (µ(x) ν(x)) + + (ν(x) µ(x)) +, x N now give (ν(x) µ(x)) + = (µ(x) ν(x)) + = d TV (µ; ν) Key observation 3: and (ν(x) γ(x)) = (µ(x) γ(x)) = (ν(x) µ(x)) + (µ(x) ν(x)) +

34 BCAM June Bernoulli rvs Pick 0 < p < p < 1. It is easy to verify that d TV (B(p), B(p )) = p p The independent coupling is not maximal The maximal coupling is achieved by taking B (p) = 1 [U p] and B (p ) = 1 [U p ] with U uniform on (0, 1). Indeed, P [1 [U p] 1 [U p ]] = P [p < U p ] = p p

35 BCAM June Poisson rvs Fact: For every λ > 0 and µ, we have d TV (Poi(λ), Poi(µ)) λ µ Assume 0 < λ < µ. Pick X Poi(λ) and Z Poi(µ λ) with X and Z independent. Thus, Y = X + Z Poi(µ). Obviously P [X Y ] = P [X X + Z] = P [Z 0] = 1 e (µ λ) and use the fact that 1 x e x for all x 0.

36 BCAM June A useful fact via coupling Proposition 1 For arbitrary pmfs µ 1,...,µ n, ν 1,...,ν n on N, it holds d TV (µ 1... µ n ; ν 1... ν n ) d TV (µ i ; ν i ) i=1

37 BCAM June An equivalent form Proposition 2 Consider mutually independent N-valued rvs X 1,...,X n defined on a common probability space with X i µ i for all i = 1,..., n. Similarly, consider mutually independent N-valued rvs Y 1,...,Y n defined on a common (possibly different) probability space with Y i ν i for all i = 1,...,n. Then, it holds d TV (X X n ; Y Y n ) d TV (X i ; Y i ) i=1

38 BCAM June For each i = 1,...,n, consider any coupling (X i, Y i ) in C(µ i, ν i ) such that the N 2 -valued rvs (X 1, Y 1 ),...,(X n, Y n ) are mutually independent pairs defined on a common probability space. By construction, X X n µ 1... µ n and Y Y n ν 1... ν n. since the rvs X 1,...,X n (resp. Y 1,...,Y n ) are mutually independent.

39 BCAM June By the coupling inequality, d TV (µ 1... µ n ; ν 1... ν n ) = d TV (X X n ; Y Y n ) P [X X n Y Y n ] P [ n i=1[x i Y i ]] P [X i Y i ]. i=1 Now use the maximal coupling for each i = 1,...,n so that d TV (µ i ; ν i ) = P [X i Y i ] and therefore, d TV (µ 1... µ n ; ν 1... ν n ) d TV (µ i ; ν i ). i=1

40 BCAM June An easy Poisson approximation result Consider a collection {B k (p k ), k = 1, 2,...,n} of mutually independent {0, 1}-valued (Bernoulli) rvs with P [B k (p k ) = 1] = 1 P [B k (p k ) = 0] = p k, k = 1,...,n and define S n := B 1 (p 1 ) B n (p n ). Also write λ n = p p n.

41 BCAM June Question: How well is S n approximated by a Poisson rv, say with parameter λ n? In particular, what can we say about the distance d TV (S n ; Poi(λ n ))? Answer: With mutually independent Poisson rvs Poi(p 1 ),...,Poi(p n ), we get d TV (S n ; Poi(λ n )) = d TV (B 1 (p 1 ) B n (p n ); Poi(p 1 ) Poi(p n )) d TV (B i (p i ); Poi(p i )) i=1 and an approximation is just within grasp!

42 BCAM June Computing d TV (B(p); Poi(p)) (0 < p < 1) The maximal coupling (B (p), Poi (p)) is given by P [B (p) = x, Poi (p) = y] 1 p if x = y = 0 = e p (1 p) if x = 1, y = 0 0 if x = 0, y = 1, 2,... p y y! e p if x = 1, y = 1, 2,... See Dobrushin s Theorem

43 BCAM June It is easy to see that P [B (p) Poi (p)] = ( e p (1 p) ) + y=2 p y y! e p = ( e p (1 p) ) + ( 1 e p pe p) = ( 1 e p) p Thus, d TV (B(p); Poi(p)) = ( 1 e p) p p 2 for all 0 < p < 1.

44 BCAM June A Poisson approximation is born! Thus, d TV (S n ; Poi(λ n )) d TV (B i (p i ); Poi(p i )) i=1 i=1 p 2 i Prohorov (1956) [Homogeneous case], LeCam (1960) [Heterogeneous case]

45 BCAM June THE STEIN-CHEN METHOD

46 BCAM June Handling lack of independence! Stein (1971) introduced a revolutionary method to build approximations to the distribution of sums of rvs as they appear in the CLT Chen (1975) extended the method to the situation encountered in the Poisson paradigm

47 BCAM June The basic ideas Fact 1: If Z = st Poi(λ), then E [λg(z + 1) Zg(Z)] = 0 for every bounded mapping g : N R. Fact 2: If Z = st Poi(λ), then any mapping f : N R satisfying can be put in the form E [f(z)] = 0 f(x) = λg(x + 1) xg(x), x N for some bounded mapping g : N R.

48 BCAM June Fact 3: The N-valued rv Z is Poisson distributed with parameter λ if and only if for every bounded mapping g : N R. E [λg(z + 1) Zg(Z)] = 0 (1) Pick k = 0, 1,... and take g k (x) = δ(x; k), x N Then, the condition (1) reads E [λg k (Z + 1) Zg k (Z)] = 0 and this is equivalent to λp [Z + 1 = k] = kp [Z = k].

49 BCAM June The leap Consider an N-valued rv W. If the distributional approximation W st Poi(λ) holds in some appropriate technical sense, then we should expect E [λg(w + 1) Wg(W)] 0 for every bounded mapping g : N R. Conversely, if the N-valued rv W has the property that E [λg(w + 1) Wg(W)] 0 for every bounded mapping g : N R, then it might be possible to tease out a formal statement for the approximation W st Poi(λ).

50 BCAM June More formally Step 1: For A N, find the mapping g : N R which solves the functional equation λg(x + 1) xg(x) = 1 [x A] Poi(λ; A), x N with g(0) = 0. Denote this solution by g λ,a. Why is this useful? Because for any N-valued rv W, we have E [λg(w + 1) Wg(W)] = P [W A] Poi(λ; A) and bounds on the total variation distance beween W and Poi(λ) are possibly within reach!

51 BCAM June Step 2: The bounds and g λ,a = sup g λ,a (x) min(1, λ 1 ) x N g λ,a = sup g λ,a (x + 1) g λ,a (x) λ 1 (1 e λ ) x N both hold uniformly in A! Question: How do we use these facts?

52 BCAM June Back to the independent case Consider a collection {B k (p k ), k = 1, 2,...,n} of mutually independent {0, 1}-valued (Bernoulli) rvs with P [B k (p k ) = 1] = 1 P [B k (p k ) = 0] = p k, k = 1,...,n and define W = B 1 (p 1 ) B n (p n ). Also, for each k = 1,...,n, write W k = B k (p k ) l=1,l k Set λ = p p n.

53 BCAM June Consider a bounded mapping g : N R. First, E [Wg(W)] = E [(B 1 (p 1 ) B n (p n )) g(w)] = E [B k (p k )g(w)] = = = k=1 E [B k (p k )g(w k + 1)] k=1 E [B k (p k )] E [g(w k + 1)] k=1 p k E [g(w k + 1)] k=1 as we note the independence of the rvs B k (p k ) and W k.

54 BCAM June On the other hand, E [λg(w + 1)] = p k E [g(w + 1)] k=1 Consequently, E [λg(w + 1) Wg(W)] = λe [g(w + 1)] E [Wg(W)] = p k E [g(w + 1) g(w k + 1)] = k=1 p k E [B k (p k ) (g(w k + 2) g(w k + 1))] k=1

55 BCAM June Without further invoking independence, we find two bounds: E [λg(w + 1) Wg(W)] p k E [B k (p k ) g(w k + 2) g(w k + 1) ] k=1 g g p k E [B k (p k )] k=1 k=1 p 2 k Also E [λg(w + 1) Wg(W)] 2 g k=1 p 2 k

56 BCAM June Now pick A N and apply this last bound with g = g λ,a : We get P [W A] Poi(λ; A) = E [λg λ,a (W + 1) Wg λ,a (W)] ( g λ,a or g λ,a ) k=1 p 2 k Conclusion: with d TV (W; Poi(λ)) C(λ) p 2 k. k=1 C(λ) = λ 1 ( 1 e λ)

57 BCAM June And the dependent case? For each k = 1, 2,...,n, we have E [B k (p k )g(w)] = E [B k (p k )g(w k + 1)] = E [B k (p k )E [g(w k + 1) B k (p k )]] = p k E [g(w k + 1) B k (p k ) = 1] = p k E [g(w) B k (p k ) = 1] Consequently, = E [λg(w + 1) Wg(W)] p k E [g(w + 1) E [g(w) B k (p k ) = 1]] k=1

58 BCAM June What if? Coupling assumption: Assume that for each k = 1,..., n, there exist a probability triple (Ω k, F k, P k ) and N-valued rvs V k and U k both defined on it such that U k = st W and V k + 1 = st [W B k (p k ) = 1] Therefore, E [g(w + 1) E [g(w) B k (p k ) = 1]] = E [g(u k + 1) g(v k + 1)] g E [ U k V k ]

59 BCAM June It is now plain that E [λg(w + 1) Wg(W)] p k E [g(u k + 1) g(v k + 1)] k=1 g p k E [ U k V k ] k=1

60 BCAM June Now pick A N and apply this last bound with g = g λ,a : We get P [W A] Poi(λ; A) = E [λg λ,a (W + 1) Wg λ,a (W)] g λ,a p k E [ U k V k ] k=1 Conclusion: d TV (W; Poi(λ)) C(λ) p k E [ U k V k ] k=1

61 BCAM June Strategy A good approximation can be obtained if we are able to construct the couplings (U 1, V 1 ),...,(U n, V n ) such that E [ U k V k ] is small since then d TV (W; Poi(λ)) C(λ) p k E [ U k V k ] k=1 U k = st W and V k + 1 = st [W B k (p k ) = 1], k = 1,...,K Many strategies have been developed to do just that!

62 BCAM June Sometimes two moments suffice An interesting situation occurs when the construction is such that V k U k, k = 1, 2,... then = = p k E [ U k V k ] k=1 p k E [U k V k ] k=1 p k E [U k ] k=1 k=1 p k E [V k ]

63 BCAM June = p k E [W] p k (E [W B k (p k ) = 1] 1) k=1 k=1 k=1 [ n ] = E [W] 2 E B k (p k )E [W B k (p k ) = 1] + λ = λ + E [W] 2 E [ W 2] = λ Var[W]

On the convergence of sequences of random variables: A primer

On the convergence of sequences of random variables: A primer BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :

More information

Chapter 1. Measure Spaces. 1.1 Algebras and σ algebras of sets Notation and preliminaries

Chapter 1. Measure Spaces. 1.1 Algebras and σ algebras of sets Notation and preliminaries Chapter 1 Measure Spaces 1.1 Algebras and σ algebras of sets 1.1.1 Notation and preliminaries We shall denote by X a nonempty set, by P(X) the set of all parts (i.e., subsets) of X, and by the empty set.

More information

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2 Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction

More information

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2 Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,

More information

Fundamental Tools - Probability Theory IV

Fundamental Tools - Probability Theory IV Fundamental Tools - Probability Theory IV MSc Financial Mathematics The University of Warwick October 1, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory IV 1 / 14 Model-independent

More information

Susceptible-Infective-Removed Epidemics and Erdős-Rényi random

Susceptible-Infective-Removed Epidemics and Erdős-Rényi random Susceptible-Infective-Removed Epidemics and Erdős-Rényi random graphs MSR-Inria Joint Centre October 13, 2015 SIR epidemics: the Reed-Frost model Individuals i [n] when infected, attempt to infect all

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then

More information

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3) STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it

More information

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,

More information

The Kadec-Pe lczynski theorem in L p, 1 p < 2

The Kadec-Pe lczynski theorem in L p, 1 p < 2 The Kadec-Pe lczynski theorem in L p, 1 p < 2 I. Berkes and R. Tichy Abstract By a classical result of Kadec and Pe lczynski (1962), every normalized weakly null sequence in L p, p > 2 contains a subsequence

More information

Invariance Principle for Variable Speed Random Walks on Trees

Invariance Principle for Variable Speed Random Walks on Trees Invariance Principle for Variable Speed Random Walks on Trees Wolfgang Löhr, University of Duisburg-Essen joint work with Siva Athreya and Anita Winter Stochastic Analysis and Applications Thoku University,

More information

Coupling AMS Short Course

Coupling AMS Short Course Coupling AMS Short Course January 2010 Distance If µ and ν are two probability distributions on a set Ω, then the total variation distance between µ and ν is Example. Let Ω = {0, 1}, and set Then d TV

More information

MATH5011 Real Analysis I. Exercise 1 Suggested Solution

MATH5011 Real Analysis I. Exercise 1 Suggested Solution MATH5011 Real Analysis I Exercise 1 Suggested Solution Notations in the notes are used. (1) Show that every open set in R can be written as a countable union of mutually disjoint open intervals. Hint:

More information

Selected Exercises on Expectations and Some Probability Inequalities

Selected Exercises on Expectations and Some Probability Inequalities Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ

More information

The coupling method - Simons Counting Complexity Bootcamp, 2016

The coupling method - Simons Counting Complexity Bootcamp, 2016 The coupling method - Simons Counting Complexity Bootcamp, 2016 Nayantara Bhatnagar (University of Delaware) Ivona Bezáková (Rochester Institute of Technology) January 26, 2016 Techniques for bounding

More information

3 Measurable Functions

3 Measurable Functions 3 Measurable Functions Notation A pair (X, F) where F is a σ-field of subsets of X is a measurable space. If µ is a measure on F then (X, F, µ) is a measure space. If µ(x) < then (X, F, µ) is a probability

More information

Things to remember when learning probability distributions:

Things to remember when learning probability distributions: SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions

More information

Chapter 4. Measure Theory. 1. Measure Spaces

Chapter 4. Measure Theory. 1. Measure Spaces Chapter 4. Measure Theory 1. Measure Spaces Let X be a nonempty set. A collection S of subsets of X is said to be an algebra on X if S has the following properties: 1. X S; 2. if A S, then A c S; 3. if

More information

Measure and Integration: Concepts, Examples and Exercises. INDER K. RANA Indian Institute of Technology Bombay India

Measure and Integration: Concepts, Examples and Exercises. INDER K. RANA Indian Institute of Technology Bombay India Measure and Integration: Concepts, Examples and Exercises INDER K. RANA Indian Institute of Technology Bombay India Department of Mathematics, Indian Institute of Technology, Bombay, Powai, Mumbai 400076,

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

1.4 Outer measures 10 CHAPTER 1. MEASURE

1.4 Outer measures 10 CHAPTER 1. MEASURE 10 CHAPTER 1. MEASURE 1.3.6. ( Almost everywhere and null sets If (X, A, µ is a measure space, then a set in A is called a null set (or µ-null if its measure is 0. Clearly a countable union of null sets

More information

Integration on Measure Spaces

Integration on Measure Spaces Chapter 3 Integration on Measure Spaces In this chapter we introduce the general notion of a measure on a space X, define the class of measurable functions, and define the integral, first on a class of

More information

II - REAL ANALYSIS. This property gives us a way to extend the notion of content to finite unions of rectangles: we define

II - REAL ANALYSIS. This property gives us a way to extend the notion of content to finite unions of rectangles: we define 1 Measures 1.1 Jordan content in R N II - REAL ANALYSIS Let I be an interval in R. Then its 1-content is defined as c 1 (I) := b a if I is bounded with endpoints a, b. If I is unbounded, we define c 1

More information

PROBABILITY THEORY II

PROBABILITY THEORY II Ruprecht-Karls-Universität Heidelberg Institut für Angewandte Mathematik Prof. Dr. Jan JOHANNES Outline of the lecture course PROBABILITY THEORY II Summer semester 2016 Preliminary version: April 21, 2016

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Probability reminders

Probability reminders CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

Approximation of the conditional number of exceedances

Approximation of the conditional number of exceedances Approximation of the conditional number of exceedances Han Liang Gan University of Melbourne July 9, 2013 Joint work with Aihua Xia Han Liang Gan Approximation of the conditional number of exceedances

More information

THEOREMS, ETC., FOR MATH 516

THEOREMS, ETC., FOR MATH 516 THEOREMS, ETC., FOR MATH 516 Results labeled Theorem Ea.b.c (or Proposition Ea.b.c, etc.) refer to Theorem c from section a.b of Evans book (Partial Differential Equations). Proposition 1 (=Proposition

More information

On the Converse Law of Large Numbers

On the Converse Law of Large Numbers On the Converse Law of Large Numbers H. Jerome Keisler Yeneng Sun This version: March 15, 2018 Abstract Given a triangular array of random variables and a growth rate without a full upper asymptotic density,

More information

Stable Lévy motion with values in the Skorokhod space: construction and approximation

Stable Lévy motion with values in the Skorokhod space: construction and approximation Stable Lévy motion with values in the Skorokhod space: construction and approximation arxiv:1809.02103v1 [math.pr] 6 Sep 2018 Raluca M. Balan Becem Saidani September 5, 2018 Abstract In this article, we

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

STAT/MATH 395 PROBABILITY II

STAT/MATH 395 PROBABILITY II STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions

More information

Annalee Gomm Math 714: Assignment #2

Annalee Gomm Math 714: Assignment #2 Annalee Gomm Math 714: Assignment #2 3.32. Verify that if A M, λ(a = 0, and B A, then B M and λ(b = 0. Suppose that A M with λ(a = 0, and let B be any subset of A. By the nonnegativity and monotonicity

More information

Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018

Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Topics: consistent estimators; sub-σ-fields and partial observations; Doob s theorem about sub-σ-field measurability;

More information

Hahn-Banach theorems. 1. Continuous Linear Functionals

Hahn-Banach theorems. 1. Continuous Linear Functionals (April 22, 2014) Hahn-Banach theorems Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/fun/notes 2012-13/07c hahn banach.pdf] 1.

More information

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

MARKOV CHAINS AND HIDDEN MARKOV MODELS

MARKOV CHAINS AND HIDDEN MARKOV MODELS MARKOV CHAINS AND HIDDEN MARKOV MODELS MERYL SEAH Abstract. This is an expository paper outlining the basics of Markov chains. We start the paper by explaining what a finite Markov chain is. Then we describe

More information

Notes on Poisson Approximation

Notes on Poisson Approximation Notes on Poisson Approximation A. D. Barbour* Universität Zürich Progress in Stein s Method, Singapore, January 2009 These notes are a supplement to the article Topics in Poisson Approximation, which appeared

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Admin and Lecture 1: Recap of Measure Theory

Admin and Lecture 1: Recap of Measure Theory Admin and Lecture 1: Recap of Measure Theory David Aldous January 16, 2018 I don t use bcourses: Read web page (search Aldous 205B) Web page rather unorganized some topics done by Nike in 205A will post

More information

Tools from Lebesgue integration

Tools from Lebesgue integration Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

I. ANALYSIS; PROBABILITY

I. ANALYSIS; PROBABILITY ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Geometrical Constants and. Norm Inequalities in Banach Spaces

Geometrical Constants and. Norm Inequalities in Banach Spaces Geometrical Constants and Norm Inequalities in Banach Spaces Hiroyasu Mizuguchi Doctoral Program in Fundamental Sciences Graduate School of Science and Technology Niigata University March 013 Contents

More information

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied

More information

2 Random Variable Generation

2 Random Variable Generation 2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods

More information

Discrete Distributions

Discrete Distributions Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

Convergence of Random Variables

Convergence of Random Variables 1 / 15 Convergence of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay March 19, 2014 2 / 15 Motivation Theorem (Weak

More information

General Theory of Large Deviations

General Theory of Large Deviations Chapter 30 General Theory of Large Deviations A family of random variables follows the large deviations principle if the probability of the variables falling into bad sets, representing large deviations

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

The Caratheodory Construction of Measures

The Caratheodory Construction of Measures Chapter 5 The Caratheodory Construction of Measures Recall how our construction of Lebesgue measure in Chapter 2 proceeded from an initial notion of the size of a very restricted class of subsets of R,

More information

Contents. Index... 15

Contents. Index... 15 Contents Filter Bases and Nets................................................................................ 5 Filter Bases and Ultrafilters: A Brief Overview.........................................................

More information

Notes on Ordered Sets

Notes on Ordered Sets Notes on Ordered Sets Mariusz Wodzicki September 10, 2013 1 Vocabulary 1.1 Definitions Definition 1.1 A binary relation on a set S is said to be a partial order if it is reflexive, x x, weakly antisymmetric,

More information

Annals of Mathematics, 168 (2008), Maharam s problem. Dedicated to J. W. Roberts

Annals of Mathematics, 168 (2008), Maharam s problem. Dedicated to J. W. Roberts Annals of Mathematics, 168 (2008), 981 1009 Maharam s problem By Michel Talagrand Dedicated to J. W. Roberts Abstract We construct an exhaustive submeasure that is not equivalent to a measure. This solves

More information

Lecture 21: Convergence of transformations and generating a random variable

Lecture 21: Convergence of transformations and generating a random variable Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous

More information

CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS. Contents

CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS. Contents CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS ARI FREEDMAN Abstract. In this expository paper, I will give an overview of the necessary conditions for convergence in Markov chains on finite state spaces.

More information

2. The Concept of Convergence: Ultrafilters and Nets

2. The Concept of Convergence: Ultrafilters and Nets 2. The Concept of Convergence: Ultrafilters and Nets NOTE: AS OF 2008, SOME OF THIS STUFF IS A BIT OUT- DATED AND HAS A FEW TYPOS. I WILL REVISE THIS MATE- RIAL SOMETIME. In this lecture we discuss two

More information

Signed Measures. Chapter Basic Properties of Signed Measures. 4.2 Jordan and Hahn Decompositions

Signed Measures. Chapter Basic Properties of Signed Measures. 4.2 Jordan and Hahn Decompositions Chapter 4 Signed Measures Up until now our measures have always assumed values that were greater than or equal to 0. In this chapter we will extend our definition to allow for both positive negative values.

More information

Chapter 8. General Countably Additive Set Functions. 8.1 Hahn Decomposition Theorem

Chapter 8. General Countably Additive Set Functions. 8.1 Hahn Decomposition Theorem Chapter 8 General Countably dditive Set Functions In Theorem 5.2.2 the reader saw that if f : X R is integrable on the measure space (X,, µ) then we can define a countably additive set function ν on by

More information

Chapter 2. Markov Chains. Introduction

Chapter 2. Markov Chains. Introduction Chapter 2 Markov Chains Introduction A Markov chain is a sequence of random variables {X n ; n = 0, 1, 2,...}, defined on some probability space (Ω, F, IP), taking its values in a set E which could be

More information

7 Convergence in R d and in Metric Spaces

7 Convergence in R d and in Metric Spaces STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a

More information

A note on adiabatic theorem for Markov chains and adiabatic quantum computation. Yevgeniy Kovchegov Oregon State University

A note on adiabatic theorem for Markov chains and adiabatic quantum computation. Yevgeniy Kovchegov Oregon State University A note on adiabatic theorem for Markov chains and adiabatic quantum computation Yevgeniy Kovchegov Oregon State University Introduction Max Born and Vladimir Fock in 1928: a physical system remains in

More information

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i := 2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section

More information

Empirical Processes: General Weak Convergence Theory

Empirical Processes: General Weak Convergence Theory Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated

More information

Useful Probability Theorems

Useful Probability Theorems Useful Probability Theorems Shiu-Tang Li Finished: March 23, 2013 Last updated: November 2, 2013 1 Convergence in distribution Theorem 1.1. TFAE: (i) µ n µ, µ n, µ are probability measures. (ii) F n (x)

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Dedicated to Michel Théra in honor of his 70th birthday

Dedicated to Michel Théra in honor of his 70th birthday VARIATIONAL GEOMETRIC APPROACH TO GENERALIZED DIFFERENTIAL AND CONJUGATE CALCULI IN CONVEX ANALYSIS B. S. MORDUKHOVICH 1, N. M. NAM 2, R. B. RECTOR 3 and T. TRAN 4. Dedicated to Michel Théra in honor of

More information

Constraint qualifications for convex inequality systems with applications in constrained optimization

Constraint qualifications for convex inequality systems with applications in constrained optimization Constraint qualifications for convex inequality systems with applications in constrained optimization Chong Li, K. F. Ng and T. K. Pong Abstract. For an inequality system defined by an infinite family

More information

Banach Spaces II: Elementary Banach Space Theory

Banach Spaces II: Elementary Banach Space Theory BS II c Gabriel Nagy Banach Spaces II: Elementary Banach Space Theory Notes from the Functional Analysis Course (Fall 07 - Spring 08) In this section we introduce Banach spaces and examine some of their

More information

Convergence Rate of Markov Chains

Convergence Rate of Markov Chains Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution

More information

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1) 1.4. CONSTRUCTION OF LEBESGUE-STIELTJES MEASURES In this section we shall put to use the Carathéodory-Hahn theory, in order to construct measures with certain desirable properties first on the real line

More information

THEOREMS, ETC., FOR MATH 515

THEOREMS, ETC., FOR MATH 515 THEOREMS, ETC., FOR MATH 515 Proposition 1 (=comment on page 17). If A is an algebra, then any finite union or finite intersection of sets in A is also in A. Proposition 2 (=Proposition 1.1). For every

More information

LECTURE 13. Last time: Lecture outline

LECTURE 13. Last time: Lecture outline LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to

More information

Normed and Banach Spaces

Normed and Banach Spaces (August 30, 2005) Normed and Banach Spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ We have seen that many interesting spaces of functions have natural structures of Banach spaces:

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Jump Processes. Richard F. Bass

Jump Processes. Richard F. Bass Jump Processes Richard F. Bass ii c Copyright 214 Richard F. Bass Contents 1 Poisson processes 1 1.1 Definitions............................. 1 1.2 Stopping times.......................... 3 1.3 Markov

More information

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9 MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended

More information

FUNDAMENTALS OF REAL ANALYSIS by. II.1. Prelude. Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as

FUNDAMENTALS OF REAL ANALYSIS by. II.1. Prelude. Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as FUNDAMENTALS OF REAL ANALYSIS by Doğan Çömez II. MEASURES AND MEASURE SPACES II.1. Prelude Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as b n f(xdx :=

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Differentiation of Measures and Functions

Differentiation of Measures and Functions Chapter 6 Differentiation of Measures and Functions This chapter is concerned with the differentiation theory of Radon measures. In the first two sections we introduce the Radon measures and discuss two

More information

1 Generating functions

1 Generating functions 1 Generating functions Even quite straightforward counting problems can lead to laborious and lengthy calculations. These are greatly simplified by using generating functions. 2 Definition 1.1. Given a

More information

Chapter 7. Basic Probability Theory

Chapter 7. Basic Probability Theory Chapter 7. Basic Probability Theory I-Liang Chern October 20, 2016 1 / 49 What s kind of matrices satisfying RIP Random matrices with iid Gaussian entries iid Bernoulli entries (+/ 1) iid subgaussian entries

More information

2.2 Some Consequences of the Completeness Axiom

2.2 Some Consequences of the Completeness Axiom 60 CHAPTER 2. IMPORTANT PROPERTIES OF R 2.2 Some Consequences of the Completeness Axiom In this section, we use the fact that R is complete to establish some important results. First, we will prove that

More information

Lecture 3: Probability Measures - 2

Lecture 3: Probability Measures - 2 Lecture 3: Probability Measures - 2 1. Continuation of measures 1.1 Problem of continuation of a probability measure 1.2 Outer measure 1.3 Lebesgue outer measure 1.4 Lebesgue continuation of an elementary

More information

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45 Division of the Humanities and Social Sciences Supergradients KC Border Fall 2001 1 The supergradient of a concave function There is a useful way to characterize the concavity of differentiable functions.

More information

Econ Slides from Lecture 1

Econ Slides from Lecture 1 Econ 205 Sobel Econ 205 - Slides from Lecture 1 Joel Sobel August 23, 2010 Warning I can t start without assuming that something is common knowledge. You can find basic definitions of Sets and Set Operations

More information

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974 LIMITS FOR QUEUES AS THE WAITING ROOM GROWS by Daniel P. Heyman Ward Whitt Bell Communications Research AT&T Bell Laboratories Red Bank, NJ 07701 Murray Hill, NJ 07974 May 11, 1988 ABSTRACT We study the

More information

Preliminaries. Probability space

Preliminaries. Probability space Preliminaries This section revises some parts of Core A Probability, which are essential for this course, and lists some other mathematical facts to be used (without proof) in the following. Probability

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

FUZZY LIE IDEALS OVER A FUZZY FIELD. M. Akram. K.P. Shum. 1. Introduction

FUZZY LIE IDEALS OVER A FUZZY FIELD. M. Akram. K.P. Shum. 1. Introduction italian journal of pure and applied mathematics n. 27 2010 (281 292) 281 FUZZY LIE IDEALS OVER A FUZZY FIELD M. Akram Punjab University College of Information Technology University of the Punjab Old Campus,

More information