Elements of Probability Theory

Size: px
Start display at page:

Download "Elements of Probability Theory"

Transcription

1 Elements of Probability Theory CHUNG-MING KUAN Department of Finance National Taiwan University December 5, 2009 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

2 Lecture Outline 1 Probability Space and Random Variables Probability Space Random Variables Moments and Norms 2 Conditional Distributions and Moments Conditional Distributions Conditional Moments 3 Modes of Convergence Almost Sure Convergence Convergence in Probability Convergence in Distribution 4 Stochastic Order Notations C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

3 Lecture Outline (cont d) 5 Law of Large Numbers 6 Central Limit Theorem 7 Stochastic Processes Brownian motion Weak Convergence 8 Functional Central Limit Theorem C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

4 Probability Space and σ-algebra A probability space is a triplet (Ω, F, IP), where 1 Ω is the outcome space, whose elements ω are outcomes of the random experiment, 2 F is a a σ-algebra, a collection of subsets of Ω, 3 IP is a a probability measure assigned to the elements in F. F is a σ-algebra if 1 Ω F, 2 if A F, then A c F, 3 if A 1, A 2, F, then n=1 A n F. By (2), Ω c = F. From de Mongan s law, ( ) c A n = A c n F. n=1 n=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

5 Probability Measure IP: F [0, 1] is a real-valued set function such that 1 IP(Ω) = 1, 2 IP(A) 0 for all A F. 3 if A 1, A 2,... F are disjoint, then IP( n=1 A n) = n=1 IP(A n). IP( ) = 0, IP(A c ) = 1 IP(A), IP(A) IP(B) if A B, and IP(A B) = IP(A) + IP(B) IP(A B). If {A n } is an increasing (decreasing) sequence in F with the limiting set A, then lim n IP(A n ) = IP(A). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

6 Borel Field Let C be a collection of subsets of Ω. The σ-algebra generated by C, σ(c), is the intersection of all σ-algebras that contain C and hence the smallest σ-algebra containing C. When Ω = R, the Borel field, B, is the σ-algebra generated by all open intervals (a, b) in R. Note that (a, b), [a, b], (a, b], and (, b] can be obtained from each other by taking complement, union and/or intersection. For example, (a, b] = n=1 ( a, b + 1 ), (a, b) = n n=1 ( a, b 1 ]. n Thus, the collection all open intervals (closed intervals, half-open intervals or half lines) generates the same Borel field. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

7 The Borel field on R d, B d, is generated by all open hypercubes: (a 1, b 1 ) (a 2, b 2 ) (a d, b d ). B d can be generated by all closed hypercubes: [a 1, b 1 ] [a 2, b 2 ] [a d, b d ], or by (, b 1 ] (, b 2 ] (, b d ]. The sets that generate the Borel field B d are all Borel sets. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

8 Random Variable A random variable z defined on (Ω, F, IP) is a function z : Ω R such that for every B in the Borel field B, its inverse image is in F: z 1 (B) = {ω : z(ω) B} F. That is, z is a F/B-measurable (or simply F-measurable) function. Given ω, the resulting value z(ω) is known as a realization of z. A R d valued random variable (random vector) z defined on (Ω, F, IP) is: z: Ω R d such that for every B B d, z 1 (B) = {ω : z(ω) B} F; i.e., z is a F/B d -measurable function. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

9 Borel Measurable All the inverse images of random vector z, z 1 (B), form a σ-algebra, denoted as σ(z). It is known as the σ-algebra generated by z, or the information set associated with z. It is the smallest σ-algebra in F such that z is measurable. A function g : R R is B-measurable or Borel measurable if {ζ R: g(ζ) b} B. For random variable z defined on (Ω, F, IP) and Borel measurable function g( ), g(z) is a random variable defined on (Ω, F, IP). The same conclusion holds for d-dimensional random vector z and B d -measurable function g( ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

10 Distribution Function The joint distribution function of z is a non-decreasing, right-continuous function F z such that for ζ = (ζ 1,..., ζ d ) R d, F z (ζ) = IP{ω Ω: z 1 (ω) ζ 1,..., z d (ω) ζ d }, with lim F z(ζ) = 0, ζ 1,..., ζ d lim F z(ζ) = 1. ζ 1,..., ζ d The marginal distribution function of the i th component of z is F zi (ζ i ) = IP{ω Ω: z i (ω) ζ i } = F z (,...,, ζ i,,..., ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

11 Independence y and z are (pairwise) independent iff for any Borel sets B 1 and B 2, IP(y B 1 and z B 2 ) = IP(y B 1 ) IP(z B 2 ). A sequence of random variables {z i } is totally independent if ( ) IP {z i B i } = IP(z i B i ). all i all i Lemma 5.1 Let {z i } be a sequence of independent random variables and h i, i = 1, 2,... be Borel-measurable functions. Then {h i (z i )} is also a sequence of independent random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

12 Expectation The expectation of Z i is the Lebesgue integral of z i wrt to IP: IE(z i ) = z i (ω) d IP(ω). Ω In terms of its distribution function, IE(z i ) = ζ i df z (ζ) = ζ i df zi (ζ i ). R d R For Borel measurable function g( ) of z, IE[g(z)] = g(z(ω)) d IP(ω) = g(ζ) df z (ζ). Ω R d For example, the covariance matrix of z IE(zz ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

13 A function g is convex on a set S if for any a [0, 1] and any x, y in S, g (ax + (1 a)y) ag(x) + (1 a)g(y); g is concave on S if the inequality above is reversed. Lemma 5.2 (Jensen) Let g be a convex function on the support of z. For an integrable random variable z such that g(z) is integrable, g(ie(z)) IE[g(z)]; the inequality reverses if g is concave. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

14 L p -Norm For random variable z with finite p th moment, its L p -norm is: z p = [IE(z p )] 1/p. The inner product of square integrable random variables z i and z j is: z i, z j = IE(z i z j ). The L 2 -norm of z i can be obtained as z i 2 = z i, z i 1/2. For any c > 0 and p > 0, note that c p IP( z c) = c p 1 {ζ: ζ c} df z (ζ) ζ p df z (ζ) IE z p, {ζ: ζ c} where 1 A is the indicator function of the event A. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

15 Inequalities Lemma 5.3 (Markov) Let z be a random variable with finite p th moment. Then, IP( z c) IE z p c p, where c is a positive real number. For p = 2, Markov s inequality is also known as Chebyshev s inequality. Markov s inequality is trivial if c is small such that IE z p /c p > 1. When c becomes large, the probability that z assumes very extreme values will be vanishing at the rate c p. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

16 Lemma 5.4 (Hölder) Let y be a random variable with finite p th moment (p > 1) and z a random variable with finite q th moment (q = p/(p 1)). Then, IE yz y p z q. Since IE(yz) IE yz, we also have: Lemma 5.5 (Cauchy-Schwatz) Let y and z be two square integrable random variables. Then, IE(yz) y 2 z 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

17 Let y = 1 and x = z p. For q > p and r = q/p, by Hölder s inequality, IE z p x r y r/(r 1) = [IE(z pr )] 1/r = [IE(z q )] p/q. Lemma 5.6 (Liapunov) Let z be a random variable with finite q th moment. Then for p < q, z p z q. Lemma 5.7 (Minkowski) Let z i, i = 1,..., n, be random variables with finite p th moment (p 1). Then, n i=1 z i p n i=1 z i p. When n = 2, this is just the triangle inequality for L p -norms. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

18 Conditional Distributions Given A, B F, suppose we know B has occurred. Given the outcome space is restricted to B, the likelihood of A is characterized by the conditional probability: IP(A B) = IP(A B)/ IP(B). The conditional density function of z given y = η is f z y (ζ y = η) = f z,y(ζ, η). f y (η) f z y (ζ y = η) is clearly non-negative. Also R d f z y (ζ y = η)dζ = 1 f y (η) R d f z,y (ζ, η)dζ = 1 f y (η) f y(η) = 1. That is, f z y (ζ y = η) is a legitimate density function. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

19 Given the conditional density function f z y, for A B d, IP(z A y = η) = f z y (ζ y = η)dζ. A This probability is defined even when IP(y = η) is zero. When A = (, ζ 1 ] (, ζ d ], the conditional distribution function is F z y (ζ y = η) = IP(z 1 ζ 1,..., z d ζ d y = η). When z and y are independent, the conditional density (distribution) reduces to the unconditional density (distribution). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

20 Let G be a sub-σ-algebra of F, the conditional expectation IE(z G) is the integrable and G-measurable random variable satisfying IE(z G) d IP = z d IP, G G. G G Suppose that G is the trivial σ-algebra {Ω, }, then IE(z G) must be a constant c, so that IE(z) = z d IP = Ω Ω c d IP = c. Consider G = σ(y), the σ-algebra generated by y. IE(z y) = IE[z σ(y)], which is interpreted as the prediction of z given all the information associated with y. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

21 By definition, IE[IE(z G)] = Ω IE(z G) d IP = Ω z d IP = IE(z); That is, only a smaller σ-algebra matters in conditional expectation. Lemma 5.9 (Law of Iterated Expectations) Let G and H be two sub-σ-algebras of F such that G H. Then for the integrable random vector z, IE[IE(z H) G] = IE[IE(z G) H] = IE(z G). If z is G-measurable, then IE[g(z)x G] = g(z) IE(x G) with prob. 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

22 Lemma 5.11 Let z be a square integrable random variable. Then IE[z IE(z G)] 2 IE(z z) 2, for any G-measurable random variable z. Proof: For any square integrable, G-measurable random variable z, IE ( [z IE(z G)] z ) = IE ( [IE(z G) IE(z G)] z ) = 0. It follows that IE(z z) 2 = IE[z IE(z G) + IE(z G) z] 2 = IE[z IE(z G)] 2 + IE[IE(z G) z] 2 IE[z IE(z G)] 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

23 The conditional variance-covariance matrix of z given y is var(z y) = IE ( [z IE(z y)][z IE(z y)] y ) = IE(zz y) IE(z y) IE(z y), which leads to decomposition of analysis of variance: var(z) = IE[var(z y)] + var ( IE(z y) ). Example 5.12: Suppose that [ ] ([ ] [ y µ N y, x µ x Σ yy Σ xy Σ xy Σ xx ]). Then, IE(y x) = µ y Σ xyσ 1 xx (x µ x ), var(y x) = var(y) var ( IE(y x) ) = Σ yy Σ xyσ 1 xx Σ xy. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

24 Almost Sure Convergence A sequence of random variables, {z n ( )} n=1,2,..., is such that for a given ω, z n (ω) is a realization of the random element ω with index n, and that for a given n, z n ( ) is a random variable. Almost Sure Convergence Suppose {z n } and z are all defined on (Ω, F, IP). {z n } is said to converge to z almost surely if, and only if, IP(ω : z n (ω) z(ω) as n ) = 1, denoted as z n a.s. z or z n z a.s. (with prob. 1). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

25 Lemma 5.13 a.s. Let g : R R be a function continuous on S g R. If z n z, where z is a random variable such that IP(z S g ) = 1, then g(z n ) a.s. g(z). Proof: Let Ω 0 = {ω : z n (ω) z(ω)} and Ω 1 = {ω : z(ω) S g }. Thus, for ω (Ω 0 Ω 1 ), continuity of g ensures that g(z n (ω)) g(z(ω)). Note that (Ω 0 Ω 1 ) c = Ω c 0 Ω c 1, which has probability zero because IP(Ω c 0 ) = IP(Ωc 1 ) = 0. (Why?) It follows that Ω 0 Ω 1 has probability one, showing that g(z n ) g(z) with probability one. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

26 Convergence in Probability Convergence in Probability {z n } is said to converge to z in probability if for every ɛ > 0, lim IP(ω : z n n(ω) z(ω) > ɛ) = 0, or equivalently, lim n IP(ω : z n (ω) z(ω) ɛ) = 1. This is denoted as IP z or z n z in probability. z n Note: In this definition, the events Ω n (ɛ) = {ω : z n (ω) z(ω) ɛ} may vary with n, and convergence is referred to the probability of such events: p n = IP(Ω n (ɛ)), rather than the random variables z n. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

27 Almost sure convergence implies convergence in probability. To see this, let Ω 0 denote the set of ω such that z n (ω) z(ω). For ω Ω 0, there is some m such that ω is in Ω n (ɛ) for all n > m. That is, Ω 0 m=1 n=m Ω n (ɛ) F. As n=mω n (ɛ) is non-decreasing in m, it follows that ( ) IP(Ω 0 ) IP Ω n (ɛ) = lim m IP m=1 n=m ( n=m Ω n (ɛ) ) lim m IP( Ω m (ɛ) ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

28 Example 5.15 Let Ω = [0, 1] and IP be the Lebesgue measure. Consider the sequence of intervals {I n } in [0, 1]: [0, 1/2), [1/2, 1], [0, 1/3), [1/3, 2/3), [2/3, 1],..., and let z n = 1 In. When n tends to infinity, I n shrinks toward a singleton. For 0 < ɛ < 1, we have IP( z n > ɛ) = IP(I n ) 0, IP which shows z n 0. On the other hand, each ω [0, 1] must be covered by infinitely many intervals, so that z n (ω) = 1 for infinitely many n. This shows that z n (ω) does not converge to zero. Note: Convergence in probability permits z n to deviate from the probability limit infinitely often, but almost sure convergence does not, except for those ω in the set of probability zero. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

29 Lemma 5.16 Let {z n } be a sequence of square integrable random variables. If IP IE(z n ) c and var(z n ) 0, then z n c. Lemma 5.17 Let g : R R be a function continuous on S g R. If z n z, where z IP is a random variable such that IP(z S g ) = 1, then g(z n ) g(z). Proof: By the continuity of g, for each ɛ > 0, we can find a δ > 0 s.t. {ω : z n (ω) z(ω) δ} {ω : z(ω) S g } {ω : g(z n (ω)) g(z(ω)) ɛ}. Taking complementation of both sides, we have IP( g(z n ) g(z) > ɛ) IP( z n z > δ) 0. IP C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

30 Lemma 5.13 and Lemma 5.17 are readily generalized to R d -valued random a.s. IP variables. For instance, z n z (z n z) implies z 1,n + z 2,n z 1,n z 2,n z 2 1,n + z 2 2,n a.s. ( ) IP z 1 + z 2, a.s. ( ) IP z 1 z 2, a.s. ( ) IP z1 2 + z2 2, where z 1,n, z 2,n are two elements of z n and z 1, z 2 are the corresponding elements of z. Also, provided that z 2 0 with probability one, z 1,n /z 2,n a.s. ( ) IP z 1 /z 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

31 Convergence in Distribution Convergence in Distribution {z n } is said to converge to z in distribution, denoted as z n D z, if lim F n z n (ζ) = F z (ζ), for every continuity point ζ of F z. We also say that z n is asymptotically distributed as F z, denoted as z n A Fz ; F z is thus known as the limiting distribution of z n. Cramér-Wold Device. Let {z n } be a sequence of random vectors in R d D. Then z n z if and only if α D z n α z for every α R d such that α α = 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

32 Lemma 5.19 If z n IP z, then z n D z. For a constant c, zn IP c iff z n D c. Proof: For some arbitrary ɛ > 0 and a continuity point ζ of F z, we have IP(z n ζ) = IP({z n ζ} { z n z ɛ}) + IP({z n ζ} { z n z > ɛ}) IP(z ζ + ɛ) + IP( z n z > ɛ). IP Similarly, IP(z ζ ɛ) IP(z n ζ) + IP( z n z > ɛ). If z n z, then by passing to the limit and noting that ɛ is arbitrary, lim IP(z n n ζ) = IP(z ζ). That is, F zn (ζ) F z (ζ). The converse is not true in general, however. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

33 Theorem 5.20 (Continuous Mapping Theorem) Let g : R R be a function continuous almost everywhere on R, except D for at most countably many points. If z n z, then g(zn ) D g(z). For example, z n Theorem 5.21 D N (0, 1) implies z 2 n D χ 2 (1). Let {y n } and {z n } be two sequences of random vectors such that IP D D y n z n 0. If z n z, then yn z. Theorem 5.22 If y n converges in probability to a constant c and z n converges in D D c + z, yn z n cz, and zn /y n distribution to z, then y n + z n if c 0. D z/c C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

34 Non-Stochastic Order Notations Order notations are used to describe the behavior of real sequences. b n is (at most) of order c n, denoted as b n = O(c n ), if there exists a < such that b n /c n for all sufficiently large n. b n is of smaller order than c n, denoted as b n = o(c n ), if b n /c n 0. An O(1) sequence in bounded; an o(1) sequence converges to zero. The product of O(1) and o(1) sequences is o(1). Theorem 5.23 (a) If a n = O(n r ) and b n = O(n s ), then a n b n = O(n r+s ), a n + b n = O(n max(r,s) ). (b) If a n = o(n r ) and b n = o(n s ), then a n b n = o(n r+s ), a n + b n = o(n max(r,s) ). (c) If a n = O(n r ) and b n = o(n s ), then a n b n = o(n r+s ), a n + b n = O(n max(r,s) ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

35 Stochastic Order Notations The order notations defined earlier easily extend to describe the behavior of sequences of random variables. {z n } is O a.s. (c n ) (or O(c n ) almost surely) if z n /c n is O(1) a.s. {z n } is O IP (c n ) (or O(c n ) in probability) if for every ɛ > 0, there is some such that IP( z n /c n ) ɛ, for all n sufficiently large. Lemma 5.23 holds for stochastic order notations. For example, y n = O IP (1) and z n = o IP (1), then y n z n is o IP (1). It is very restrictive to require a random variable being bounded almost surely, but a well defined random variable is typically bounded in probability, i.e., O IP (1). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

36 D Let {z n } be a sequence of random variables such that z n z and ζ be a continuity point of F z. Then for any ɛ > 0, we can choose a sufficiently D large ζ such that IP( z > ζ) < ɛ/2. As z n z, we can also choose n large enough such that IP( z n > ζ) IP( z > ζ) < ɛ/2, which implies IP( z n > ζ) < ɛ. We have proved: Lemma 5.24 Let {z n } be a sequence of random variables such that z n z n = O IP (1). D z. Then C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

37 Law of Large Numbers When a law of large numbers holds almost surely, it is a strong law of large numbers (SLLN); when a law of large numbers holds in probability, it is a weak law of large numbers (WLLN). A sequence of random variables obeys a LLN when its sample average essentially follows its mean behavior; random irregularities (deviations from the mean) are wiped out in the limit by averaging. Kolmogorov s SLLN : Let {z t } be a sequence of i.i.d. random variables with mean µ o. Then, T 1 T t=1 z a.s. t µ o. Note that i.i.d. random variables need not obey Kolmogorov s SLLN if they do not have a finite mean, e.g., i.i.d. Cauchy random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

38 Theorem 5.26 (Markov s SLLN) Let {z t } be a sequence of independent random variables such that for some δ > 0, IE z t 1+δ is bounded for all t. Then, 1 T T [z t IE(z t )] a.s. 0. t=1 Note that here z t need not have a common mean, and the average of their means need not converge. Compared with Kolmogorov s SLLN, Markov s SLLN requires a stronger moment condition but not identical distribution. A LLN usually obtains by regulating the moments of and dependence across random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

39 Examples Example 5.27 Suppose that y t = α o y t 1 + u t with α o < 1. Then, var(y t ) = σ 2 u/(1 α 2 o), and cov(y t, y t j ) = α j o ( T ) var y t = t=1 T t=1 σ2 u 1 α 2 o. Thus, T 1 var(y t ) + 2 (T τ) cov(y t, y t τ ) T var(y t ) + 2T t=1 ( so that var T 1 ) T t=1 y t 1 T T t=1 y t IP 0. τ=1 T 1 τ=1 by Lemma That is, {y t } obeys a WLLN. cov(y t, y t τ ) = O(T ), = O(T 1 ). As IE(T 1 T t=1 y t) = 0, C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

40 Lemma 5.28 Let y t = i=0 π iu t i, where u t are i.i.d. random variables with mean zero and variance σu. 2 If i= π i <, then T 1 T t=1 y a.s. t 0. In Example 5.27, y t = i=0 αi ou t i with α o < 1, so that i=0 αi o < Lemma 5.28 is quite general and applicable to processes that can be expressed as an MA process with absolutely summable weights, e.g., weakly stationary AR(p) processes. For random variables with strong correlations over time, the variation of their partial sums may grow too rapidly and cannot be eliminated by simple averaging. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

41 Example 5.29: For the sequences {t} and {t 2 }, T t=1 t = T (T + 1)/2, T t=1 t2 = T (T + 1)(2T + 1)/6. Hence, T 1 T t=1 t and T 1 T t=1 t2 both diverge. Example 5.30: u t are i.i.d. with mean zero and variance σ 2 u. Consider now {tu t }, which does not have bounded (1 + δ) th moment and does not obey Markov s SLLN. Moreover, ( T ) T var tu t = t 2 var(u t ) = σu 2 t=1 t=1 T (T + 1)(2T + 1), 6 so that T t=1 tu t = O IP (T 3/2 ). It follows that T 1 T t=1 tu t = O IP (T 1/2 ). That is, {tu t } does not obey a WLLN. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

42 Example 5.31: y t is a random walk: y t = y t 1 + u t. For s < t, y t = y s + t i=s+1 u i = y s + v t s, where v t s is independent of y s and cov(y t, y s ) = IE(y 2 s ) = sσ 2 u. Thus, ( T ) var y t = t=1 T t=1 T 1 var(y t ) + 2 T τ=1 t=τ+1 for T t=1 var(y t) = T t=1 tσ2 u = O(T 2 ) and T 1 2 T τ=1 t=τ+1 T 1 cov(y t, y t τ ) = 2 T τ=1 t=τ+1 cov(y t, y t τ ) = O(T 3 ), (t τ)σ 2 u = O(T 3 ). Then, T t=1 y t = O IP (T 3/2 ) and T 1 T t=1 y t diverges in probability. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

43 Example 5.32: y t is the random walk in Example Then, IE(y t 1 u t ) = 0, var(y t 1 u t ) = IE(y 2 t 1 ) IE(u2 t ) = (t 1)σ 4 u, and for s < t, cov(y t 1 u t, y s 1 u s ) = IE(y t 1 y s 1 u s ) IE(u t ) = 0. This yields ( T ) var y t 1 u t = t=1 T var(y t 1 u t ) = t=1 T (t 1)σu 4 = O(T 2 ), and T t=1 y t 1u t = O IP (T ). As var(t 1 T t=1 y t 1u t ) converges to σ 4 u/2, rather than 0, {y t 1 u t } does not obey a WLLN, even though its partial sums are O IP (T ). t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

44 Central Limit Theorem (CLT) Lemma 5.35 (Lindeberg-Lévy s CLT) Let {z t } be a sequence of i.i.d. random variables with mean µ o and variance σo 2 > 0. Then, D T ( z T µ o )/σ o N (0, 1). i.i.d. random variables need not obey this CLT if they do not have a finite variance, e.g., t(2) r.v. Note that z T converges to µ o in probability, and its variance σo/t 2 vanishes when T tends to infinity. A normalizing factor T 1/2 suffices to prevent a degenerate distribution in the limit. When {z t } obeys a CLT, z T is said to converge to µ o at the rate T 1/2, and z T is understood as a root-t consistent estimator. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

45 Lemma 5.36 (Liapunov s CLT) Let {z Tt } be a triangular array of independent random variables with mean µ Tt and variance σtt 2 > 0 such that σ2 T = 1 T T t=1 σ2 Tt σ2 o > 0. If for some δ > 0, IE z Tt 2+δ are bounded, then D T ( z T µ T )/σ o N (0, 1). A CLT usually requires stronger conditions on the moment of and dependence across random variables than those needed to ensure a LLN. Moreover, every random variable must also be asymptotically negligible, in the sense that no random variable is influential in affecting the partial sums. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

46 Examples Example 5.37: {u t } is a sequence of independent random variables with mean zero, variance σu, 2 and bounded (2 + δ) th moment. we know var( T t=1 tu t) is O(T 3 ), which implies that variance of T 1/2 T t=1 tu t is diverging at the rate O(T 2 ). On the other hand, observe that ( 1 T var T 1/2 It follows that 3 T 1/2 σ u T t=1 t=1 ) t T u t = t T u t D N (0, 1). T (T + 1)(2T + 1) 6T 3 σu 2 σ2 u 3. These results show that {(t/t )u t } obeys a CLT, whereas {tu t } does not. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

47 Example 5.38: y t is a random walk: y t = y t 1 + u t, where u t are i.i.d. with mean zero and variance σu. 2 We know y t do not obey a LLN and hence do not obey a CLT. CLT for Triangular Array {z Tt } is a triangular array of random variables and obeys a CLT if 1 σ o T T [z Tt IE(z Tt )] = t=1 T ( zt µ T ) σ o D N (0, 1), where z T = T 1 T t=1 z Tt, µ T = IE( z T ), and ( T ) σt 2 = var T 1/2 z Tt σo 2 > 0. t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

48 Consider an array of square integrable random vectors z Tt in R d. Let z T denote the average of z Tt, µ T = IE( z T ), and ( 1 Σ T = var T T t=1 z Tt ) Σ o, a positive definite matrix. Using the Cramér-Wold device, {z Tt } is said to obey a multivariate CLT, in the sense that Σ 1/2 o 1 T T t=1 [z Tt IE(z Tt )] = Σ 1/2 o T ( zt µ T ) D N (0, I d ), if {α z Tt } obeys a CLT, for any α R d such that α α = 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

49 Stochastic Processes A d-dimensional stochastic process with the index set T is a measurable mapping z: Ω (R d ) T such that z(ω) = {z t (ω), t T }. For each t T, z t ( ) is a R d -valued r.v.; for each ω, z(ω) is a sample path (realization) of z, a R d -valued function on T. The finite-dimensional distributions of {z(t, ), t T } is IP(z t1 a 1,..., z tn a n ) = F t1,...,t n (a 1,..., a n ). z is stationary if F t1,...,t n are invariant under index displacement. z is Gaussian if F t1,...,t n are all (multivariate) normal. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

50 Brownian motion The process {w(t), t [0, )} is the standard Wiener process (standard Brownian motion) if it has continuous sample paths almost surely and satisfies: 1 IP ( w(0) = 0 ) = 1. 2 For 0 t 0 t 1 t k, IP ( w(t i ) w(t i 1 ) B i, i k ) = i k IP( ) w(t i ) w(t i 1 ) B i, where B i are Borel sets. 3 For 0 s < t, w(t) w(s) N (0, t s). Note: w here has independent and Gaussian increments. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

51 w(t) N (0, t) such that for r t, cov ( w(r), w(t) ) = IE [ w(r) ( w(t) w(r) )] + IE [ w(r) 2] = r. The sample paths of w are a.s. continuous but highly irregular (nowhere differentiable). To see this, note w c (t) = w(c 2 t)/c for c > 0 is also a standard Wiener process. (Why?) Then, w c (1/c) = w(c)/c. For a large c such that w(c)/c > 1, wc(1/c) 1/c = w(c) > c. That is, the sample path of w c has a slope larger than c on a very small interval (0, 1/c). The difference quotient: [w(t + h) w(t)]/h N (0, 1/ h ) can not converge to a finite limit (as h 0) with a positive prob. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

52 The d-dimensional, standard Wiener process w consists of d mutually independent, standard Wiener processes, so that for s < t, w(t) w(s) N (0, (t s) I d ). Lemma 5.39 Let w be the d-dimensional, standard Wiener process. 1 w(t) N (0, t I d ). 2 cov(w(r), w(t)) = min(r, t) I d. The Brownian bridge w 0 on [0, 1] is w 0 (t) = w(t) tw(1). Clearly, IE[w 0 (t)] = 0, and for r < t, cov ( w 0 (r), w 0 (t) ) = cov ( w(r) rw(1), w(t) tw(1) ) = r(1 t) I d. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

53 Weak Convergence IP n converges weakly to IP, denoted as IP n IP, if for every bounded, continuous real function f on S, f (s) dip n (s) f (s) d IP(s), where {IP n } and IP are probability measures on (S, S). When z n and z are all R d -valued random variables, IP n IP reduces D to the usual notion of convergence in distribution: z n z. When z n and z are d-dimensional stochastic processes with the D distributions induced by IP n and IP, z n z, also denoted as zn z, implies that all the finite-dimensional distributions of z n converge to the corresponding distributions of z. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

54 Continuous Mapping Theorem Lemma 5.40 (Continuous Mapping Theorem) Let g : R d R be a function continuous almost everywhere on R d, except for at most countably many points. If z n z, then g(z n ) g(z). Proof: Let S and S be two metric spaces with Borel σ-algebras S and S and g : S S be a measurable mapping. For IP on (S, S), define IP on (S, S ) as IP (A ) = IP(g 1 (A )), A S. For every bounded, continuous f on S, f g is also bounded and continuous on S. IP n IP now implies that f g(s) dip n (s) f g(s) d IP(s), which is equivalent to f (a) dip n(a) f (a) dip (a), proving IP n IP. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

55 Functional Central Limit Theorem (FCLT) ζ i are i.i.d. with mean zero and variance σ 2. Let s n = ζ ζ n and z n (i/n) = (σ n) 1 s i. For t [(i 1)/n, i/n), the constant interpolations of z n (i/n) is z n (t) = z n ((i 1)/n) = 1 σ n s [nt], where [nt] is the the largest integer less than or equal to nt. From Lindeberg-Lévy s CLT, 1 σ n s [nt] = ( [nt] n ) 1/2 1 σ [nt] s D [nt] t N (0, 1), which is just N (0, t), the distribution of w(t). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

56 For r < t, we have (z n (r), z n (t) z n (r)) D ( w(r), w(t) w(r) ), and hence (z n (r), z n (t)) D (w(r), w(t)). This is easily extended to establish convergence of any finite-dimensional distributions and leads to the functional central limit theorem. Lemma 5.41 (Donsker) Let ζ t be i.i.d. with mean µ o and variance σ 2 o > 0 and z T (r) = 1 σ o T [Tr] (ζ t µ o ), r [0, 1]. t=1 Then, z T w as T. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

57 Let ζ t be r.v.s with mean µ t and variance σ 2 t > 0. Define long-run variance of ζ t as σ 2 = lim T var ( 1 T T t=1 ζ t ) {ζ t } is said to obey an FCLT if z T w as T, where z T (r) = 1 σ T [Tr] ( ) ζt µ t, r [0, 1]. t=1 In the multivariate context, FCLT is z T w as T, where z T (r) = 1 Σ 1/2 ( ) ζt µ t, r [0, 1], T [Tr] t=1 w is the d-dimensional, standard Wiener process, and ( T ) ( 1 Σ = lim T T IE T ) (ζ t µ t ) (ζ t µ t ), t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58, t=1

58 Example 5.43 y t = y t 1 + u t, t = 1, 2,..., with y 0 = 0, where u t are i.i.d. with mean zero and variance σ 2 u. By Donsker s FCLT, the partial sum y [Tr] = [Tr] t=1 u t is such that 1 T 3/2 T y t = σ u t=1 T t/t t=1 (t 1)/T 1 1 y [Tr] dr σ u w(r) dr, T σu 0 This result also verifies that T t=1 y t is O IP (T 3/2 ). Similarly, 1 T 2 T yt 2 = 1 T t=1 T t=1 ( yt ) 2 1 σ 2 u w(r) 2 dr, T 0 so that T t=1 y 2 t is O IP (T 2 ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58

Elements of Probability Theory

Elements of Probability Theory Chapter 5 Elements of Probability Theory The purpose of this chapter is to summarize some important concepts and results in probability theory. Of particular interest to us are the limit theorems which

More information

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the

More information

Asymptotic Least Squares Theory

Asymptotic Least Squares Theory Asymptotic Least Squares Theory CHUNG-MING KUAN Department of Finance & CRETA December 5, 2011 C.-M. Kuan (National Taiwan Univ.) Asymptotic Least Squares Theory December 5, 2011 1 / 85 Lecture Outline

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Probability Theory I: Syllabus and Exercise

Probability Theory I: Syllabus and Exercise Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will

More information

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1 Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Economics 620, Lecture 8: Asymptotics I

Economics 620, Lecture 8: Asymptotics I Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators

More information

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

4 Expectation & the Lebesgue Theorems

4 Expectation & the Lebesgue Theorems STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does

More information

1 Introduction. 2 Diffusion equation and central limit theorem. The content of these notes is also covered by chapter 3 section B of [1].

1 Introduction. 2 Diffusion equation and central limit theorem. The content of these notes is also covered by chapter 3 section B of [1]. 1 Introduction The content of these notes is also covered by chapter 3 section B of [1]. Diffusion equation and central limit theorem Consider a sequence {ξ i } i=1 i.i.d. ξ i = d ξ with ξ : Ω { Dx, 0,

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

STAT 7032 Probability. Wlodek Bryc

STAT 7032 Probability. Wlodek Bryc STAT 7032 Probability Wlodek Bryc Revised for Spring 2019 Printed: January 14, 2019 File: Grad-Prob-2019.TEX Department of Mathematical Sciences, University of Cincinnati, Cincinnati, OH 45221 E-mail address:

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with

More information

Some functional (Hölderian) limit theorems and their applications (II)

Some functional (Hölderian) limit theorems and their applications (II) Some functional (Hölderian) limit theorems and their applications (II) Alfredas Račkauskas Vilnius University Outils Statistiques et Probabilistes pour la Finance Université de Rouen June 1 5, Rouen (Rouen

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

Univariate Unit Root Process (May 14, 2018)

Univariate Unit Root Process (May 14, 2018) Ch. Univariate Unit Root Process (May 4, 8) Introduction Much conventional asymptotic theory for least-squares estimation (e.g. the standard proofs of consistency and asymptotic normality of OLS estimators)

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,

More information

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality. 88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal

More information

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research

More information

THEOREMS, ETC., FOR MATH 515

THEOREMS, ETC., FOR MATH 515 THEOREMS, ETC., FOR MATH 515 Proposition 1 (=comment on page 17). If A is an algebra, then any finite union or finite intersection of sets in A is also in A. Proposition 2 (=Proposition 1.1). For every

More information

5 Birkhoff s Ergodic Theorem

5 Birkhoff s Ergodic Theorem 5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even

More information

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna

More information

FE 5204 Stochastic Differential Equations

FE 5204 Stochastic Differential Equations Instructor: Jim Zhu e-mail:zhu@wmich.edu http://homepages.wmich.edu/ zhu/ January 20, 2009 Preliminaries for dealing with continuous random processes. Brownian motions. Our main reference for this lecture

More information

Lecture 1: Review of Basic Asymptotic Theory

Lecture 1: Review of Basic Asymptotic Theory Lecture 1: Instructor: Department of Economics Stanfor University Prepare by Wenbo Zhou, Renmin University Basic Probability Theory Takeshi Amemiya, Avance Econometrics, 1985, Harvar University Press.

More information

Tools from Lebesgue integration

Tools from Lebesgue integration Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given

More information

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With

More information

1 Appendix A: Matrix Algebra

1 Appendix A: Matrix Algebra Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix

More information

Introduction to Stochastic processes

Introduction to Stochastic processes Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Gaussian Processes. 1. Basic Notions

Gaussian Processes. 1. Basic Notions Gaussian Processes 1. Basic Notions Let T be a set, and X : {X } T a stochastic process, defined on a suitable probability space (Ω P), that is indexed by T. Definition 1.1. We say that X is a Gaussian

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Lecture 4 Lebesgue spaces and inequalities

Lecture 4 Lebesgue spaces and inequalities Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how

More information

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction Chapter 1 Probability Theory: Introduction Basic Probability General In a probability space (Ω, Σ, P), the set Ω is the set of all possible outcomes of a probability experiment. Mathematically, Ω is just

More information

ECON 616: Lecture 1: Time Series Basics

ECON 616: Lecture 1: Time Series Basics ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters

More information

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3. 1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t

More information

Rudin Real and Complex Analysis - Harmonic Functions

Rudin Real and Complex Analysis - Harmonic Functions Rudin Real and Complex Analysis - Harmonic Functions Aaron Lou December 2018 1 Notes 1.1 The Cauchy-Riemann Equations 11.1: The Operators and Suppose f is a complex function defined in a plane open set

More information

Stochastic Processes

Stochastic Processes Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space

More information

EE514A Information Theory I Fall 2013

EE514A Information Theory I Fall 2013 EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/

More information

MATHS 730 FC Lecture Notes March 5, Introduction

MATHS 730 FC Lecture Notes March 5, Introduction 1 INTRODUCTION MATHS 730 FC Lecture Notes March 5, 2014 1 Introduction Definition. If A, B are sets and there exists a bijection A B, they have the same cardinality, which we write as A, #A. If there exists

More information

CHAPTER 3: LARGE SAMPLE THEORY

CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually

More information

18.175: Lecture 3 Integration

18.175: Lecture 3 Integration 18.175: Lecture 3 Scott Sheffield MIT Outline Outline Recall definitions Probability space is triple (Ω, F, P) where Ω is sample space, F is set of events (the σ-algebra) and P : F [0, 1] is the probability

More information

Generalized Method of Moment

Generalized Method of Moment Generalized Method of Moment CHUNG-MING KUAN Department of Finance & CRETA National Taiwan University June 16, 2010 C.-M. Kuan (Finance & CRETA, NTU Generalized Method of Moment June 16, 2010 1 / 32 Lecture

More information

7 Convergence in R d and in Metric Spaces

7 Convergence in R d and in Metric Spaces STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a

More information

Lecture 4: Introduction to stochastic processes and stochastic calculus

Lecture 4: Introduction to stochastic processes and stochastic calculus Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

Reminder Notes for the Course on Measures on Topological Spaces

Reminder Notes for the Course on Measures on Topological Spaces Reminder Notes for the Course on Measures on Topological Spaces T. C. Dorlas Dublin Institute for Advanced Studies School of Theoretical Physics 10 Burlington Road, Dublin 4, Ireland. Email: dorlas@stp.dias.ie

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

Gradient Gibbs measures with disorder

Gradient Gibbs measures with disorder Gradient Gibbs measures with disorder Codina Cotar University College London April 16, 2015, Providence Partly based on joint works with Christof Külske Outline 1 The model 2 Questions 3 Known results

More information

4th Preparation Sheet - Solutions

4th Preparation Sheet - Solutions Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space

More information

Example 4.1 Let X be a random variable and f(t) a given function of time. Then. Y (t) = f(t)x. Y (t) = X sin(ωt + δ)

Example 4.1 Let X be a random variable and f(t) a given function of time. Then. Y (t) = f(t)x. Y (t) = X sin(ωt + δ) Chapter 4 Stochastic Processes 4. Definition In the previous chapter we studied random variables as functions on a sample space X(ω), ω Ω, without regard to how these might depend on parameters. We now

More information

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015 EC402: Serial Correlation Danny Quah Economics Department, LSE Lent 2015 OUTLINE 1. Stationarity 1.1 Covariance stationarity 1.2 Explicit Models. Special cases: ARMA processes 2. Some complex numbers.

More information

Bootstrap - theoretical problems

Bootstrap - theoretical problems Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Review of measure theory

Review of measure theory 209: Honors nalysis in R n Review of measure theory 1 Outer measure, measure, measurable sets Definition 1 Let X be a set. nonempty family R of subsets of X is a ring if, B R B R and, B R B R hold. bove,

More information

Homework 11. Solutions

Homework 11. Solutions Homework 11. Solutions Problem 2.3.2. Let f n : R R be 1/n times the characteristic function of the interval (0, n). Show that f n 0 uniformly and f n µ L = 1. Why isn t it a counterexample to the Lebesgue

More information

Basic Measure and Integration Theory. Michael L. Carroll

Basic Measure and Integration Theory. Michael L. Carroll Basic Measure and Integration Theory Michael L. Carroll Sep 22, 2002 Measure Theory: Introduction What is measure theory? Why bother to learn measure theory? 1 What is measure theory? Measure theory is

More information

ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes

ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ED HERBST September 11, 2017 Background Hamilton, chapters 15-16 Trends vs Cycles A commond decomposition of macroeconomic time series

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

Mi-Hwa Ko. t=1 Z t is true. j=0

Mi-Hwa Ko. t=1 Z t is true. j=0 Commun. Korean Math. Soc. 21 (2006), No. 4, pp. 779 786 FUNCTIONAL CENTRAL LIMIT THEOREMS FOR MULTIVARIATE LINEAR PROCESSES GENERATED BY DEPENDENT RANDOM VECTORS Mi-Hwa Ko Abstract. Let X t be an m-dimensional

More information

Exercises in stochastic analysis

Exercises in stochastic analysis Exercises in stochastic analysis Franco Flandoli, Mario Maurelli, Dario Trevisan The exercises with a P are those which have been done totally or partially) in the previous lectures; the exercises with

More information

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration? Lebesgue Integration: A non-rigorous introduction What is wrong with Riemann integration? xample. Let f(x) = { 0 for x Q 1 for x / Q. The upper integral is 1, while the lower integral is 0. Yet, the function

More information

Notes 1 : Measure-theoretic foundations I

Notes 1 : Measure-theoretic foundations I Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

1 Independent increments

1 Independent increments Tel Aviv University, 2008 Brownian motion 1 1 Independent increments 1a Three convolution semigroups........... 1 1b Independent increments.............. 2 1c Continuous time................... 3 1d Bad

More information

Lecture 21: Convergence of transformations and generating a random variable

Lecture 21: Convergence of transformations and generating a random variable Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous

More information

Empirical Processes: General Weak Convergence Theory

Empirical Processes: General Weak Convergence Theory Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated

More information

Probability and Measure

Probability and Measure Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real

More information

Lecture 2: Convergence of Random Variables

Lecture 2: Convergence of Random Variables Lecture 2: Convergence of Random Variables Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Introduction to Stochastic Processes, Fall 2013 1 / 9 Convergence of Random Variables

More information

Notes on Time Series Modeling

Notes on Time Series Modeling Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g

More information

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define Homework, Real Analysis I, Fall, 2010. (1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define ρ(f, g) = 1 0 f(x) g(x) dx. Show that

More information

Stochastic Processes

Stochastic Processes Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic

More information

Conditional independence, conditional mixing and conditional association

Conditional independence, conditional mixing and conditional association Ann Inst Stat Math (2009) 61:441 460 DOI 10.1007/s10463-007-0152-2 Conditional independence, conditional mixing and conditional association B. L. S. Prakasa Rao Received: 25 July 2006 / Revised: 14 May

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

Summary of Real Analysis by Royden

Summary of Real Analysis by Royden Summary of Real Analysis by Royden Dan Hathaway May 2010 This document is a summary of the theorems and definitions and theorems from Part 1 of the book Real Analysis by Royden. In some areas, such as

More information