Elements of Probability Theory
|
|
- Priscilla Barrett
- 5 years ago
- Views:
Transcription
1 Elements of Probability Theory CHUNG-MING KUAN Department of Finance National Taiwan University December 5, 2009 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
2 Lecture Outline 1 Probability Space and Random Variables Probability Space Random Variables Moments and Norms 2 Conditional Distributions and Moments Conditional Distributions Conditional Moments 3 Modes of Convergence Almost Sure Convergence Convergence in Probability Convergence in Distribution 4 Stochastic Order Notations C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
3 Lecture Outline (cont d) 5 Law of Large Numbers 6 Central Limit Theorem 7 Stochastic Processes Brownian motion Weak Convergence 8 Functional Central Limit Theorem C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
4 Probability Space and σ-algebra A probability space is a triplet (Ω, F, IP), where 1 Ω is the outcome space, whose elements ω are outcomes of the random experiment, 2 F is a a σ-algebra, a collection of subsets of Ω, 3 IP is a a probability measure assigned to the elements in F. F is a σ-algebra if 1 Ω F, 2 if A F, then A c F, 3 if A 1, A 2, F, then n=1 A n F. By (2), Ω c = F. From de Mongan s law, ( ) c A n = A c n F. n=1 n=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
5 Probability Measure IP: F [0, 1] is a real-valued set function such that 1 IP(Ω) = 1, 2 IP(A) 0 for all A F. 3 if A 1, A 2,... F are disjoint, then IP( n=1 A n) = n=1 IP(A n). IP( ) = 0, IP(A c ) = 1 IP(A), IP(A) IP(B) if A B, and IP(A B) = IP(A) + IP(B) IP(A B). If {A n } is an increasing (decreasing) sequence in F with the limiting set A, then lim n IP(A n ) = IP(A). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
6 Borel Field Let C be a collection of subsets of Ω. The σ-algebra generated by C, σ(c), is the intersection of all σ-algebras that contain C and hence the smallest σ-algebra containing C. When Ω = R, the Borel field, B, is the σ-algebra generated by all open intervals (a, b) in R. Note that (a, b), [a, b], (a, b], and (, b] can be obtained from each other by taking complement, union and/or intersection. For example, (a, b] = n=1 ( a, b + 1 ), (a, b) = n n=1 ( a, b 1 ]. n Thus, the collection all open intervals (closed intervals, half-open intervals or half lines) generates the same Borel field. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
7 The Borel field on R d, B d, is generated by all open hypercubes: (a 1, b 1 ) (a 2, b 2 ) (a d, b d ). B d can be generated by all closed hypercubes: [a 1, b 1 ] [a 2, b 2 ] [a d, b d ], or by (, b 1 ] (, b 2 ] (, b d ]. The sets that generate the Borel field B d are all Borel sets. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
8 Random Variable A random variable z defined on (Ω, F, IP) is a function z : Ω R such that for every B in the Borel field B, its inverse image is in F: z 1 (B) = {ω : z(ω) B} F. That is, z is a F/B-measurable (or simply F-measurable) function. Given ω, the resulting value z(ω) is known as a realization of z. A R d valued random variable (random vector) z defined on (Ω, F, IP) is: z: Ω R d such that for every B B d, z 1 (B) = {ω : z(ω) B} F; i.e., z is a F/B d -measurable function. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
9 Borel Measurable All the inverse images of random vector z, z 1 (B), form a σ-algebra, denoted as σ(z). It is known as the σ-algebra generated by z, or the information set associated with z. It is the smallest σ-algebra in F such that z is measurable. A function g : R R is B-measurable or Borel measurable if {ζ R: g(ζ) b} B. For random variable z defined on (Ω, F, IP) and Borel measurable function g( ), g(z) is a random variable defined on (Ω, F, IP). The same conclusion holds for d-dimensional random vector z and B d -measurable function g( ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
10 Distribution Function The joint distribution function of z is a non-decreasing, right-continuous function F z such that for ζ = (ζ 1,..., ζ d ) R d, F z (ζ) = IP{ω Ω: z 1 (ω) ζ 1,..., z d (ω) ζ d }, with lim F z(ζ) = 0, ζ 1,..., ζ d lim F z(ζ) = 1. ζ 1,..., ζ d The marginal distribution function of the i th component of z is F zi (ζ i ) = IP{ω Ω: z i (ω) ζ i } = F z (,...,, ζ i,,..., ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
11 Independence y and z are (pairwise) independent iff for any Borel sets B 1 and B 2, IP(y B 1 and z B 2 ) = IP(y B 1 ) IP(z B 2 ). A sequence of random variables {z i } is totally independent if ( ) IP {z i B i } = IP(z i B i ). all i all i Lemma 5.1 Let {z i } be a sequence of independent random variables and h i, i = 1, 2,... be Borel-measurable functions. Then {h i (z i )} is also a sequence of independent random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
12 Expectation The expectation of Z i is the Lebesgue integral of z i wrt to IP: IE(z i ) = z i (ω) d IP(ω). Ω In terms of its distribution function, IE(z i ) = ζ i df z (ζ) = ζ i df zi (ζ i ). R d R For Borel measurable function g( ) of z, IE[g(z)] = g(z(ω)) d IP(ω) = g(ζ) df z (ζ). Ω R d For example, the covariance matrix of z IE(zz ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
13 A function g is convex on a set S if for any a [0, 1] and any x, y in S, g (ax + (1 a)y) ag(x) + (1 a)g(y); g is concave on S if the inequality above is reversed. Lemma 5.2 (Jensen) Let g be a convex function on the support of z. For an integrable random variable z such that g(z) is integrable, g(ie(z)) IE[g(z)]; the inequality reverses if g is concave. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
14 L p -Norm For random variable z with finite p th moment, its L p -norm is: z p = [IE(z p )] 1/p. The inner product of square integrable random variables z i and z j is: z i, z j = IE(z i z j ). The L 2 -norm of z i can be obtained as z i 2 = z i, z i 1/2. For any c > 0 and p > 0, note that c p IP( z c) = c p 1 {ζ: ζ c} df z (ζ) ζ p df z (ζ) IE z p, {ζ: ζ c} where 1 A is the indicator function of the event A. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
15 Inequalities Lemma 5.3 (Markov) Let z be a random variable with finite p th moment. Then, IP( z c) IE z p c p, where c is a positive real number. For p = 2, Markov s inequality is also known as Chebyshev s inequality. Markov s inequality is trivial if c is small such that IE z p /c p > 1. When c becomes large, the probability that z assumes very extreme values will be vanishing at the rate c p. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
16 Lemma 5.4 (Hölder) Let y be a random variable with finite p th moment (p > 1) and z a random variable with finite q th moment (q = p/(p 1)). Then, IE yz y p z q. Since IE(yz) IE yz, we also have: Lemma 5.5 (Cauchy-Schwatz) Let y and z be two square integrable random variables. Then, IE(yz) y 2 z 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
17 Let y = 1 and x = z p. For q > p and r = q/p, by Hölder s inequality, IE z p x r y r/(r 1) = [IE(z pr )] 1/r = [IE(z q )] p/q. Lemma 5.6 (Liapunov) Let z be a random variable with finite q th moment. Then for p < q, z p z q. Lemma 5.7 (Minkowski) Let z i, i = 1,..., n, be random variables with finite p th moment (p 1). Then, n i=1 z i p n i=1 z i p. When n = 2, this is just the triangle inequality for L p -norms. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
18 Conditional Distributions Given A, B F, suppose we know B has occurred. Given the outcome space is restricted to B, the likelihood of A is characterized by the conditional probability: IP(A B) = IP(A B)/ IP(B). The conditional density function of z given y = η is f z y (ζ y = η) = f z,y(ζ, η). f y (η) f z y (ζ y = η) is clearly non-negative. Also R d f z y (ζ y = η)dζ = 1 f y (η) R d f z,y (ζ, η)dζ = 1 f y (η) f y(η) = 1. That is, f z y (ζ y = η) is a legitimate density function. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
19 Given the conditional density function f z y, for A B d, IP(z A y = η) = f z y (ζ y = η)dζ. A This probability is defined even when IP(y = η) is zero. When A = (, ζ 1 ] (, ζ d ], the conditional distribution function is F z y (ζ y = η) = IP(z 1 ζ 1,..., z d ζ d y = η). When z and y are independent, the conditional density (distribution) reduces to the unconditional density (distribution). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
20 Let G be a sub-σ-algebra of F, the conditional expectation IE(z G) is the integrable and G-measurable random variable satisfying IE(z G) d IP = z d IP, G G. G G Suppose that G is the trivial σ-algebra {Ω, }, then IE(z G) must be a constant c, so that IE(z) = z d IP = Ω Ω c d IP = c. Consider G = σ(y), the σ-algebra generated by y. IE(z y) = IE[z σ(y)], which is interpreted as the prediction of z given all the information associated with y. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
21 By definition, IE[IE(z G)] = Ω IE(z G) d IP = Ω z d IP = IE(z); That is, only a smaller σ-algebra matters in conditional expectation. Lemma 5.9 (Law of Iterated Expectations) Let G and H be two sub-σ-algebras of F such that G H. Then for the integrable random vector z, IE[IE(z H) G] = IE[IE(z G) H] = IE(z G). If z is G-measurable, then IE[g(z)x G] = g(z) IE(x G) with prob. 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
22 Lemma 5.11 Let z be a square integrable random variable. Then IE[z IE(z G)] 2 IE(z z) 2, for any G-measurable random variable z. Proof: For any square integrable, G-measurable random variable z, IE ( [z IE(z G)] z ) = IE ( [IE(z G) IE(z G)] z ) = 0. It follows that IE(z z) 2 = IE[z IE(z G) + IE(z G) z] 2 = IE[z IE(z G)] 2 + IE[IE(z G) z] 2 IE[z IE(z G)] 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
23 The conditional variance-covariance matrix of z given y is var(z y) = IE ( [z IE(z y)][z IE(z y)] y ) = IE(zz y) IE(z y) IE(z y), which leads to decomposition of analysis of variance: var(z) = IE[var(z y)] + var ( IE(z y) ). Example 5.12: Suppose that [ ] ([ ] [ y µ N y, x µ x Σ yy Σ xy Σ xy Σ xx ]). Then, IE(y x) = µ y Σ xyσ 1 xx (x µ x ), var(y x) = var(y) var ( IE(y x) ) = Σ yy Σ xyσ 1 xx Σ xy. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
24 Almost Sure Convergence A sequence of random variables, {z n ( )} n=1,2,..., is such that for a given ω, z n (ω) is a realization of the random element ω with index n, and that for a given n, z n ( ) is a random variable. Almost Sure Convergence Suppose {z n } and z are all defined on (Ω, F, IP). {z n } is said to converge to z almost surely if, and only if, IP(ω : z n (ω) z(ω) as n ) = 1, denoted as z n a.s. z or z n z a.s. (with prob. 1). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
25 Lemma 5.13 a.s. Let g : R R be a function continuous on S g R. If z n z, where z is a random variable such that IP(z S g ) = 1, then g(z n ) a.s. g(z). Proof: Let Ω 0 = {ω : z n (ω) z(ω)} and Ω 1 = {ω : z(ω) S g }. Thus, for ω (Ω 0 Ω 1 ), continuity of g ensures that g(z n (ω)) g(z(ω)). Note that (Ω 0 Ω 1 ) c = Ω c 0 Ω c 1, which has probability zero because IP(Ω c 0 ) = IP(Ωc 1 ) = 0. (Why?) It follows that Ω 0 Ω 1 has probability one, showing that g(z n ) g(z) with probability one. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
26 Convergence in Probability Convergence in Probability {z n } is said to converge to z in probability if for every ɛ > 0, lim IP(ω : z n n(ω) z(ω) > ɛ) = 0, or equivalently, lim n IP(ω : z n (ω) z(ω) ɛ) = 1. This is denoted as IP z or z n z in probability. z n Note: In this definition, the events Ω n (ɛ) = {ω : z n (ω) z(ω) ɛ} may vary with n, and convergence is referred to the probability of such events: p n = IP(Ω n (ɛ)), rather than the random variables z n. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
27 Almost sure convergence implies convergence in probability. To see this, let Ω 0 denote the set of ω such that z n (ω) z(ω). For ω Ω 0, there is some m such that ω is in Ω n (ɛ) for all n > m. That is, Ω 0 m=1 n=m Ω n (ɛ) F. As n=mω n (ɛ) is non-decreasing in m, it follows that ( ) IP(Ω 0 ) IP Ω n (ɛ) = lim m IP m=1 n=m ( n=m Ω n (ɛ) ) lim m IP( Ω m (ɛ) ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
28 Example 5.15 Let Ω = [0, 1] and IP be the Lebesgue measure. Consider the sequence of intervals {I n } in [0, 1]: [0, 1/2), [1/2, 1], [0, 1/3), [1/3, 2/3), [2/3, 1],..., and let z n = 1 In. When n tends to infinity, I n shrinks toward a singleton. For 0 < ɛ < 1, we have IP( z n > ɛ) = IP(I n ) 0, IP which shows z n 0. On the other hand, each ω [0, 1] must be covered by infinitely many intervals, so that z n (ω) = 1 for infinitely many n. This shows that z n (ω) does not converge to zero. Note: Convergence in probability permits z n to deviate from the probability limit infinitely often, but almost sure convergence does not, except for those ω in the set of probability zero. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
29 Lemma 5.16 Let {z n } be a sequence of square integrable random variables. If IP IE(z n ) c and var(z n ) 0, then z n c. Lemma 5.17 Let g : R R be a function continuous on S g R. If z n z, where z IP is a random variable such that IP(z S g ) = 1, then g(z n ) g(z). Proof: By the continuity of g, for each ɛ > 0, we can find a δ > 0 s.t. {ω : z n (ω) z(ω) δ} {ω : z(ω) S g } {ω : g(z n (ω)) g(z(ω)) ɛ}. Taking complementation of both sides, we have IP( g(z n ) g(z) > ɛ) IP( z n z > δ) 0. IP C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
30 Lemma 5.13 and Lemma 5.17 are readily generalized to R d -valued random a.s. IP variables. For instance, z n z (z n z) implies z 1,n + z 2,n z 1,n z 2,n z 2 1,n + z 2 2,n a.s. ( ) IP z 1 + z 2, a.s. ( ) IP z 1 z 2, a.s. ( ) IP z1 2 + z2 2, where z 1,n, z 2,n are two elements of z n and z 1, z 2 are the corresponding elements of z. Also, provided that z 2 0 with probability one, z 1,n /z 2,n a.s. ( ) IP z 1 /z 2. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
31 Convergence in Distribution Convergence in Distribution {z n } is said to converge to z in distribution, denoted as z n D z, if lim F n z n (ζ) = F z (ζ), for every continuity point ζ of F z. We also say that z n is asymptotically distributed as F z, denoted as z n A Fz ; F z is thus known as the limiting distribution of z n. Cramér-Wold Device. Let {z n } be a sequence of random vectors in R d D. Then z n z if and only if α D z n α z for every α R d such that α α = 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
32 Lemma 5.19 If z n IP z, then z n D z. For a constant c, zn IP c iff z n D c. Proof: For some arbitrary ɛ > 0 and a continuity point ζ of F z, we have IP(z n ζ) = IP({z n ζ} { z n z ɛ}) + IP({z n ζ} { z n z > ɛ}) IP(z ζ + ɛ) + IP( z n z > ɛ). IP Similarly, IP(z ζ ɛ) IP(z n ζ) + IP( z n z > ɛ). If z n z, then by passing to the limit and noting that ɛ is arbitrary, lim IP(z n n ζ) = IP(z ζ). That is, F zn (ζ) F z (ζ). The converse is not true in general, however. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
33 Theorem 5.20 (Continuous Mapping Theorem) Let g : R R be a function continuous almost everywhere on R, except D for at most countably many points. If z n z, then g(zn ) D g(z). For example, z n Theorem 5.21 D N (0, 1) implies z 2 n D χ 2 (1). Let {y n } and {z n } be two sequences of random vectors such that IP D D y n z n 0. If z n z, then yn z. Theorem 5.22 If y n converges in probability to a constant c and z n converges in D D c + z, yn z n cz, and zn /y n distribution to z, then y n + z n if c 0. D z/c C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
34 Non-Stochastic Order Notations Order notations are used to describe the behavior of real sequences. b n is (at most) of order c n, denoted as b n = O(c n ), if there exists a < such that b n /c n for all sufficiently large n. b n is of smaller order than c n, denoted as b n = o(c n ), if b n /c n 0. An O(1) sequence in bounded; an o(1) sequence converges to zero. The product of O(1) and o(1) sequences is o(1). Theorem 5.23 (a) If a n = O(n r ) and b n = O(n s ), then a n b n = O(n r+s ), a n + b n = O(n max(r,s) ). (b) If a n = o(n r ) and b n = o(n s ), then a n b n = o(n r+s ), a n + b n = o(n max(r,s) ). (c) If a n = O(n r ) and b n = o(n s ), then a n b n = o(n r+s ), a n + b n = O(n max(r,s) ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
35 Stochastic Order Notations The order notations defined earlier easily extend to describe the behavior of sequences of random variables. {z n } is O a.s. (c n ) (or O(c n ) almost surely) if z n /c n is O(1) a.s. {z n } is O IP (c n ) (or O(c n ) in probability) if for every ɛ > 0, there is some such that IP( z n /c n ) ɛ, for all n sufficiently large. Lemma 5.23 holds for stochastic order notations. For example, y n = O IP (1) and z n = o IP (1), then y n z n is o IP (1). It is very restrictive to require a random variable being bounded almost surely, but a well defined random variable is typically bounded in probability, i.e., O IP (1). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
36 D Let {z n } be a sequence of random variables such that z n z and ζ be a continuity point of F z. Then for any ɛ > 0, we can choose a sufficiently D large ζ such that IP( z > ζ) < ɛ/2. As z n z, we can also choose n large enough such that IP( z n > ζ) IP( z > ζ) < ɛ/2, which implies IP( z n > ζ) < ɛ. We have proved: Lemma 5.24 Let {z n } be a sequence of random variables such that z n z n = O IP (1). D z. Then C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
37 Law of Large Numbers When a law of large numbers holds almost surely, it is a strong law of large numbers (SLLN); when a law of large numbers holds in probability, it is a weak law of large numbers (WLLN). A sequence of random variables obeys a LLN when its sample average essentially follows its mean behavior; random irregularities (deviations from the mean) are wiped out in the limit by averaging. Kolmogorov s SLLN : Let {z t } be a sequence of i.i.d. random variables with mean µ o. Then, T 1 T t=1 z a.s. t µ o. Note that i.i.d. random variables need not obey Kolmogorov s SLLN if they do not have a finite mean, e.g., i.i.d. Cauchy random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
38 Theorem 5.26 (Markov s SLLN) Let {z t } be a sequence of independent random variables such that for some δ > 0, IE z t 1+δ is bounded for all t. Then, 1 T T [z t IE(z t )] a.s. 0. t=1 Note that here z t need not have a common mean, and the average of their means need not converge. Compared with Kolmogorov s SLLN, Markov s SLLN requires a stronger moment condition but not identical distribution. A LLN usually obtains by regulating the moments of and dependence across random variables. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
39 Examples Example 5.27 Suppose that y t = α o y t 1 + u t with α o < 1. Then, var(y t ) = σ 2 u/(1 α 2 o), and cov(y t, y t j ) = α j o ( T ) var y t = t=1 T t=1 σ2 u 1 α 2 o. Thus, T 1 var(y t ) + 2 (T τ) cov(y t, y t τ ) T var(y t ) + 2T t=1 ( so that var T 1 ) T t=1 y t 1 T T t=1 y t IP 0. τ=1 T 1 τ=1 by Lemma That is, {y t } obeys a WLLN. cov(y t, y t τ ) = O(T ), = O(T 1 ). As IE(T 1 T t=1 y t) = 0, C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
40 Lemma 5.28 Let y t = i=0 π iu t i, where u t are i.i.d. random variables with mean zero and variance σu. 2 If i= π i <, then T 1 T t=1 y a.s. t 0. In Example 5.27, y t = i=0 αi ou t i with α o < 1, so that i=0 αi o < Lemma 5.28 is quite general and applicable to processes that can be expressed as an MA process with absolutely summable weights, e.g., weakly stationary AR(p) processes. For random variables with strong correlations over time, the variation of their partial sums may grow too rapidly and cannot be eliminated by simple averaging. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
41 Example 5.29: For the sequences {t} and {t 2 }, T t=1 t = T (T + 1)/2, T t=1 t2 = T (T + 1)(2T + 1)/6. Hence, T 1 T t=1 t and T 1 T t=1 t2 both diverge. Example 5.30: u t are i.i.d. with mean zero and variance σ 2 u. Consider now {tu t }, which does not have bounded (1 + δ) th moment and does not obey Markov s SLLN. Moreover, ( T ) T var tu t = t 2 var(u t ) = σu 2 t=1 t=1 T (T + 1)(2T + 1), 6 so that T t=1 tu t = O IP (T 3/2 ). It follows that T 1 T t=1 tu t = O IP (T 1/2 ). That is, {tu t } does not obey a WLLN. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
42 Example 5.31: y t is a random walk: y t = y t 1 + u t. For s < t, y t = y s + t i=s+1 u i = y s + v t s, where v t s is independent of y s and cov(y t, y s ) = IE(y 2 s ) = sσ 2 u. Thus, ( T ) var y t = t=1 T t=1 T 1 var(y t ) + 2 T τ=1 t=τ+1 for T t=1 var(y t) = T t=1 tσ2 u = O(T 2 ) and T 1 2 T τ=1 t=τ+1 T 1 cov(y t, y t τ ) = 2 T τ=1 t=τ+1 cov(y t, y t τ ) = O(T 3 ), (t τ)σ 2 u = O(T 3 ). Then, T t=1 y t = O IP (T 3/2 ) and T 1 T t=1 y t diverges in probability. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
43 Example 5.32: y t is the random walk in Example Then, IE(y t 1 u t ) = 0, var(y t 1 u t ) = IE(y 2 t 1 ) IE(u2 t ) = (t 1)σ 4 u, and for s < t, cov(y t 1 u t, y s 1 u s ) = IE(y t 1 y s 1 u s ) IE(u t ) = 0. This yields ( T ) var y t 1 u t = t=1 T var(y t 1 u t ) = t=1 T (t 1)σu 4 = O(T 2 ), and T t=1 y t 1u t = O IP (T ). As var(t 1 T t=1 y t 1u t ) converges to σ 4 u/2, rather than 0, {y t 1 u t } does not obey a WLLN, even though its partial sums are O IP (T ). t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
44 Central Limit Theorem (CLT) Lemma 5.35 (Lindeberg-Lévy s CLT) Let {z t } be a sequence of i.i.d. random variables with mean µ o and variance σo 2 > 0. Then, D T ( z T µ o )/σ o N (0, 1). i.i.d. random variables need not obey this CLT if they do not have a finite variance, e.g., t(2) r.v. Note that z T converges to µ o in probability, and its variance σo/t 2 vanishes when T tends to infinity. A normalizing factor T 1/2 suffices to prevent a degenerate distribution in the limit. When {z t } obeys a CLT, z T is said to converge to µ o at the rate T 1/2, and z T is understood as a root-t consistent estimator. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
45 Lemma 5.36 (Liapunov s CLT) Let {z Tt } be a triangular array of independent random variables with mean µ Tt and variance σtt 2 > 0 such that σ2 T = 1 T T t=1 σ2 Tt σ2 o > 0. If for some δ > 0, IE z Tt 2+δ are bounded, then D T ( z T µ T )/σ o N (0, 1). A CLT usually requires stronger conditions on the moment of and dependence across random variables than those needed to ensure a LLN. Moreover, every random variable must also be asymptotically negligible, in the sense that no random variable is influential in affecting the partial sums. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
46 Examples Example 5.37: {u t } is a sequence of independent random variables with mean zero, variance σu, 2 and bounded (2 + δ) th moment. we know var( T t=1 tu t) is O(T 3 ), which implies that variance of T 1/2 T t=1 tu t is diverging at the rate O(T 2 ). On the other hand, observe that ( 1 T var T 1/2 It follows that 3 T 1/2 σ u T t=1 t=1 ) t T u t = t T u t D N (0, 1). T (T + 1)(2T + 1) 6T 3 σu 2 σ2 u 3. These results show that {(t/t )u t } obeys a CLT, whereas {tu t } does not. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
47 Example 5.38: y t is a random walk: y t = y t 1 + u t, where u t are i.i.d. with mean zero and variance σu. 2 We know y t do not obey a LLN and hence do not obey a CLT. CLT for Triangular Array {z Tt } is a triangular array of random variables and obeys a CLT if 1 σ o T T [z Tt IE(z Tt )] = t=1 T ( zt µ T ) σ o D N (0, 1), where z T = T 1 T t=1 z Tt, µ T = IE( z T ), and ( T ) σt 2 = var T 1/2 z Tt σo 2 > 0. t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
48 Consider an array of square integrable random vectors z Tt in R d. Let z T denote the average of z Tt, µ T = IE( z T ), and ( 1 Σ T = var T T t=1 z Tt ) Σ o, a positive definite matrix. Using the Cramér-Wold device, {z Tt } is said to obey a multivariate CLT, in the sense that Σ 1/2 o 1 T T t=1 [z Tt IE(z Tt )] = Σ 1/2 o T ( zt µ T ) D N (0, I d ), if {α z Tt } obeys a CLT, for any α R d such that α α = 1. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
49 Stochastic Processes A d-dimensional stochastic process with the index set T is a measurable mapping z: Ω (R d ) T such that z(ω) = {z t (ω), t T }. For each t T, z t ( ) is a R d -valued r.v.; for each ω, z(ω) is a sample path (realization) of z, a R d -valued function on T. The finite-dimensional distributions of {z(t, ), t T } is IP(z t1 a 1,..., z tn a n ) = F t1,...,t n (a 1,..., a n ). z is stationary if F t1,...,t n are invariant under index displacement. z is Gaussian if F t1,...,t n are all (multivariate) normal. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
50 Brownian motion The process {w(t), t [0, )} is the standard Wiener process (standard Brownian motion) if it has continuous sample paths almost surely and satisfies: 1 IP ( w(0) = 0 ) = 1. 2 For 0 t 0 t 1 t k, IP ( w(t i ) w(t i 1 ) B i, i k ) = i k IP( ) w(t i ) w(t i 1 ) B i, where B i are Borel sets. 3 For 0 s < t, w(t) w(s) N (0, t s). Note: w here has independent and Gaussian increments. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
51 w(t) N (0, t) such that for r t, cov ( w(r), w(t) ) = IE [ w(r) ( w(t) w(r) )] + IE [ w(r) 2] = r. The sample paths of w are a.s. continuous but highly irregular (nowhere differentiable). To see this, note w c (t) = w(c 2 t)/c for c > 0 is also a standard Wiener process. (Why?) Then, w c (1/c) = w(c)/c. For a large c such that w(c)/c > 1, wc(1/c) 1/c = w(c) > c. That is, the sample path of w c has a slope larger than c on a very small interval (0, 1/c). The difference quotient: [w(t + h) w(t)]/h N (0, 1/ h ) can not converge to a finite limit (as h 0) with a positive prob. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
52 The d-dimensional, standard Wiener process w consists of d mutually independent, standard Wiener processes, so that for s < t, w(t) w(s) N (0, (t s) I d ). Lemma 5.39 Let w be the d-dimensional, standard Wiener process. 1 w(t) N (0, t I d ). 2 cov(w(r), w(t)) = min(r, t) I d. The Brownian bridge w 0 on [0, 1] is w 0 (t) = w(t) tw(1). Clearly, IE[w 0 (t)] = 0, and for r < t, cov ( w 0 (r), w 0 (t) ) = cov ( w(r) rw(1), w(t) tw(1) ) = r(1 t) I d. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
53 Weak Convergence IP n converges weakly to IP, denoted as IP n IP, if for every bounded, continuous real function f on S, f (s) dip n (s) f (s) d IP(s), where {IP n } and IP are probability measures on (S, S). When z n and z are all R d -valued random variables, IP n IP reduces D to the usual notion of convergence in distribution: z n z. When z n and z are d-dimensional stochastic processes with the D distributions induced by IP n and IP, z n z, also denoted as zn z, implies that all the finite-dimensional distributions of z n converge to the corresponding distributions of z. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
54 Continuous Mapping Theorem Lemma 5.40 (Continuous Mapping Theorem) Let g : R d R be a function continuous almost everywhere on R d, except for at most countably many points. If z n z, then g(z n ) g(z). Proof: Let S and S be two metric spaces with Borel σ-algebras S and S and g : S S be a measurable mapping. For IP on (S, S), define IP on (S, S ) as IP (A ) = IP(g 1 (A )), A S. For every bounded, continuous f on S, f g is also bounded and continuous on S. IP n IP now implies that f g(s) dip n (s) f g(s) d IP(s), which is equivalent to f (a) dip n(a) f (a) dip (a), proving IP n IP. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
55 Functional Central Limit Theorem (FCLT) ζ i are i.i.d. with mean zero and variance σ 2. Let s n = ζ ζ n and z n (i/n) = (σ n) 1 s i. For t [(i 1)/n, i/n), the constant interpolations of z n (i/n) is z n (t) = z n ((i 1)/n) = 1 σ n s [nt], where [nt] is the the largest integer less than or equal to nt. From Lindeberg-Lévy s CLT, 1 σ n s [nt] = ( [nt] n ) 1/2 1 σ [nt] s D [nt] t N (0, 1), which is just N (0, t), the distribution of w(t). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
56 For r < t, we have (z n (r), z n (t) z n (r)) D ( w(r), w(t) w(r) ), and hence (z n (r), z n (t)) D (w(r), w(t)). This is easily extended to establish convergence of any finite-dimensional distributions and leads to the functional central limit theorem. Lemma 5.41 (Donsker) Let ζ t be i.i.d. with mean µ o and variance σ 2 o > 0 and z T (r) = 1 σ o T [Tr] (ζ t µ o ), r [0, 1]. t=1 Then, z T w as T. C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
57 Let ζ t be r.v.s with mean µ t and variance σ 2 t > 0. Define long-run variance of ζ t as σ 2 = lim T var ( 1 T T t=1 ζ t ) {ζ t } is said to obey an FCLT if z T w as T, where z T (r) = 1 σ T [Tr] ( ) ζt µ t, r [0, 1]. t=1 In the multivariate context, FCLT is z T w as T, where z T (r) = 1 Σ 1/2 ( ) ζt µ t, r [0, 1], T [Tr] t=1 w is the d-dimensional, standard Wiener process, and ( T ) ( 1 Σ = lim T T IE T ) (ζ t µ t ) (ζ t µ t ), t=1 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58, t=1
58 Example 5.43 y t = y t 1 + u t, t = 1, 2,..., with y 0 = 0, where u t are i.i.d. with mean zero and variance σ 2 u. By Donsker s FCLT, the partial sum y [Tr] = [Tr] t=1 u t is such that 1 T 3/2 T y t = σ u t=1 T t/t t=1 (t 1)/T 1 1 y [Tr] dr σ u w(r) dr, T σu 0 This result also verifies that T t=1 y t is O IP (T 3/2 ). Similarly, 1 T 2 T yt 2 = 1 T t=1 T t=1 ( yt ) 2 1 σ 2 u w(r) 2 dr, T 0 so that T t=1 y 2 t is O IP (T 2 ). C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, / 58
Elements of Probability Theory
Chapter 5 Elements of Probability Theory The purpose of this chapter is to summarize some important concepts and results in probability theory. Of particular interest to us are the limit theorems which
More informationLarge Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the
More informationAsymptotic Least Squares Theory
Asymptotic Least Squares Theory CHUNG-MING KUAN Department of Finance & CRETA December 5, 2011 C.-M. Kuan (National Taiwan Univ.) Asymptotic Least Squares Theory December 5, 2011 1 / 85 Lecture Outline
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationProbability Theory I: Syllabus and Exercise
Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will
More informationLecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1
Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).
More informationThe main results about probability measures are the following two facts:
Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More informationProbability and Measure
Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability
More informationEconomics 620, Lecture 8: Asymptotics I
Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators
More informationWeak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij
Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More information4 Expectation & the Lebesgue Theorems
STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does
More information1 Introduction. 2 Diffusion equation and central limit theorem. The content of these notes is also covered by chapter 3 section B of [1].
1 Introduction The content of these notes is also covered by chapter 3 section B of [1]. Diffusion equation and central limit theorem Consider a sequence {ξ i } i=1 i.i.d. ξ i = d ξ with ξ : Ω { Dx, 0,
More information9 Brownian Motion: Construction
9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of
More informationSTAT 7032 Probability. Wlodek Bryc
STAT 7032 Probability Wlodek Bryc Revised for Spring 2019 Printed: January 14, 2019 File: Grad-Prob-2019.TEX Department of Mathematical Sciences, University of Cincinnati, Cincinnati, OH 45221 E-mail address:
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationGAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM
GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationLECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT
MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationSome functional (Hölderian) limit theorems and their applications (II)
Some functional (Hölderian) limit theorems and their applications (II) Alfredas Račkauskas Vilnius University Outils Statistiques et Probabilistes pour la Finance Université de Rouen June 1 5, Rouen (Rouen
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationUnivariate Unit Root Process (May 14, 2018)
Ch. Univariate Unit Root Process (May 4, 8) Introduction Much conventional asymptotic theory for least-squares estimation (e.g. the standard proofs of consistency and asymptotic normality of OLS estimators)
More informationBrownian Motion and Conditional Probability
Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationHILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define
HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,
More information5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.
88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal
More informationMeasure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond
Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationProduct measure and Fubini s theorem
Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω
More informationIf g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get
18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More informationTHEOREMS, ETC., FOR MATH 515
THEOREMS, ETC., FOR MATH 515 Proposition 1 (=comment on page 17). If A is an algebra, then any finite union or finite intersection of sets in A is also in A. Proposition 2 (=Proposition 1.1). For every
More information5 Birkhoff s Ergodic Theorem
5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even
More informationReview (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology
Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna
More informationFE 5204 Stochastic Differential Equations
Instructor: Jim Zhu e-mail:zhu@wmich.edu http://homepages.wmich.edu/ zhu/ January 20, 2009 Preliminaries for dealing with continuous random processes. Brownian motions. Our main reference for this lecture
More informationLecture 1: Review of Basic Asymptotic Theory
Lecture 1: Instructor: Department of Economics Stanfor University Prepare by Wenbo Zhou, Renmin University Basic Probability Theory Takeshi Amemiya, Avance Econometrics, 1985, Harvar University Press.
More informationTools from Lebesgue integration
Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given
More informationENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM
c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationIntroduction to Stochastic processes
Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationProf. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis
Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation
More informationPart II Probability and Measure
Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationGaussian Processes. 1. Basic Notions
Gaussian Processes 1. Basic Notions Let T be a set, and X : {X } T a stochastic process, defined on a suitable probability space (Ω P), that is indexed by T. Definition 1.1. We say that X is a Gaussian
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationLecture 4 Lebesgue spaces and inequalities
Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how
More informationRS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction
Chapter 1 Probability Theory: Introduction Basic Probability General In a probability space (Ω, Σ, P), the set Ω is the set of all possible outcomes of a probability experiment. Mathematically, Ω is just
More informationECON 616: Lecture 1: Time Series Basics
ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters
More informationIn terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.
1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t
More informationRudin Real and Complex Analysis - Harmonic Functions
Rudin Real and Complex Analysis - Harmonic Functions Aaron Lou December 2018 1 Notes 1.1 The Cauchy-Riemann Equations 11.1: The Operators and Suppose f is a complex function defined in a plane open set
More informationStochastic Processes
Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space
More informationEE514A Information Theory I Fall 2013
EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/
More informationMATHS 730 FC Lecture Notes March 5, Introduction
1 INTRODUCTION MATHS 730 FC Lecture Notes March 5, 2014 1 Introduction Definition. If A, B are sets and there exists a bijection A B, they have the same cardinality, which we write as A, #A. If there exists
More informationCHAPTER 3: LARGE SAMPLE THEORY
CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually
More information18.175: Lecture 3 Integration
18.175: Lecture 3 Scott Sheffield MIT Outline Outline Recall definitions Probability space is triple (Ω, F, P) where Ω is sample space, F is set of events (the σ-algebra) and P : F [0, 1] is the probability
More informationGeneralized Method of Moment
Generalized Method of Moment CHUNG-MING KUAN Department of Finance & CRETA National Taiwan University June 16, 2010 C.-M. Kuan (Finance & CRETA, NTU Generalized Method of Moment June 16, 2010 1 / 32 Lecture
More information7 Convergence in R d and in Metric Spaces
STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a
More informationLecture 4: Introduction to stochastic processes and stochastic calculus
Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London
More informationLecture 17 Brownian motion as a Markov process
Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is
More informationReminder Notes for the Course on Measures on Topological Spaces
Reminder Notes for the Course on Measures on Topological Spaces T. C. Dorlas Dublin Institute for Advanced Studies School of Theoretical Physics 10 Burlington Road, Dublin 4, Ireland. Email: dorlas@stp.dias.ie
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information1. Stochastic Processes and Stationarity
Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed
More informationGradient Gibbs measures with disorder
Gradient Gibbs measures with disorder Codina Cotar University College London April 16, 2015, Providence Partly based on joint works with Christof Külske Outline 1 The model 2 Questions 3 Known results
More information4th Preparation Sheet - Solutions
Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space
More informationExample 4.1 Let X be a random variable and f(t) a given function of time. Then. Y (t) = f(t)x. Y (t) = X sin(ωt + δ)
Chapter 4 Stochastic Processes 4. Definition In the previous chapter we studied random variables as functions on a sample space X(ω), ω Ω, without regard to how these might depend on parameters. We now
More informationEC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015
EC402: Serial Correlation Danny Quah Economics Department, LSE Lent 2015 OUTLINE 1. Stationarity 1.1 Covariance stationarity 1.2 Explicit Models. Special cases: ARMA processes 2. Some complex numbers.
More informationBootstrap - theoretical problems
Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationReview of measure theory
209: Honors nalysis in R n Review of measure theory 1 Outer measure, measure, measurable sets Definition 1 Let X be a set. nonempty family R of subsets of X is a ring if, B R B R and, B R B R hold. bove,
More informationHomework 11. Solutions
Homework 11. Solutions Problem 2.3.2. Let f n : R R be 1/n times the characteristic function of the interval (0, n). Show that f n 0 uniformly and f n µ L = 1. Why isn t it a counterexample to the Lebesgue
More informationBasic Measure and Integration Theory. Michael L. Carroll
Basic Measure and Integration Theory Michael L. Carroll Sep 22, 2002 Measure Theory: Introduction What is measure theory? Why bother to learn measure theory? 1 What is measure theory? Measure theory is
More informationECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes
ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ED HERBST September 11, 2017 Background Hamilton, chapters 15-16 Trends vs Cycles A commond decomposition of macroeconomic time series
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More informationMi-Hwa Ko. t=1 Z t is true. j=0
Commun. Korean Math. Soc. 21 (2006), No. 4, pp. 779 786 FUNCTIONAL CENTRAL LIMIT THEOREMS FOR MULTIVARIATE LINEAR PROCESSES GENERATED BY DEPENDENT RANDOM VECTORS Mi-Hwa Ko Abstract. Let X t be an m-dimensional
More informationExercises in stochastic analysis
Exercises in stochastic analysis Franco Flandoli, Mario Maurelli, Dario Trevisan The exercises with a P are those which have been done totally or partially) in the previous lectures; the exercises with
More informationLebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?
Lebesgue Integration: A non-rigorous introduction What is wrong with Riemann integration? xample. Let f(x) = { 0 for x Q 1 for x / Q. The upper integral is 1, while the lower integral is 0. Yet, the function
More informationNotes 1 : Measure-theoretic foundations I
Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More information1 Independent increments
Tel Aviv University, 2008 Brownian motion 1 1 Independent increments 1a Three convolution semigroups........... 1 1b Independent increments.............. 2 1c Continuous time................... 3 1d Bad
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationEmpirical Processes: General Weak Convergence Theory
Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated
More informationProbability and Measure
Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real
More informationLecture 2: Convergence of Random Variables
Lecture 2: Convergence of Random Variables Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Introduction to Stochastic Processes, Fall 2013 1 / 9 Convergence of Random Variables
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More information(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define
Homework, Real Analysis I, Fall, 2010. (1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define ρ(f, g) = 1 0 f(x) g(x) dx. Show that
More informationStochastic Processes
Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic
More informationConditional independence, conditional mixing and conditional association
Ann Inst Stat Math (2009) 61:441 460 DOI 10.1007/s10463-007-0152-2 Conditional independence, conditional mixing and conditional association B. L. S. Prakasa Rao Received: 25 July 2006 / Revised: 14 May
More informationMultiple Random Variables
Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x
More informationSummary of Real Analysis by Royden
Summary of Real Analysis by Royden Dan Hathaway May 2010 This document is a summary of the theorems and definitions and theorems from Part 1 of the book Real Analysis by Royden. In some areas, such as
More information