Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
|
|
- Adele Harrison
- 6 years ago
- Views:
Transcription
1 Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the properties of estimators when the sample size becomes arbitrarily large. The idea is that given a reonably large datet, the properties of an estimator even when the sample size is finite are similar to the properties of an estimator when the sample size is arbitrarily large. In these notes we focus on the large sample properties of sample averages formed from i.i.d. data. That is, sume that X i i.i.d.f, for i = 1,..., n,.... Assume EX i = µ, for all i. The sample average after n draws is X n 1 n i X i. We focus on two important sets of large sample results: (1) Law of large numbers: Xn EX n. () Central limit theorem: n( X n EX) N(0, ). That is, n times a sample average looks like (in a precise sense to be defined later) a normal random variable n gets large. An important endeavor of ymptotic statistics is to show (1) and () under various sumptions on the data sampling process. Consider a sequence of random variables Z 1, Z,...,. Convergence in probability: p Z for all ϵ > 0, limn P rob ( Z < ϵ) = 1. More formally: for all ϵ > 0 and δ > 0, there exists n 0 (ϵ, δ) such that for all n > n 0 (ϵ, δ): P rob ( Z < ϵ) > 1 δ. Note that the limiting variable Z can also be a random variable. Furthermore, for convergence in probability, the random variables Z 1, Z,... and Z should be defined on the same probability space: if this common sample space is Ω with elements ω, then the statement of convergence in probability is that: P rob (ω Ω : (ω) Z(ω) < ϵ) > 1 δ. Corresponding to this convergence concept, we have the Weak Law of Large Numbers (WLLN), which is the result that X n p µ. 1
2 Earlier, we had used Chebyshev s inequality to prove a version of the WLLN, under the sumption that X i are i.i.d. with mean µ and variance σ. Khinchine s law of large numbers only requires that the mean exists (i.e., is finite), but does not require existence of variances. Recall Markov s inequality: for positive random variables Z, and p > 0, ϵ > 0, we have P (Z > ϵ) E(Z) p /ϵ p. Take Z = X n X. Then if we have: E[ X n X p ] 0 : that is, X n converges in p-th mean to X, then by Markov s inequality, we also have P ( X n X ϵ) 0. That is, convergence in p-th mean implies convergence in probability. Some definitions and results: Define: the probability limit of a random sequence, denoted plim n, is a non-random quantity α such that p α. Stochtic orders of magnitude: big-o.p. (bounded in probability): = O p (n λ ) for every δ, there exists a finite (δ) and n (δ) such that P rob( Zn > ) < δ for all n n λ n (δ). little-o.p. : = o p (n λ ) Zn n λ p 0. = o p (1) p 0. Plim operator theorem: Let be a k-dimensional random vector, and g( ) be a function which is continuous at a constant k-vector point α. Then p α = g(zn ) p g(α). In other words: g(plim ) = plim g( ). Proof: See Serfling, Approximation Theorems of Mathematical Statistics, p. 4. Don t confuse this: plim 1 n i g(z i) = Eg(Z i ), from the LLN, but this is distinct from plim g( 1 n i Z i) = g(ez i ), using plim operator result. The two are generally not the same; if g( ) is convex, then plim 1 n i g(z i) plim g( 1 n i Z i).
3 p A useful intermediate result ( squeezing ): if α (a constant) and Z n lies between and α with probability 1, then Zn p α. (Quick proof: for all ϵ > 0, X n α < ϵ Xn α < ϵ. Hence P rob( Xn α < ϵ) P rob( X n α < ϵ) 1.) Convergence almost surely Z P rob (ω : lim n (ω) = Z(ω)) = 1. As with convergence in probability, almost sure convergence makes sense when the random variables Z 1,...,,... and Z are all defined on the same sample space. Hence, for each element in the sample space ω Ω, we can sociate the sequence Z 1 (ω),..., (ω),... well the limit point Z(ω). To understand the probability statement more clearly, sume that all these RVs are defined on the sample space (Ω, B(Ω), P ). Define some set-theoretic concepts. Consider a sequence of sets S 1,..., S n, all B(Ω). Unless this sequence is monotonic, it s difficult to talk about a limit of this sequence. So we define the liminf and limsup of this set sequence. lim inf n S n lim n m=n S m = n=1 m=n S m = {ω Ω : ω S n, n n 0 (ω)}. Note that the sequence of sets m=n S m, for n = 1,, 3,... is a non-decreing sequence of sets. By taking the union, the liminf is this the limit of this monotone sequence of sets. That is, for all ω lim infs n, there exists some number n 0 (ω) such that ω S n for all n n 0 (ω). Hence we can say that lim infs n is the set of outcomes which occur eventually. lim sup n S n S m = lim n=1 m=n S m n m=n = { ω Ω : for every m, ω S n0 (ω,m) for some n 0 (ω, m) m }. 3
4 Note that the sequence of sets m=n S n, for n = 1,, 3,... is a non-increing sequence of sets. Hence, limsup is limit of this monotone sequence of sets. The lim sups n is the set of outcomes ω which occurs within every tail of the set sequence S n. Hence we say that an outcome ω lim sups n occurs infinitely often. 1 Note that Hence and liminf S n limsup S n. P (limsup S n ) P (liminf S n ) P (limsup S n ) = 0 P (liminf S n ) = 0 P (liminf S n ) = 1 P (limsup S n ) = 1. Borel-Cantelli Lemma: if i=1 P (S i) < then P (lim sup n S n ) = 0. Proof: for each m, we have P (lim sup n S n ) = P (lim { m=n S m }) = lim n P ( which equals zero (by the sumption that i=1 P (S i) < ). m=n S m ) lim P (S m ) n m n The first equality (interchange of limit and Prob operations) is an application of the Monotone Convergence Theorem, which holds only because the sequence m=n S m is monotone. To apply this to almost-sure convergence, consider the sequence of sets S n {ω : (ω) Z(ω) > ϵ}, for ϵ > 0. Almost sure convergence is the statement that P (limsup S n ) = 0: Then m=n S m denotes the ω s such that the sequence (ω) Z(ω) exceeds ϵ in the tail beyond n. 1 However, note that all outcomes ω lim infs n also occur an infinite number of times along the infinite sequence S n. 4
5 Then lim n m=n S m denotes all ω s such that the sequence (ω) Z(ω) exceeds ϵ in every tail: those ω for which (ω) escapes the ϵ-ball around Z(ω) infinitely often. For these ω s, lim n (ω) Z(ω). For almost-sure convergence, you require the probability of this set to equal zero, i.e. Pr(lim n m=n S m) = 0. Corresponding to this convergence concept, we have the Strong Law of Large Numbers (SLLN), which is the result that X n µ. Consider that the sample averages are random variables X 1 (ω), X (ω),... defined on the same probability space, say (Ω, B, P ), where each ω indexes a sequence a sequence X 1 (ω), X (ω), X 3 (ω),.... Consider the set sequence S n = { ω : X n (ω) µ > ϵ }. SLLN is the statement that P (limsups n ) = 0. Proof (sketch; Davidson, pg. 96): For convenience, consider µ = 0. From the above discussion, we see that the two statements X n 0 and P ( X n > ϵ, i.o.) = 0 for any ϵ > 0 are equivalent. Therefore, we proceed in proving the SLLN by verifying that n=1 P ( X n > ϵ) < for all ϵ > 0, and then apply the Borel-Cantelli Lemma. As a starting point, we see that Chebyshev s inequality is not good enough by itself; it tells us P ( X n > ϵ) σ nϵ ; but 1 nϵ =. However, consider the subsequence with indices n : {1, 4, 9, 16,...}. Chebyshev s inequality, we have P ( X n > ϵ) n=1 σ n ϵ ; and 1 n ϵ = 1 ϵ (π /6) 1.64 <. ϵ n=1 Again using Hence, by BC lemma, we have that X n 0. Next, examine the omitted terms from the sum n=1 P ( X n > ϵ). Define D n = max n k<(n+1) X k X n. Note that: ( ) X k X n n = k 1 X n + 1 k Another of Euler s greatest hits: i=1 1 n = π 6 k t=n +1 X t (the Riemann zeta function). 5
6 and, by independence of X i, the two terms on RHS are uncorrelated. Hence V ar( X k X n ) = ) (1 n σ k n + k n σ k ) = σ ( 1 n 1 k By Chebyshev s inequality, then, we have n P (D n > ϵ) σ ϵ σ ( 1 n 1 (n + 1) ( ) 1 n 1 ; so (n + 1) P (D n > ϵ) σ ( ) 1 ϵ n 1 < σ ( ) 1 (n + 1) ϵ n 1 = σ (n + 1) ϵ n implying, using BC Lemma, that D n Now, consider, for n l < (n + 1), n 0. X l = X l X n + X n X l X n + X n D n + X n. By the above discussion, the RHS converges a.s. to 0. Hence, for all integers l, we have that X l is bounded by a random variable which converges a.s. to 0. Thus, X l 0. Example: X i i.i.d. U[0, 1]. Show that X (1:n) min i=1,...,n X i 0. Take S n { X (1:n) > ϵ }. For all ϵ > 0, P ( X (1:n) > ϵ) = P (X (1:n) > ϵ) = P (X i > ϵ, i = 1,..., n) = (1 ϵ) n. Hence, n=1 (1 ϵ)n = 1/ϵ <. So the conclusion follows by the BC Lemma. Theorem: Z = p Z. ). 6
7 Proof: 0 = P ( lim = lim n P n m n ( m n {ω : Z m (ω) Z(ω) > ϵ} {ω : Z m (ω) Z(ω) > ϵ} lim n P ({ω : (ω) Z(ω) > ϵ}). ) ) Convergence in Distribution: d Z. A sequence of real-valued random variables Z 1, Z, Z 3,... converges in distribution to a random variable Z if lim F (z) = F Z (z) z s.t. F Z (z) is continuous. (1) n This is a statement about the CDFs of the random variables Z, and Z 1, Z,.... These random variables do not need to be defined on the same probability space. F Z is also called the limiting distribution of the random sequence. Alternative definitions of convergence in distribution ( Portmanteau theorem ): Letting F ([a, b]) F (b) F (a) for a < b, convergence (1) is equivalent to F Zn ([a, b]) F Z ([a, b]) [a, b] s.t. F Z continuous at a, b. For all bounded, continuous functions g( ), g(z)df Zn (z) g(z)df Z (z) Eg( ) Eg(Z). () This definition of distributional convergence is more useful in advanced settings, because it is extendable to setting where and Z are general random elements taking values in metric space. Levy s continuity theorem: A sequence {X n } of random variables converges in distribution to random variable X if and only if the sequence of characteristic function {ϕ Xj (t)} converges pointwise (in t) to a function ϕ X (t) which is continuous at the origin. Then ϕ X is the characteristic function of X. 7
8 Hence, this ties together convergence of characteristic functions, and convergence in distribution. This theorem will be used to prove the CLT later. For proofs of most results here, see Serfling, Approximation Theorems in Mathematical Statistics, ch. 1. Distributional convergence, defined above, is called weak convergence. This is because there are stronger notions of distributional convergence. One such notion is: sup A B(R) P Zn (A) P Z (A) 0 which is convergence in total variation norm. Example: Multinomial distribution on [0, 1]. X n { 1,, 3,..., n = 1} each with n n n n probability 1. Consider A = Q (set of rational numbers in [0, 1]). n Some definitions and results: Slutsky Theorem: If Z d p n Z, and Y n α (a constant), then (a) Y n Z d n αz (b) + Y d n Z + α. Theorem *: (a) p Z = Zn d Z (b) d Z = p Z if Z is a constant. Note: convergence in probability implies that (roughly speaking), the random variables (for n large enough) and Z frequently have the same numerical value. Convergence in distribution need not imply this, only that the CDF s of and Z are similar. Z d n Z = = O p (1). Use = max( F 1 1 Z (1 ϵ), FZ (ϵ) ) in definition of O p (1). Note that the LLN tells us that X p, n µ, which implies (trivially) that X n degenerate limiting distribution. d µ, a 8
9 This is not very useful for our purposes, because we are interested in knowing (say) how far X n is from µ, which is unknown. How do we fix X n, so that it h a non-degenerate limiting distribution? Central Limit Theorem: (Lindeberg-Levy) Let X i be i.i.d. with mean µ and variance σ. Then n( Xn µ) d N(0, 1). σ n is also called the rate of convergence of the sequence Xn. By definition, a rate of convergence is the lowest polynomial in n for which n p ( X n µ) converges in distribution to a nondegenerate distribution. The rate of convergence n make sense: if you blow up X n by a constant (no matter how big), you still get a degenerate limiting distribution. If you blow up by n, then the sequence S n n X n = n i=1 X n will diverge. Proof via characteristic functions: Consider X i i.i.d. with mean µ and variance σ. Define: n( Xn µ) = X i µ 1 W i. σ i nσ n Note that, W i (X i µ)/σ are i.i.d. with mean EW i = 0 and variance VW i = EW i = 1. Let ϕ(t) denote characteristic function. Then [ n [ ϕ Zn (t) = ϕ n W (t)] = ϕw (t/ n) ] n. Using the second-order Taylor expansion for ϕ W (t/ n) around 0, we have: ϕ W (t/ n) = ϕ W (0) + t ( ) iew + t t n n i EW + o n ( ) = ϕ W (0) + 0 t t n 1 + o n i 9
10 Now we have ( )} log ϕ Zn(t)(t) = n log {ϕ W (0) (it) t n 1 + o n ( )} = n log {1 + (it) t n + o n { [ ( )] ( )} (it) t t n log 1 + n + o log (1) + o n n ( ) = 0 + t t n o n t n The third line comes from Taylor-expanding log { } around 1. Hence ( ) t ϕ Zn(t)(t) n exp = ϕ N(0,1) (t). Since ϕ N(0,1) (t) is continuous at t = 0, we have d N(0, 1). Asymptotic approximations for X n CLT tells us that n( X n µ)/σ h a limiting standard normal distribution. We can use this true result to say something about the distribution of Xn, even when n is finite. That is, we use the CLT to derive an ymptotic approximation for the finite-sample distribution of Xn. The approximation is follows: starting with the result of the CLT: n( Xn µ)/σ d N(0, 1) we flip over the result of the CLT: X n a σ n N(0, 1) + µ X n a N(µ, 1 n σ ). The notation a makes explicit that what is on the RHS is an approximation. Note that X d n N(µ, 1 n σ ) is definitely not true! This approximation intuitively makes sense: under the sumptions of the LLCLT, we know that E X n = µ and V ar( X n ) = σ /n. What the ymptotic approximation tells us is that the distribution of Xn is approximately normal. 3 3 Note that the approximation is exactly right if we sumed that X i N(µ, σ ), for all i. 10
11 Asymptotic approximations for functions of X n Oftentimes, we are not interested per se in approximating the finite-sample distribution of the sample mean X n, but rather functions of a sample mean. (Later, you will see that the ymptotic approximations for many statistics and estimators that you run across are derived by expressing them sample averages.) Continuous mapping theorem: Let g( ) be a continuous function. Then d Z = g( ) d g(z). Proof: Serfling, Approximation Theorems of Mathematical Statistics, p. 5. (Note: you still have to figure out what the limiting distribution of g(z) is. But if you know F X, then you can get F g(x) by the change of variables formul.) Note that for any linear function g( X n ) = a X n +b, deriving the limiting distribution of g( X n ) is no problem (just use Slutsky s Theorem to get a X n + b a N(aµ + b, 1 n a σ )). The problem in deriving the distribution of g( X n ) arises when g( ) is nonlinear so that the X n is inside the g function: 1. Use the Mean Value Theorem: Let g be a continuous function on [a, b] that is differentiable on (a, b). Then, there exists (at let one) λ (a, b) such that g (λ) = g(b) g(a) b a g(b) g(a) = g (λ)(b a). Using the MVT, we can write where X n is an RV strictly between X n and µ.. On the RHS of Eq. (3): g( X n ) g(µ) = g (X n)( X n µ) (3) (a) g (X n) p g (µ) by the squeezing result, and the plim operator theorem. (b) If we multiply by n and divide by σ, we can apply the CLT to get n( Xn µ)/σ d N(0, 1). (c) Hence, n RHS = n[g( Xn ) g(µ)] using Slutsky s theorem. 11 d N(0, g (µ) σ ), (4)
12 3. Now, in order to get the ymptotic approximation for the distribution of g( X n ), we flip over to get g( X n ) a N(g(µ), 1 n g (µ) σ ). 4. Check that g satisfies sumptions for these results to go through: continuity and differentiability for MVT, and continuous at µ for plim operator theorem. Examples: 1/ X n, exp( X n ), ( X n ), etc. Eq. (4) is a general result, known the Delta method. For the purposes of this cls, I want you to derive the approximate distributions for g( X n ) from first principles ( we did above). 1
13 1 Some extra topics (CAN SKIP) 1.1 Triangular array CLT LLCLT sumes independence (across n) well identically distributed. We extend this to the independent, non-identically distributed setting. Lindeberg-Feller CLT: For each n, let Y n,1,..., Y n,n be independent random variables with finite (possibly non-identical) means and variances σ n,i such that (1/C n ) n E Y n,i EY n,i 1 { Yn,i EY n,i /C n>ϵ} 0 every ϵ > 0 i=1 where C n [ n i=1 σ n,i] 1/. Then [ n i=1 (Y n,i EY n,i )] C n d N(0, 1) The sampling framework is known a triangular array. Note that the (n 1)-th observation in the n-th sample (Y n,n 1 ) need not coincide with Y n 1,n 1, the (n 1)-th observation in the (n 1)-th sample. is a stan- C n is just the standard deviation of n i=1 Y n,i. Hence [ n i=1 (Y n,i EY n,i )] C n dardized sum. This is useful for showing ymptotic normality of Let-Squares regression. Let y = β 0 X + ϵ with ϵ i.i.d. with mean zero and variance σ, and x being an n 1 vector of covariates (here we have just 1 RHS variable). The OLS estimator is ˆβ = (X X) 1 X Y. To apply the LFCLT, we consider the normalized difference between the estimated and true β s: ((X X) 1/ /σ) ( ˆβ β 0 ) = (1/σ) (X X) 1/ X ϵ (1/σ) n a n,i ϵ i i=1 where a n,i corresponds to the i-th component of the 1 n vector (1/σ) (X X) 1/ X. So we just need to show that the Lindeberg condition applies to Y n,i = a n,i ϵ i. Note that n i=1 V (a n,iϵ i ) = V ( n i=1 a n,iϵ i ) = (1/σ ) V ((X X) 1/ X ϵ) = σ /σ 1 = Cn. 1. Convergence in distribution vs. a.s. Combining two results above, we have Z = d Z. 13
14 The converse is not generally true. (Indeed, {Z 1, Z, Z 3,...} need not be defined on the same probability space, in which ce s.d. convergence makes no sense.) However, consider Z n = F 1 (U), Z = F 1 Z (U); U U[0, 1]. Here F 1 (U) denotes the quantile function corresponding to the CDF F (z): Z F 1 (τ) = inf {z : F (z) > τ} τ [0, 1]. We have F (F 1 (τ)) = τ. (Note that quantile function is also right-continuous; discontinuity points of the quantile function arise where the CDF function is flat.) Then P (Z n z) = P (F 1 (U) z) = P (U F Zn (z)) = F Zn (z) so that Zn d =. (Even though their domains are different!) Similarly, Z = d Z. The notation = d means identically distributed. Moreover, it turns out (quite intuitive) that the convergence F 1 (U) F 1 Z (U) fails only at points where F 1 Z (U) is discontinuous (corresponding to flat portions of F Z (z)). Since these points of discontinuity are a countable set, their probability (under U[0, 1]) is equal to zero, so that F 1 (U) F 1 Z (U) for almost-all U. So what we have here, is a result that for real-valued random variables Z d n Z, we can construct identically distributed variables such that both Zn d Z and Zn Z. This is called the Skorokhod construction. Skorokhod representation: Let (n = 1,,...) be random elements defined on probability spaces (Ω n, B(Ω n ), P n ) and d Z, where X is defined on (Ω, B(Ω), P ). Then there exist random variables Z n (n = 1,,...) and Z defined on a common probability space ( Ω, B( Ω), P ) such that Z n d = (n = 1,,...); Z = d Z; Zn Z. Applications: Continuous mapping theorem: Z d n Z Zn Z ; for h( ) continuous, we have h(zn) h(z ) h(zn) d h(z ) which implies h( ) d h(z). 14
15 Building on the above, if h is bounded, then we get Eh( ) = Eh(Z n) Eh(Z ) = Eh(Z) under the bounded convergence theorem, which shows one direction of the Portmanteau theorem, Eq. (). 1.3 Functional Central limit theorems There are a set of distributional convergence results, which are known functional CLT s (or Donsker theorems), because they deal with convergence of random functions (or, interchangeably, processes). These are indispensible tools in finance. One of the simplest random functions is the Wiener process W(t), which is viewed a random function on the unit interval t [0, 1]. This is also known a Brownian motion process. Features of the Wiener process: 1. W(0) = 0. Gaussian marginals: W(t) = d N(0, t); that is, P (W(t) a) = 1 a ( ) u exp du. πt t 3. Independent increments: Define x t W(t). For any set 0 t 0 t , the differences x t1 x t0, x t x t1,... are independent. 4. Given the two above features, we have that the increments are themselves normally distributed: x ti x ti 1 d = N(0, t i t i 1 ). Moreover, from t 1 t 0 = V (x 1 x 0 ) = E[(x 1 x 0 ) ] = Ex 1 Ex 1 x 0 + Ex 0 = t 1 + t 0 Ex 1 x 0 implying Ex 1 x 0 = Cov(x 1, x 0 ) = t 0. 15
16 5. Furthermore, we know that any finite collection (x t1, x t,...) is jointly distributed multivariate normal, with mean 0 and variance matrix Σ described above. Take the conditions of the LLCLT: X 1, X,... are iid with mean zero and finite variance σ. For a given n, define the partial sum S k = X 1 + X + + X k for k n. Now, for t [0, 1], define the normalized partial sum process S n (t) = S [tn] σ n + (tn [tn]) 1 σ n X [tn]+1 where [a] denotes the largest integer a. Sn (t) is a random function on [0, 1]. (Graph) Then the functional CLT is the following result: Functional CLT: S n d W. Note that S n is a random element taking values in C[0, 1], the space of continuous functions on [0,1]. Proving this result is beyond our focus in this cls. 4 Note that the functional CLT implies more than the LLCLT. For example, take t = 1. Then S n (1) = By the FCLT, we have n i=1 X i σ n = n X n σ which is the same result the LLCLT. Similarly, for k n < n but kn n ( ) k S n = n P ( S n (1) a) P (W(1) a) = P (N(0, 1) a) τ [0, 1], we have k i=1 X i σ n = k X k σ n d N (0, τ). Also, it turns out, by a functional version of the continuous mapping theorem, we can get convergence results for functionals of the partial-sum process. For instance, sup t Sn (t) d sup t W(t), and it turns out P (sup t W(t) a) = a exp( u /) du; a 0. π (Recall that W 0 = 0, so a < 0 makes no sense.) 4 See, for instance, Billingsley, Convergence of Probability Meures, ch., section
Lecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationElements of Probability Theory
Elements of Probability Theory CHUNG-MING KUAN Department of Finance National Taiwan University December 5, 2009 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, 2009 1 / 58
More informationCHAPTER 3: LARGE SAMPLE THEORY
CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More informationLecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1
Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More informationProduct measure and Fubini s theorem
Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationProbability and Measure
Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real
More informationProbability and Measure
Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability
More informationSTAT 7032 Probability. Wlodek Bryc
STAT 7032 Probability Wlodek Bryc Revised for Spring 2019 Printed: January 14, 2019 File: Grad-Prob-2019.TEX Department of Mathematical Sciences, University of Cincinnati, Cincinnati, OH 45221 E-mail address:
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationEconomics 620, Lecture 8: Asymptotics I
Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators
More informationThe main results about probability measures are the following two facts:
Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More information4 Expectation & the Lebesgue Theorems
STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationNotes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770
Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November
More informationFundamental Tools - Probability Theory IV
Fundamental Tools - Probability Theory IV MSc Financial Mathematics The University of Warwick October 1, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory IV 1 / 14 Model-independent
More informationLecture 1: Review of Basic Asymptotic Theory
Lecture 1: Instructor: Department of Economics Stanfor University Prepare by Wenbo Zhou, Renmin University Basic Probability Theory Takeshi Amemiya, Avance Econometrics, 1985, Harvar University Press.
More informationP n. This is called the law of large numbers but it comes in two forms: Strong and Weak.
Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationStochastic Convergence, Delta Method & Moment Estimators
Stochastic Convergence, Delta Method & Moment Estimators Seminar on Asymptotic Statistics Daniel Hoffmann University of Kaiserslautern Department of Mathematics February 13, 2015 Daniel Hoffmann (TU KL)
More informationConvergence in Distribution
Convergence in Distribution Undergraduate version of central limit theorem: if X 1,..., X n are iid from a population with mean µ and standard deviation σ then n 1/2 ( X µ)/σ has approximately a normal
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationPart II Probability and Measure
Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationGAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM
GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationOn the convergence of sequences of random variables: A primer
BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :
More information9 Brownian Motion: Construction
9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of
More informationStochastic Models (Lecture #4)
Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence
More informationNotes 1 : Measure-theoretic foundations I
Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationTopic 7: Convergence of Random Variables
Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information
More informationEconomics 241B Review of Limit Theorems for Sequences of Random Variables
Economics 241B Review of Limit Theorems for Sequences of Random Variables Convergence in Distribution The previous de nitions of convergence focus on the outcome sequences of a random variable. Convergence
More informationIf g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get
18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationLECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT
MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More information4 Sums of Independent Random Variables
4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables
More informationMath Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT
Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set
More information3 Integration and Expectation
3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationProbability Theory I: Syllabus and Exercise
Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will
More informationFourth Week: Lectures 10-12
Fourth Week: Lectures 10-12 Lecture 10 The fact that a power series p of positive radius of convergence defines a function inside its disc of convergence via substitution is something that we cannot ignore
More informationThe strictly 1/2-stable example
The strictly 1/2-stable example 1 Direct approach: building a Lévy pure jump process on R Bert Fristedt provided key mathematical facts for this example. A pure jump Lévy process X is a Lévy process such
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationCONVERGENCE OF RANDOM SERIES AND MARTINGALES
CONVERGENCE OF RANDOM SERIES AND MARTINGALES WESLEY LEE Abstract. This paper is an introduction to probability from a measuretheoretic standpoint. After covering probability spaces, it delves into the
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and
More informationAnalysis Finite and Infinite Sets The Real Numbers The Cantor Set
Analysis Finite and Infinite Sets Definition. An initial segment is {n N n n 0 }. Definition. A finite set can be put into one-to-one correspondence with an initial segment. The empty set is also considered
More informationBootstrap - theoretical problems
Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem
More informationEconometrics I. September, Part I. Department of Economics Stanford University
Econometrics I Deartment of Economics Stanfor University Setember, 2008 Part I Samling an Data Poulation an Samle. ineenent an ientical samling. (i.i..) Samling with relacement. aroximates samling without
More information7 Convergence in R d and in Metric Spaces
STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a
More informationSTAT 331. Martingale Central Limit Theorem and Related Results
STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal
More information1 Stat 605. Homework I. Due Feb. 1, 2011
The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationSpectral representations and ergodic theorems for stationary stochastic processes
AMS 263 Stochastic Processes (Fall 2005) Instructor: Athanasios Kottas Spectral representations and ergodic theorems for stationary stochastic processes Stationary stochastic processes Theory and methods
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationGaussian Random Fields
Gaussian Random Fields Mini-Course by Prof. Voijkan Jaksic Vincent Larochelle, Alexandre Tomberg May 9, 009 Review Defnition.. Let, F, P ) be a probability space. Random variables {X,..., X n } are called
More informationLecture 6 Basic Probability
Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic
More informationLecture 12. F o s, (1.1) F t := s>t
Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let
More informationPart 2 Continuous functions and their properties
Part 2 Continuous functions and their properties 2.1 Definition Definition A function f is continuous at a R if, and only if, that is lim f (x) = f (a), x a ε > 0, δ > 0, x, x a < δ f (x) f (a) < ε. Notice
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationThe Canonical Gaussian Measure on R
The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where
More informationLimit Thoerems and Finitiely Additive Probability
Limit Thoerems and Finitiely Additive Probability Director Chennai Mathematical Institute rlk@cmi.ac.in rkarandikar@gmail.com Limit Thoerems and Finitiely Additive Probability - 1 Most of us who study
More informationContents Ordered Fields... 2 Ordered sets and fields... 2 Construction of the Reals 1: Dedekind Cuts... 2 Metric Spaces... 3
Analysis Math Notes Study Guide Real Analysis Contents Ordered Fields 2 Ordered sets and fields 2 Construction of the Reals 1: Dedekind Cuts 2 Metric Spaces 3 Metric Spaces 3 Definitions 4 Separability
More informationLecture 2: Random Variables and Expectation
Econ 514: Probability and Statistics Lecture 2: Random Variables and Expectation Definition of function: Given sets X and Y, a function f with domain X and image Y is a rule that assigns to every x X one
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationProbability Background
CS76 Spring 0 Advanced Machine Learning robability Background Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu robability Meure A sample space Ω is the set of all possible outcomes. Elements ω Ω are called sample
More information1 Sequences of events and their limits
O.H. Probability II (MATH 2647 M15 1 Sequences of events and their limits 1.1 Monotone sequences of events Sequences of events arise naturally when a probabilistic experiment is repeated many times. For
More informationBrownian Motion and Conditional Probability
Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical
More informationProbability and Measure
Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More informationA Primer on Asymptotics
A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these
More informationLecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.
Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal
More informationUses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationLecture 17 Brownian motion as a Markov process
Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is
More informationLecture Notes on Asymptotic Statistics. Changliang Zou
Lecture Notes on Asymptotic Statistics Changliang Zou Prologue Why asymptotic statistics? The use of asymptotic approximation is two-fold. First, they enable us to find approximate tests and confidence
More informationMod-φ convergence I: examples and probabilistic estimates
Mod-φ convergence I: examples and probabilistic estimates Valentin Féray (joint work with Pierre-Loïc Méliot and Ashkan Nikeghbali) Institut für Mathematik, Universität Zürich Summer school in Villa Volpi,
More information17. Convergence of Random Variables
7. Convergence of Random Variables In elementary mathematics courses (such as Calculus) one speaks of the convergence of functions: f n : R R, then lim f n = f if lim f n (x) = f(x) for all x in R. This
More informationA PECULIAR COIN-TOSSING MODEL
A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or
More informationSUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES
SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,
More informationIntroduction and Preliminaries
Chapter 1 Introduction and Preliminaries This chapter serves two purposes. The first purpose is to prepare the readers for the more systematic development in later chapters of methods of real analysis
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationNotes for Expansions/Series and Differential Equations
Notes for Expansions/Series and Differential Equations In the last discussion, we considered perturbation methods for constructing solutions/roots of algebraic equations. Three types of problems were illustrated
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationLecture Notes for MA 623 Stochastic Processes. Ionut Florescu. Stevens Institute of Technology address:
Lecture Notes for MA 623 Stochastic Processes Ionut Florescu Stevens Institute of Technology E-mail address: ifloresc@stevens.edu 2000 Mathematics Subject Classification. 60Gxx Stochastic Processes Abstract.
More informationChapter 7: Special Distributions
This chater first resents some imortant distributions, and then develos the largesamle distribution theory which is crucial in estimation and statistical inference Discrete distributions The Bernoulli
More information