Point Estimation: properties of estimators 1 FINITE-SAMPLE PROPERTIES. finite-sample properties (CB 7.3) large-sample properties (CB 10.
|
|
- Joy Stephens
- 6 years ago
- Views:
Transcription
1 Poit Estimatio: properties of estimators fiite-sample properties CB 7.3) large-sample properties CB 10.1) 1 FINITE-SAMPLE PROPERTIES How a estimator performs for fiite umber of observatios. Estimator: W Parameter: θ Criteria for evaluatig estimators: Bias: does EW = θ? Variace of W you would like a estimator with a smaller variace) Example: X 1,..., X i.i.d. µ, σ 2 ) Ukow parameters are µ ad σ 2. Cosider: ˆµ 1 i X i, estimator of µ ˆσ 2 1 i X i X ) 2, estimator of σ 2. Bias: E ˆµ = 1 µ = µ. So ubiased. Var ˆµ = 1 2 σ 2 = 1 σ2. Eˆσ 2 = E 1 X i X ) 2 ) i = 1 EXi 2 2EX i X + E X ) 2 i = 1 [ ] µ 2 + σ 2 ) 2µ 2 + σ2 ) + σ2 + µ2 = 1 σ2. 1
2 Hece it is biased. To fix this bias, cosider the estimator s 2 1 ubiased). 1 i X i X ) 2, ad Es 2 = σ 2 Mea-squared error MSE) of W is EW θ) 2. Commo criterio for comparig estimators. Decompose: MSEW ) = V W + EW θ) 2 =Variace + Bias) 2. Hece, for a ubiased estimator: MSEW ) = V W. Example: X 1,..., X U[0, θ]. fx) = 1/θ, x [0, θ]. Cosider estimator ˆθ 2 X. E ˆθ = 2 1 E i X i = MSEˆθ N ) = V ˆθ = 4 2 i V X i = θ2 3 θ = θ. So ubiased Cosider estimator θ max X 1,..., X ). I order to derive momets, start by derivig CDF: P θ z) = P X 1 z, X 2 z,..., X z) = P X i z) i=1 { z ) if z θ = θ 1 otherwise Therefore z) = z 1 1 f θ θ), for 0 x θ. θ Bias θ ) = θ/ + 1) E θ ) 2 = θ θ 0 z+1 dz = Hece V θ = θ 2 +2 θ z ) 1 1 E θ ) = z 0 θ θ dz = θ z dz = θ + 1 θ. +2 θ2. +1 Accordigly, MSE= 2θ 2 +2)+1) ) 2 ) 0 = θ 2 +2)+1) 2. 2
3 Cotiue the previous example. Redefie θ = +1 maxx 1,..., X ). Now both estimators ˆθ ad θ are ubiased. Which is better? V ˆθ = θ2 = O1/). 3 V θ = ) ) +1 2 V maxx1,..., X )) = θ 2 1 = O1/ 2 ). +2) Hece, for large eough, θ has a smaller variace, ad i this sese it is better. Best ubiased estimator: if you choose the best i terms of MSE) estimator, ad restrict yourself to ubiased estimators, the the best estimator is the oe with the lowest variace. A best ubiased estimator is also called the Uiform miimum variace ubiased estimator UMVUE). Formally: a estimator W is a UMVUE of θ satisfies: i) E θ W = θ, for all θ ubiasedess) ii) V θ W V θ W, for all θ, ad all other ubiased estimators W. The uiform coditio is crucial, because it is always possible to fid estimators which have zero variace for a specific value of θ. It is difficult i geeral to verify that a estimator W is UMVUE, sice you have to verify coditio ii) of the defiitio, that V W is smaller tha all other ubiased estimators. Luckily, we have a importat result for the lowest attaiable variace of a estimator. Theorem Cramer-Rao Iequality): Let X 1,..., X be a sample with joit pdf f X θ), ad let W X) be ay estimator satisfyig d i) dθ E θw X) [ = W X) θ f X θ) ] dx; The V θ W X) ii) V θ W X) <. E θ d E dθ θw X) ) 2 log f X θ) θ ) 2. 3
4 The RHS above is called the Cramer-Rao Lower Boud. Proof: CB, pg I short, we apply the Cauchy-Schwarz iequality V S) covs, T )/V T ) with S = W X) ad T = θ log f X θ). The choice of T here may seem a bit arbitrary. To get some ituitio, cosider Cramer s derivatio. 1 Start with the followig maipulatio of the equality E θ W X) = W X)fX θ)dx: d dθ E θw X) = d W X)fX θ)dx dθ = W X) θ fx θ)dx = W X) E θ W X)) θ fx θ)dx ote E θ W X) fx θ)dx = 0) θ ) = W X) E θ W X)) log fx θ) fx θ)dx θ Applyig the Cauchy-Schwarz iequality, we have [ d dθ E θw X)] 2 V ar θ W X) E θ [ log fx θ)]2 θ or V ar θ W X) [ d dθ E θw X)] 2 E θ [ θ log fx θ)]2. The LHS of coditio i) above is d dθ W X)fX θ)dx, so by Leibiz rule, this coditio rules out cases where the support of X is depedet o θ. The crucial step i the derivatio of the CR-boud is the iterchage of differetiatio ad itegratio which implies E θ θ log f X θ) = 1 f X θ) = f X θ)dx θ = θ 1 = 0 1 Cramer, Mathematical Methods of Statistics, p. 475ff. f X θ) f X θ)dx θ 1) 4
5 skip) The above derivatio is oteworthy, because θ log f X θ) = 0 is the FOC of maximum likelihood estimatio problem. Alteratively, as i CB, apply coditio i) of CR result, usig W X) = 1.) I the i.i.d. case, this becomes the sample average 1 θ log fx i θ) = 0. i Ad by the LLN: 1 i θ log fx i θ) p E θ0 θ log fx i θ), where θ 0 is the true value of θ 0. This shows that maximum likelihood estimatio of θ is equivalet to estimatio based o the momet coditio E θ0 θ log fx i θ) = 0 which holds oly at the true value θ = θ 0. Thus MLE is cosistet for the true value θ 0, as we ll see later.) However, ote that Eq. 1) holds at all values of θ, ot just θ 0.) [Thik about] What if model is misspecified, i the sese that true desity of X is g x), ad that for all θ Θ, f x θ) g x) that is, there is o value of the parameter θ such that the postulated model f coicides with the true model g)? Does Eq. 1) still hold? What is MLE lookig for? I the iid case, the CR lower boud ca be simplified Corollary : if X 1,..., X i.i.d. fx θ), the V θ W X) d E dθ θw X) ) 2 E θ θ log fx θ)) 2. Up to this poit, Cramer-Rao results ot that operatioal for us to fid a best estimator, because the estimator W X) is o both sides of the iequality. However, for a ubiased estimator, how ca you simplify the expressio further? 5
6 Example: X 1,..., X i.i.d. Nµ, σ 2 ). What is CRLB for a ubiased estimator of µ? Ubiased umerator =1. log fx θ) = log 2π log σ 1 ) 2 x µ 2 σ ) ) x µ 1 log fx θ) = = x µ µ σ σ σ 2 ) 2 E log fx θ) = E X µ) 2 σ 4) = 1 θ σ V X = 1 4 σ. 2 Hece the CRLB = 1 1 σ 2 = σ2. This is the variace of the sample mea, so that the sample mea is a UMVUE for µ. Sometimes we ca simplify the deomiator of the CRLB further: Lemma Iformatio iequality): if fx θ) satisfies ) [ ) ] d *) dθ E θ log fx θ) = log fx θ) fx θ) dx, θ θ θ the ) 2 ) 2 E θ θ log fx θ) = E θ log fx θ). θ2 Rough proof: LHS of *): Usig Eq. 1) above, we get that LHS of *) =0. RHS of *): [ ) ] θ θ log f f dx 2 ) 2 log f 1 f = fdx + dx θ 2 f θ ) 2 =E 2 log f log f + E. θ 2 θ Puttig the LHS ad RHS together yields the desired result. 6
7 The LHS of the above coditio *) is just d log fx θ)) fx θ)dx. As dθ θ before, the crucial step is the iterchage of differetiatio ad itegratio. skip) Also, the iformatio iequality depeds crucially o the equality E θ log fx θ) = θ 0, which depeds o the correct specificatio of the model. Thus iformatio iequality ca be used as basis of specificatio test. How?) Example: for the previous example, cosider CRLB for ubiased estimator of σ 2. We ca use the iformatio iequality, because coditio *) is satisfied for the ormal distributio. Hece: E 1 log fx θ) = σ2 2σ + 1 x µ) 2 ) 2 2 σ 4 log fx θ) = 1 x µ)2 σ 2 σ2 2σ4 σ )) 6 1 log fx θ) = σ2 2σ 1 ) 4 σ 4 σ 2 Hece the CRLB is 2σ4. = 1 2σ 4. Example: X 1,..., X U[0, θ]. Check coditios for CRLB for a ubiased estimator W X) of θ. d EW X) = 1 because it is ubiased) dθ [ W X)f θ X θ) ] dx = ) W X) 1 dx d dθ EW X) = 1 θ 2 Hece, coditio i) of theorem ot satisfied. skip) But whe ca CRLB if it exists) be attaied? Corollary : X 1,..., X i.i.d. fx θ), satisfyig the coditios of CR theorem. Let Lθ X) = i=1 fx θ) deote the likelihood fuctio. Estimator W X) ubiased for θ W X) attais CRLB iff you ca write θ log Lθ X) [ = aθ) W X) ] θ for some fuctio aθ). 7
8 Example: X 1,..., X i.i.d. Nµ, σ 2 ) Cosider estimatig µ: µ log Lθ X) = µ ) log fx θ) i=1 = log 2π log σ 1 ) ) X µ) 2 µ 2 σ 2 i=1 ) X µ = i=1 σ 2 = σ 2 X µ). Hece, CRLB ca be attaied i fact, we showed earlier that CRLB attaied by X ) Loss fuctio optimality Let X f X θ). Cosider a loss fuctio Lθ, W X)), takig values i [0, + ), which pealizes you whe your W X) estimator is far from the true parameter θ. Note that Lθ, W X)) is a radom variable, sice X ad W X)) are radom. Cosider estimators which miimize expected loss: that is mi E θlθ, W X)) mi Rθ, W )) W ) W ) where Rθ, W )) is the risk fuctio. Note: the risk fuctio is ot a radom variable, because X has bee itegrated out.) Loss fuctio optimality is a more geeral criterio tha miimum MSE. I fact, 2, because MSEW X)) = E θ W X) θ) the MSE is actually the risk fuctio associated with the quadratic loss fuctio Lθ, W X)) 2. = W X) θ) Other examples of loss fuctios: Absolute error loss: W X) θ 8
9 Relative quadratic error loss: W X) θ) 2 θ +1 The exercise of miimizig risk takes a give value of θ as give, so that the miimized risk of a estimator depeds o whichever value of θ you are cosiderig. You might be iterested i a estimator which does well regardless of which value of θ you are cosiderig. Aalogous to the focus o the uiform miimal variace.) For this differet problem, you wat to cosider a otio of risk which does ot deped o θ. Two criteria which have bee cosidered are: Average risk: mi W ) Rθ, W ))hθ)dθ. where hθ) is some weightig fuctio across θ. I a Bayesia iterpretatio, hθ) is a prior desity over θ.) Mimax criterio: mi max Rθ, W )). W ) θ Here you choose the estimator W ) to miimize the maximum risk = max θ Rθ, W )), where θ is set to the worse value. So mimax optimizer is the best that ca be achieved i a worst-case sceario. Example: X 1,..., X i.i.d. Nµ, σ 2 ). Sample mea X is: ubiased miimum MSE UMVUE attais CRLB miimizes expected quadratic loss 9
10 2 LARGE SAMPLE PROPERTIES OF ESTIMATORS It ca be difficult to compute MSE, risk fuctios, etc., for some estimators, especially whe estimator does ot resemble a sample average. Large-sample properties: exploit LLN, CLT Cosider { data {X 1, X 2,...} by which we costruct a sequece of estimators W W X 1 ), W X } 2 ),.... W is a radom sequece. Defie: we say that W is cosistet for a parameter θ iff the radom sequece W as coverges i some stochastic sese) to θ. Strog cosistecy obtais whe W θ. p Weak cosistecy obtais whe W θ. For estimators like sample-meas, cosistecy either weak or strog) follows easily usig a LLN. Cosistecy ca be thought of as the large-sample versio of ubiasedess. Defie: a M-estimator is a estimator of θ which a maximizer of a objective fuctio Q θ). Examples: MLE: Q θ) = 1 i=1 log fx i θ) Least squares: Q θ) = i=1 [y i gx i ; θ)] 2. OLS is special case whe gx i ; θ) = α + X iβ. GMM: Q θ) = G θ) W θ)g θ) where G θ) = [ 1 m 1 x i ; θ), 1 i=1 m 2 x i ; θ),..., 1 i=1 m M x i ; θ)], i=1 a M 1 vector of sample momet coditios, ad W is a M M weightig matrix. Notatio: For each θ Θ, let f θ fx 1,..., x,... ; θ) deote the joit desity of the data for the give value of θ. For θ 0 Θ, we deote the limit objective fuctio Q 0 θ) = plim,fθ0 Q θ) at each θ). 10
11 Cosistecy of M-estimators Make the followig assumptios: 1. For each θ 0 Θ, the limitig objective fuctio Q 0 θ) is uiquely maximized at θ 0 idetificatio ) 2. Parameter space Θ is a compact subset of R K. 3. Q 0 θ) is cotiuous i θ 4. Q θ) coverges uiformly i probability to Q 0 θ); that is: p sup Q θ) Q 0 θ) 0. θ Θ Theorem: Cosistecy of M-Estimator) Uder assumptio 1,2,3,4, θ p θ0. Proof: We eed to show: for ay arbitrarily small eighborhood N cotaiig θ 0, P θ N ) 1. For large eough, the uiform covergece coditios that, for all ɛ, δ > 0, ) P sup Q θ) Q 0 θ) < ɛ/2 θ Θ > 1 δ. The evet sup θ Θ Q θ) Q 0 θ) < ɛ/2 implies Similarly, Q θ ) Q 0 θ ) < ɛ/2 Q 0 θ ) > Q θ ) ɛ/2 2) Q θ 0 ) Q 0 θ 0 ) > ɛ/2 Q θ 0 ) > Q 0 θ 0 ) ɛ/2. 3) Sice θ = argmax θ Q θ), Eq. 2) implies Hece, addig Eqs. 3) ad 4), we have So we have show that Q 0 θ ) > Q θ 0 ) ɛ/2. 4) Q 0 θ ) > Q 0 θ 0 ) ɛ. 5) sup Q θ) Q 0 θ) < ɛ/2 = Q 0 θ ) > Q 0 θ 0 ) ɛ θ Θ ) P Q 0 θ ) > Q 0 θ 0 ) ɛ) P sup Q θ) Q 0 θ) < ɛ/2 θ Θ 1. 11
12 Now defie N as ay ope eighborhood of R K, which cotais θ 0, ad N is the complemet of N i R K. The Θ N is compact, so that max θ Θ N Q 0 θ) exists. Set ɛ = Q 0 θ 0 ) max θ Θ N Q 0 θ). The Q 0 θ ) > Q 0 θ 0 ) ɛ Q 0 θ ) > max Q 0 θ) θ Θ N θ N P θ N ) P Q 0 θ ) > Q 0 θ 0 ) ɛ) 1. Sice the argumet above holds for ay arbitrarily small eighborhood of θ 0, we are doe. I geeral, the limit objective fuctio Q 0 θ) = plim Q θ) may ot be that straightforward to determie. But i may cases, Q θ) is a sample average of some sort: Q θ) = 1 qx i θ) eg. least squares, MLE). The by a law of large umbers, we coclude that for all θ) Q 0 θ) = plim 1 qx i θ) = E xi qx i θ) i where E xi deote expectatio with respect to the true but uobserved) distributio of x i. skip) Most of the time, θ 0 ca be iterpreted as a true value. But if model is misspecified, the this iterpretatio does t hold ideed, uder misspecificatio, ot eve clear what the true value is). So a more cautious way to iterpret the cosistecy result is that i θ p argmaxθ Q 0 θ) which holds give the coditios) o matter whether model is correctly specified. skip) Cosider the abstract) mappig from parameter values θ to joit desities of the data f θ = fx 1,..., x,... ; θ). The idetificatio coditio posits: for each θ 1, the limit objective fuctio Q 1 plim fθ1 Q θ) i uiquely optimized at θ 1. This implies that the mappig θ f θ is oe-to-oe. Assume ot: that is, there are two values θ 1 θ 2 such that f θ1 = f θ2 a.s.). 2 The θ 1 = argmi θ Q 1 θ) = argmi θ Q 2 θ) = θ 2. 2 I other words, θ 1 ad θ 2 are observatioally equivalet. 12
13 ** Let s upack the uiform covergece coditio. Sufficiet coditios for this coditios are: 1. Poitwise covergece: For each θ Θ, Q θ) Q 0 θ) = o p 1). 2. Q θ) is stochastically equicotiuous: for every ɛ > 0, η > 0 there exists a sequece of radom variable ɛ, η) ad ɛ, η) such that for all >, P > ɛ) < η ad for each θ there is a ope set N cotaiig θ with sup θ N Q θ) Q θ), >. Note that both ad do ot deped o θ: it is uiform result. This is a i probability versio of the determiistic otio of uiform equicotiuity: we say a sequece of determiistic fuctios R θ) is uiformly equicotiuous if, for every ɛ > 0 there exists δɛ) ad ɛ) such that for all θ sup θ: θ θ <δ R θ) R θ) ɛ, >. To uderstad this more ituitively, cosider a aive argumet for cosistecy. By cotiuity of Q 0, we kow that Q 0 θ) is close to Q 0 θ 0 ) for θ N θ 0 ). By poitwise covergece, we have Q θ) covergig to Q 0 θ) for all θ; hece, eve if Q θ) is ot optimized by θ 0, the optimizer θ = argmax θ Q θ) should ot be far from θ 0. Implicitly, for the last part, we have assumed that Q θ) is equally close to Q 0 θ) for all θ, because the the optimizers of Q ad Q 0 caot be too far apart. However, poitwise covergece is ot eough to esure this equally closeess. At ay give, Q θ 0 ) beig close to Q 0 θ 0 ) does ot imply this at other poits. Uiform covergece esures, roughly speakig, that at ay give, Q ad Q 0 are equally close at all poits θ. As the above discussio shows, uiform covergece is related to a cotiuity coditio o Q ); if Q ) is ot cotiuous, the it ca be maximized far from θ 0. Stochastic equicotiuity is the right otio of cotiuity here sice Q ) ad θ are both radom. Asymptotic ormality for M-estimators Defie the score vector [ Q θ) θq θ) =,..., Q θ) θ 1 θ K. θ= θ] θ= θ 13
14 Similarly, defie the K K Hessia matrix [ θ θ Q θ) ] i,j = 2 Q θ) θ i θ j Note that the Hessia is symmetric. Make the followig assumptios: 1 i, j K. θ= θ, 1. θ = argmax θ Q θ) p θ 0 2. θ 0 iteriorθ) 3. Q θ) is twice cotiuously differetiable i a eighborhood N of θ θ0 Q θ) d N0, Σ) 5. Uiform covergece of Hessia: there exists the matrix Hθ) which is cotiuous at θ 0 ad sup θ N θθ Q θ) Hθ) 0. p 6. Hθ 0 ) is osigular Theorem Asymptotic ormality for M-estimator): Uder assumptios 1,2,3,4,5, where H 0 Hθ 0 ). θ θ 0 ) d N0, H 1 0 ΣH 1 0 ) Proof: sketch) By Assumptios 1,2,3, θ Q θ) = 0 this is FOC of maximizatio problem). The usig mea-value theorem with θ deotig mea value): 0 = θ Q θ) = θ0 Q θ) + θ θ Q θ)θ θ 0 ) θ θ 0 ) = θ0 Q θ) }{{} θ θ Q θ) }{{} p H 0 usig A5) d N0,Σ) usig A4) θ θ 0 ) d Hθ 0 ) 1 N0, Σ) = N0, H 1 0 ΣH 1 0 ). Note: A5 is a uiform covergece assumptio o the sample Hessia. Give previous discussio, it esures that the sample Hessia θθ Q θ) evaluated at θ which is close to θ 0 ) does ot vary far from the limit Hessia Hθ) at θ 0, which is implied by a type of cotiuity of the sample Hessia close to θ 0. 14
15 2.1 Maximum likelihood estimatio The cosistecy of MLE ca follow by applicatio of the theorem above for cosistecy of M-estimators. Essetially, as we oted above,what the cosistecy theorem showed above was that, for ay M-estimator sequece θ : plim θ = argmax θ Q 0 θ). For MLE, there is a distict ad earlier argumet due to Wald 1949), who shows that, i the i.i.d. case, the limitig likelihood fuctio correspodig to Q 0 θ)) is ideed globally maximized at θ 0, the true value. Thus, we ca directly cofirm the idetificatio assumptio of the M-estimator cosistecy theorem. This argumet is of iterest by itself. Argumet: summary of Amemiya, pp ) MLE 1 Defie ˆθ argmax θ i log fx i θ). Let θ 0 deote the true value. By LLN: 1 i log fx i θ) p E θ0 log fx i θ), for all θ ot ecessarily the true θ 0 ). ) ) By Jese s iequality: E θ0 log fx θ) fx θ 0 < log E fx θ) ) θ0 fx θ 0 ) ) But E fx θ) θ0 fx θ 0 = ) fx θ) ) fx θ 0 fx θ ) 0 ) = 1, sice fx θ) is a desity fuctio, for all θ. 3 Hece: ) fx θ) E θ0 log < 0, θ fx θ 0 ) = E θ0 log fx θ) < E θ0 log fx θ 0 ), θ = E θ0 log fx θ) is maximized at the true θ 0. This is the idetificatio assumptio from the M-estimator cosistecy theorem. 3 I this step, ote the importace of assumptio A3 i CB, pg If x has support depedig o θ, the it will ot itegrate to 1 for all θ. 15
16 skip) Aalogously, we also kow that, for δ > 0, µ 1 = E θ0 log fx; θ ) 0 δ) < 0; µ 2 = E θ0 log fx; θ ) 0 + δ) fx; θ 0 ) fx; θ 0 ) i < 0. By the SLLN, we kow that 1 log fx ) i; θ 0 δ) = 1 fx i ; θ 0 ) [log L x; θ 0 δ) log L x : θ 0 )] as µ 1 so that, with probability 1, log L x; θ 0 δ) < log L x; θ 0 ) for large eough. Similarly, for large eough, log L x; θ 0 + δ) < log L x; θ 0 ) with probability 1. Hece, for large, ˆθ argmax θ log L x; θ) θ 0 δ, θ 0 + δ). That is, the MLE ˆθ is strogly cosistet for θ 0. Note that this argumet requires weaker assumptios tha the M-estimator cosistecy theorem above. Now we cosider aother idea, efficiecy, which is a large-sample aalogue of the miimum variace cocept. For the sequece of estimators W, suppose that k)w θ) d N0, σ 2 ) where k) is a polyomial i. The σ 2 is deoted the asymptotic variace of W. I usual cases, k) =. For example, by the CLT, we kow that X µ) d N0, σ 2 ). Hece, σ 2 is the asymptotic variace of the sample mea X. Defiitio : A estimator sequece W is asymptotically efficiet for θ if W θ) d N0, vθ)), where 1 the asymptotic variace vθ) = the E θ0 log fx θ))2 θ CRLB) 16
17 By asymptotic ormality result for M-estimator, we kow what the asymptotic distributio for the MLE should be. However, it turs out give the iformatio iequality, the MLE s asymptotic distributio ca be further simplified. Theorem : Asymptotic efficiecy of MLE Proof: followig Amemiya, pp ) MLE ˆθ satisfies the FOC of the MLE problem: Usig the mea value theorem: 0 = log Lθ X ) θ 0 = log Lθ X ) θ θ=θ0 + 2 log Lθ X ) θ 2 = ) ˆθ θ 0 = log Lθ X ) θ θ=θ0 2 log Lθ X ) = θ=θ θ 2 θ=ˆθmle. θ=θ ˆθMLE 1 1 ) θ 0 i log fx i θ) i θ θ=θ0 2 log fx i θ) θ 2 θ=θ ) Note that, by the LLN, 1 i log fx i θ) θ θ=θ0 p log fx θ) Eθ0 θ θ=θ0 = fxi θ) θ θ=θ0 dx. Usig same argumet as i the iformatio iequality result above, the last term is: f θ dx = fdx = 0. θ Hece, the CLT ca be applied to the umerator of **): ) ) umerator of **) d log fxi θ) 2 N 0, E θ0 θ θ=θ0. By LLN, ad uiform covergece of Hessia term: deomiator of **) p E θ0 2 log fx θ) θ 2 θ=θ0. 17
18 Hece, by Slutsky theorem: ˆθ θ 0 ) d N By the iformatio iequality: so that E log fxi ) 2 θ) θ0 θ θ=θ0 0, [ ] 2. E 2 log fx θ) θ=θ0 θ0 θ 2 ) log fxi θ) 2 E θ0 2 log fx θ) θ θ=θ0 = E θ0 θ 2 ˆθ θ 0 ) d N 0, so that the asymptotic variace is the CRLB. 1 E θ0 log fxi θ) θ θ=θ0 ) 2 θ=θ0 Hece, the asymptotic approximatio for the fiite-sample distributio is ˆθ MLE a N θ 0, 1 1 E θ0 log fxi θ) θ ) 2. θ=θ0 18
Lecture 19: Convergence
Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may
More informationLet us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.
Lecture 5 Let us give oe more example of MLE. Example 3. The uiform distributio U[0, ] o the iterval [0, ] has p.d.f. { 1 f(x =, 0 x, 0, otherwise The likelihood fuctio ϕ( = f(x i = 1 I(X 1,..., X [0,
More informationConvergence of random variables. (telegram style notes) P.J.C. Spreij
Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space
More informationUnbiased Estimation. February 7-12, 2008
Ubiased Estimatio February 7-2, 2008 We begi with a sample X = (X,..., X ) of radom variables chose accordig to oe of a family of probabilities P θ where θ is elemet from the parameter space Θ. For radom
More informationMATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4
MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.
More informationEECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1
EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum
More informationIntroductory statistics
CM9S: Machie Learig for Bioiformatics Lecture - 03/3/06 Itroductory statistics Lecturer: Sriram Sakararama Scribe: Sriram Sakararama We will provide a overview of statistical iferece focussig o the key
More informationChapter 6 Infinite Series
Chapter 6 Ifiite Series I the previous chapter we cosidered itegrals which were improper i the sese that the iterval of itegratio was ubouded. I this chapter we are goig to discuss a topic which is somewhat
More informationLecture 12: September 27
36-705: Itermediate Statistics Fall 207 Lecturer: Siva Balakrisha Lecture 2: September 27 Today we will discuss sufficiecy i more detail ad the begi to discuss some geeral strategies for costructig estimators.
More information6. Sufficient, Complete, and Ancillary Statistics
Sufficiet, Complete ad Acillary Statistics http://www.math.uah.edu/stat/poit/sufficiet.xhtml 1 of 7 7/16/2009 6:13 AM Virtual Laboratories > 7. Poit Estimatio > 1 2 3 4 5 6 6. Sufficiet, Complete, ad Acillary
More informationChapter 3. Strong convergence. 3.1 Definition of almost sure convergence
Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i
More informationSequences and Series of Functions
Chapter 6 Sequeces ad Series of Fuctios 6.1. Covergece of a Sequece of Fuctios Poitwise Covergece. Defiitio 6.1. Let, for each N, fuctio f : A R be defied. If, for each x A, the sequece (f (x)) coverges
More informationLecture 8: Convergence of transformations and law of large numbers
Lecture 8: Covergece of trasformatios ad law of large umbers Trasformatio ad covergece Trasformatio is a importat tool i statistics. If X coverges to X i some sese, we ofte eed to check whether g(x ) coverges
More informationEconomics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator
Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS
MASSACHUSTTS INSTITUT OF TCHNOLOGY 6.436J/5.085J Fall 2008 Lecture 9 /7/2008 LAWS OF LARG NUMBRS II Cotets. The strog law of large umbers 2. The Cheroff boud TH STRONG LAW OF LARG NUMBRS While the weak
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationEmpirical Processes: Glivenko Cantelli Theorems
Empirical Processes: Gliveko Catelli Theorems Mouliath Baerjee Jue 6, 200 Gliveko Catelli classes of fuctios The reader is referred to Chapter.6 of Weller s Torgo otes, Chapter??? of VDVW ad Chapter 8.3
More informationProbability and Statistics
ICME Refresher Course: robability ad Statistics Staford Uiversity robability ad Statistics Luyag Che September 20, 2016 1 Basic robability Theory 11 robability Spaces A probability space is a triple (Ω,
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationLecture 10 October Minimaxity and least favorable prior sequences
STATS 300A: Theory of Statistics Fall 205 Lecture 0 October 22 Lecturer: Lester Mackey Scribe: Brya He, Rahul Makhijai Warig: These otes may cotai factual ad/or typographic errors. 0. Miimaxity ad least
More informationTopic 9: Sampling Distributions of Estimators
Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be
More informationLecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett
Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets
More informationAsymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values
of the secod half Biostatistics 6 - Statistical Iferece Lecture 6 Fial Exam & Practice Problems for the Fial Hyu Mi Kag Apil 3rd, 3 Hyu Mi Kag Biostatistics 6 - Lecture 6 Apil 3rd, 3 / 3 Rao-Blackwell
More informationRandom Variables, Sampling and Estimation
Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig
More informationProblem Set 4 Due Oct, 12
EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios
More information1.010 Uncertainty in Engineering Fall 2008
MIT OpeCourseWare http://ocw.mit.edu.00 Ucertaity i Egieerig Fall 2008 For iformatio about citig these materials or our Terms of Use, visit: http://ocw.mit.edu.terms. .00 - Brief Notes # 9 Poit ad Iterval
More informationDirection: This test is worth 250 points. You are required to complete this test within 50 minutes.
Term Test October 3, 003 Name Math 56 Studet Number Directio: This test is worth 50 poits. You are required to complete this test withi 50 miutes. I order to receive full credit, aswer each problem completely
More informationStat410 Probability and Statistics II (F16)
Some Basic Cocepts of Statistical Iferece (Sec 5.) Suppose we have a rv X that has a pdf/pmf deoted by f(x; θ) or p(x; θ), where θ is called the parameter. I previous lectures, we focus o probability problems
More informationLECTURE 2 LEAST SQUARES CROSS-VALIDATION FOR KERNEL DENSITY ESTIMATION
Jauary 3 07 LECTURE LEAST SQUARES CROSS-VALIDATION FOR ERNEL DENSITY ESTIMATION Noparametric kerel estimatio is extremely sesitive to te coice of badwidt as larger values of result i averagig over more
More information7.1 Convergence of sequences of random variables
Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite
More informationDistribution of Random Samples & Limit theorems
STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to
More informationAn Introduction to Asymptotic Theory
A Itroductio to Asymptotic Theory Pig Yu School of Ecoomics ad Fiace The Uiversity of Hog Kog Pig Yu (HKU) Asymptotic Theory 1 / 20 Five Weapos i Asymptotic Theory Five Weapos i Asymptotic Theory Pig Yu
More informationStatistical Theory MT 2008 Problems 1: Solution sketches
Statistical Theory MT 008 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. a) Let 0 < θ < ad put fx, θ) = θ)θ x ; x = 0,,,... b) c) where α
More informationStatistical Theory MT 2009 Problems 1: Solution sketches
Statistical Theory MT 009 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. (a) Let 0 < θ < ad put f(x, θ) = ( θ)θ x ; x = 0,,,... (b) (c) where
More informationLecture 11 and 12: Basic estimation theory
Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis
More informationLECTURE 14 NOTES. A sequence of α-level tests {ϕ n (x)} is consistent if
LECTURE 14 NOTES 1. Asymptotic power of tests. Defiitio 1.1. A sequece of -level tests {ϕ x)} is cosistet if β θ) := E θ [ ϕ x) ] 1 as, for ay θ Θ 1. Just like cosistecy of a sequece of estimators, Defiitio
More informationEcon 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.
Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio
More informationThis exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.
Probability ad Statistics FS 07 Secod Sessio Exam 09.0.08 Time Limit: 80 Miutes Name: Studet ID: This exam cotais 9 pages (icludig this cover page) ad 0 questios. A Formulae sheet is provided with the
More informationChapter 6 Principles of Data Reduction
Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a
More informationECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors
ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic
More informationOutput Analysis and Run-Length Control
IEOR E4703: Mote Carlo Simulatio Columbia Uiversity c 2017 by Marti Haugh Output Aalysis ad Ru-Legth Cotrol I these otes we describe how the Cetral Limit Theorem ca be used to costruct approximate (1 α%
More informationStatistical Inference Based on Extremum Estimators
T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0
More informationKernel density estimator
Jauary, 07 NONPARAMETRIC ERNEL DENSITY ESTIMATION I this lecture, we discuss kerel estimatio of probability desity fuctios PDF Noparametric desity estimatio is oe of the cetral problems i statistics I
More informationSOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker
SOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker CHAPTER 9. POINT ESTIMATION 9. Covergece i Probability. The bases of poit estimatio have already bee laid out i previous chapters. I chapter 5
More informationEcon 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara
Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio
More information11 THE GMM ESTIMATION
Cotets THE GMM ESTIMATION 2. Cosistecy ad Asymptotic Normality..................... 3.2 Regularity Coditios ad Idetificatio..................... 4.3 The GMM Iterpretatio of the OLS Estimatio.................
More informationSlide Set 13 Linear Model with Endogenous Regressors and the GMM estimator
Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday
More informationThis section is optional.
4 Momet Geeratig Fuctios* This sectio is optioal. The momet geeratig fuctio g : R R of a radom variable X is defied as g(t) = E[e tx ]. Propositio 1. We have g () (0) = E[X ] for = 1, 2,... Proof. Therefore
More informationMath 341 Lecture #31 6.5: Power Series
Math 341 Lecture #31 6.5: Power Series We ow tur our attetio to a particular kid of series of fuctios, amely, power series, f(x = a x = a 0 + a 1 x + a 2 x 2 + where a R for all N. I terms of a series
More information(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3
MATH 337 Sequeces Dr. Neal, WKU Let X be a metric space with distace fuctio d. We shall defie the geeral cocept of sequece ad limit i a metric space, the apply the results i particular to some special
More informationFall 2013 MTH431/531 Real analysis Section Notes
Fall 013 MTH431/531 Real aalysis Sectio 8.1-8. Notes Yi Su 013.11.1 1. Defiitio of uiform covergece. We look at a sequece of fuctios f (x) ad study the coverget property. Notice we have two parameters
More informationLECTURE 8: ASYMPTOTICS I
LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece
More informationJanuary 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS
Jauary 25, 207 INTRODUCTION TO MATHEMATICAL STATISTICS Abstract. A basic itroductio to statistics assumig kowledge of probability theory.. Probability I a typical udergraduate problem i probability, we
More informationLast Lecture. Biostatistics Statistical Inference Lecture 16 Evaluation of Bayes Estimator. Recap - Example. Recap - Bayes Estimator
Last Lecture Biostatistics 60 - Statistical Iferece Lecture 16 Evaluatio of Bayes Estimator Hyu Mi Kag March 14th, 013 What is a Bayes Estimator? Is a Bayes Estimator the best ubiased estimator? Compared
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013 Fuctioal Law of Large Numbers. Costructio of the Wieer Measure Cotet. 1. Additioal techical results o weak covergece
More informationReview Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn
Stat 366 Lab 2 Solutios (September 2, 2006) page TA: Yury Petracheko, CAB 484, yuryp@ualberta.ca, http://www.ualberta.ca/ yuryp/ Review Questios, Chapters 8, 9 8.5 Suppose that Y, Y 2,..., Y deote a radom
More informationEFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS
EFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS Ryszard Zieliński Ist Math Polish Acad Sc POBox 21, 00-956 Warszawa 10, Polad e-mail: rziel@impagovpl ABSTRACT Weak laws of large umbers (W LLN), strog
More informationLecture 11 October 27
STATS 300A: Theory of Statistics Fall 205 Lecture October 27 Lecturer: Lester Mackey Scribe: Viswajith Veugopal, Vivek Bagaria, Steve Yadlowsky Warig: These otes may cotai factual ad/or typographic errors..
More informationLast Lecture. Wald Test
Last Lecture Biostatistics 602 - Statistical Iferece Lecture 22 Hyu Mi Kag April 9th, 2013 Is the exact distributio of LRT statistic typically easy to obtai? How about its asymptotic distributio? For testig
More informationSolutions: Homework 3
Solutios: Homework 3 Suppose that the radom variables Y,...,Y satisfy Y i = x i + " i : i =,..., IID where x,...,x R are fixed values ad ",...," Normal(0, )with R + kow. Fid ˆ = MLE( ). IND Solutio: Observe
More informationf n (x) f m (x) < ɛ/3 for all x A. By continuity of f n and f m we can find δ > 0 such that d(x, x 0 ) < δ implies that
Lecture 15 We have see that a sequece of cotiuous fuctios which is uiformly coverget produces a limit fuctio which is also cotiuous. We shall stregthe this result ow. Theorem 1 Let f : X R or (C) be a
More informationRates of Convergence by Moduli of Continuity
Rates of Covergece by Moduli of Cotiuity Joh Duchi: Notes for Statistics 300b March, 017 1 Itroductio I this ote, we give a presetatio showig the importace, ad relatioship betwee, the modulis of cotiuity
More informationECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization
ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where
More informationLecture 10: Universal coding and prediction
0-704: Iformatio Processig ad Learig Sprig 0 Lecture 0: Uiversal codig ad predictio Lecturer: Aarti Sigh Scribes: Georg M. Goerg Disclaimer: These otes have ot bee subjected to the usual scrutiy reserved
More informationREAL ANALYSIS II: PROBLEM SET 1 - SOLUTIONS
REAL ANALYSIS II: PROBLEM SET 1 - SOLUTIONS 18th Feb, 016 Defiitio (Lipschitz fuctio). A fuctio f : R R is said to be Lipschitz if there exists a positive real umber c such that for ay x, y i the domai
More informationlim za n n = z lim a n n.
Lecture 6 Sequeces ad Series Defiitio 1 By a sequece i a set A, we mea a mappig f : N A. It is customary to deote a sequece f by {s } where, s := f(). A sequece {z } of (complex) umbers is said to be coverget
More informationResampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.
Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator
More informationLecture 3: MLE and Regression
STAT/Q SCI 403: Itroductio to Resamplig Methods Sprig 207 Istructor: Ye-Chi Che Lecture 3: MLE ad Regressio 3. Parameters ad Distributios Some distributios are idexed by their uderlyig parameters. Thus,
More information4. Partial Sums and the Central Limit Theorem
1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems
More informationNotes 5 : More on the a.s. convergence of sums
Notes 5 : More o the a.s. covergece of sums Math 733-734: Theory of Probability Lecturer: Sebastie Roch Refereces: Dur0, Sectios.5; Wil9, Sectio 4.7, Shi96, Sectio IV.4, Dur0, Sectio.. Radom series. Three-series
More informationECE 901 Lecture 13: Maximum Likelihood Estimation
ECE 90 Lecture 3: Maximum Likelihood Estimatio R. Nowak 5/7/009 The focus of this lecture is to cosider aother approach to learig based o maximum likelihood estimatio. Ulike earlier approaches cosidered
More informationMaximum Likelihood Estimation
ECE 645: Estimatio Theory Sprig 2015 Istructor: Prof. Staley H. Cha Maximum Likelihood Estimatio (LaTeX prepared by Shaobo Fag) April 14, 2015 This lecture ote is based o ECE 645(Sprig 2015) by Prof. Staley
More informationLecture 2: Monte Carlo Simulation
STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?
More informationMATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED
MATH 47 / SPRING 013 ASSIGNMENT : DUE FEBRUARY 4 FINALIZED Please iclude a cover sheet that provides a complete setece aswer to each the followig three questios: (a) I your opiio, what were the mai ideas
More informationDetailed proofs of Propositions 3.1 and 3.2
Detailed proofs of Propositios 3. ad 3. Proof of Propositio 3. NB: itegratio sets are geerally omitted for itegrals defied over a uit hypercube [0, s with ay s d. We first give four lemmas. The proof of
More informationST5215: Advanced Statistical Theory
ST525: Advaced Statistical Theory Departmet of Statistics & Applied Probability Tuesday, September 7, 2 ST525: Advaced Statistical Theory Lecture : The law of large umbers The Law of Large Numbers The
More information7.1 Convergence of sequences of random variables
Chapter 7 Limit theorems Throughout this sectio we will assume a probability space (Ω, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite
More informationLecture 6 Ecient estimators. Rao-Cramer bound.
Lecture 6 Eciet estimators. Rao-Cramer boud. 1 MSE ad Suciecy Let X (X 1,..., X) be a radom sample from distributio f θ. Let θ ˆ δ(x) be a estimator of θ. Let T (X) be a suciet statistic for θ. As we have
More informationMachine Learning Brett Bernstein
Machie Learig Brett Berstei Week Lecture: Cocept Check Exercises Starred problems are optioal. Statistical Learig Theory. Suppose A = Y = R ad X is some other set. Furthermore, assume P X Y is a discrete
More informationEmpirical Process Theory and Oracle Inequalities
Stat 928: Statistical Learig Theory Lecture: 10 Empirical Process Theory ad Oracle Iequalities Istructor: Sham Kakade 1 Risk vs Risk See Lecture 0 for a discussio o termiology. 2 The Uio Boud / Boferoi
More informationFirst Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise
First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >
More informationAsymptotic Results for the Linear Regression Model
Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/5.070J Fall 203 Lecture 3 9//203 Large deviatios Theory. Cramér s Theorem Cotet.. Cramér s Theorem. 2. Rate fuctio ad properties. 3. Chage of measure techique.
More informationAdvanced Stochastic Processes.
Advaced Stochastic Processes. David Gamarik LECTURE 2 Radom variables ad measurable fuctios. Strog Law of Large Numbers (SLLN). Scary stuff cotiued... Outlie of Lecture Radom variables ad measurable fuctios.
More informationEstimation for Complete Data
Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of
More informationCS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5
CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio
More informationApplication to Random Graphs
A Applicatio to Radom Graphs Brachig processes have a umber of iterestig ad importat applicatios. We shall cosider oe of the most famous of them, the Erdős-Réyi radom graph theory. 1 Defiitio A.1. Let
More informationNotes 19 : Martingale CLT
Notes 9 : Martigale CLT Math 733-734: Theory of Probability Lecturer: Sebastie Roch Refereces: [Bil95, Chapter 35], [Roc, Chapter 3]. Sice we have ot ecoutered weak covergece i some time, we first recall
More informationQuestions and Answers on Maximum Likelihood
Questios ad Aswers o Maximum Likelihood L. Magee Fall, 2008 1. Give: a observatio-specific log likelihood fuctio l i (θ) = l f(y i x i, θ) the log likelihood fuctio l(θ y, X) = l i(θ) a data set (x i,
More informationProduct measures, Tonelli s and Fubini s theorems For use in MAT3400/4400, autumn 2014 Nadia S. Larsen. Version of 13 October 2014.
Product measures, Toelli s ad Fubii s theorems For use i MAT3400/4400, autum 2014 Nadia S. Larse Versio of 13 October 2014. 1. Costructio of the product measure The purpose of these otes is to preset the
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013 Large Deviatios for i.i.d. Radom Variables Cotet. Cheroff boud usig expoetial momet geeratig fuctios. Properties of a momet
More information2.1. Convergence in distribution and characteristic functions.
3 Chapter 2. Cetral Limit Theorem. Cetral limit theorem, or DeMoivre-Laplace Theorem, which also implies the wea law of large umbers, is the most importat theorem i probability theory ad statistics. For
More informationSTAT Homework 1 - Solutions
STAT-36700 Homework 1 - Solutios Fall 018 September 11, 018 This cotais solutios for Homework 1. Please ote that we have icluded several additioal commets ad approaches to the problems to give you better
More informationStudy the bias (due to the nite dimensional approximation) and variance of the estimators
2 Series Methods 2. Geeral Approach A model has parameters (; ) where is ite-dimesioal ad is oparametric. (Sometimes, there is o :) We will focus o regressio. The fuctio is approximated by a series a ite
More information62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 +
62. Power series Defiitio 16. (Power series) Give a sequece {c }, the series c x = c 0 + c 1 x + c 2 x 2 + c 3 x 3 + is called a power series i the variable x. The umbers c are called the coefficiets of
More informationLecture 3 : Random variables and their distributions
Lecture 3 : Radom variables ad their distributios 3.1 Radom variables Let (Ω, F) ad (S, S) be two measurable spaces. A map X : Ω S is measurable or a radom variable (deoted r.v.) if X 1 (A) {ω : X(ω) A}
More informationNotes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley
Notes O Media ad Quatile Regressio James L. Powell Departmet of Ecoomics Uiversity of Califoria, Berkeley Coditioal Media Restrictios ad Least Absolute Deviatios It is well-kow that the expected value
More informationUniversity of Colorado Denver Dept. Math. & Stat. Sciences Applied Analysis Preliminary Exam 13 January 2012, 10:00 am 2:00 pm. Good luck!
Uiversity of Colorado Dever Dept. Math. & Stat. Scieces Applied Aalysis Prelimiary Exam 13 Jauary 01, 10:00 am :00 pm Name: The proctor will let you read the followig coditios before the exam begis, ad
More informationMA Advanced Econometrics: Properties of Least Squares Estimators
MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample
More informationExponential Families and Bayesian Inference
Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where
More information