Point Estimation: properties of estimators 1 FINITE-SAMPLE PROPERTIES. finite-sample properties (CB 7.3) large-sample properties (CB 10.

Size: px
Start display at page:

Download "Point Estimation: properties of estimators 1 FINITE-SAMPLE PROPERTIES. finite-sample properties (CB 7.3) large-sample properties (CB 10."

Transcription

1 Poit Estimatio: properties of estimators fiite-sample properties CB 7.3) large-sample properties CB 10.1) 1 FINITE-SAMPLE PROPERTIES How a estimator performs for fiite umber of observatios. Estimator: W Parameter: θ Criteria for evaluatig estimators: Bias: does EW = θ? Variace of W you would like a estimator with a smaller variace) Example: X 1,..., X i.i.d. µ, σ 2 ) Ukow parameters are µ ad σ 2. Cosider: ˆµ 1 i X i, estimator of µ ˆσ 2 1 i X i X ) 2, estimator of σ 2. Bias: E ˆµ = 1 µ = µ. So ubiased. Var ˆµ = 1 2 σ 2 = 1 σ2. Eˆσ 2 = E 1 X i X ) 2 ) i = 1 EXi 2 2EX i X + E X ) 2 i = 1 [ ] µ 2 + σ 2 ) 2µ 2 + σ2 ) + σ2 + µ2 = 1 σ2. 1

2 Hece it is biased. To fix this bias, cosider the estimator s 2 1 ubiased). 1 i X i X ) 2, ad Es 2 = σ 2 Mea-squared error MSE) of W is EW θ) 2. Commo criterio for comparig estimators. Decompose: MSEW ) = V W + EW θ) 2 =Variace + Bias) 2. Hece, for a ubiased estimator: MSEW ) = V W. Example: X 1,..., X U[0, θ]. fx) = 1/θ, x [0, θ]. Cosider estimator ˆθ 2 X. E ˆθ = 2 1 E i X i = MSEˆθ N ) = V ˆθ = 4 2 i V X i = θ2 3 θ = θ. So ubiased Cosider estimator θ max X 1,..., X ). I order to derive momets, start by derivig CDF: P θ z) = P X 1 z, X 2 z,..., X z) = P X i z) i=1 { z ) if z θ = θ 1 otherwise Therefore z) = z 1 1 f θ θ), for 0 x θ. θ Bias θ ) = θ/ + 1) E θ ) 2 = θ θ 0 z+1 dz = Hece V θ = θ 2 +2 θ z ) 1 1 E θ ) = z 0 θ θ dz = θ z dz = θ + 1 θ. +2 θ2. +1 Accordigly, MSE= 2θ 2 +2)+1) ) 2 ) 0 = θ 2 +2)+1) 2. 2

3 Cotiue the previous example. Redefie θ = +1 maxx 1,..., X ). Now both estimators ˆθ ad θ are ubiased. Which is better? V ˆθ = θ2 = O1/). 3 V θ = ) ) +1 2 V maxx1,..., X )) = θ 2 1 = O1/ 2 ). +2) Hece, for large eough, θ has a smaller variace, ad i this sese it is better. Best ubiased estimator: if you choose the best i terms of MSE) estimator, ad restrict yourself to ubiased estimators, the the best estimator is the oe with the lowest variace. A best ubiased estimator is also called the Uiform miimum variace ubiased estimator UMVUE). Formally: a estimator W is a UMVUE of θ satisfies: i) E θ W = θ, for all θ ubiasedess) ii) V θ W V θ W, for all θ, ad all other ubiased estimators W. The uiform coditio is crucial, because it is always possible to fid estimators which have zero variace for a specific value of θ. It is difficult i geeral to verify that a estimator W is UMVUE, sice you have to verify coditio ii) of the defiitio, that V W is smaller tha all other ubiased estimators. Luckily, we have a importat result for the lowest attaiable variace of a estimator. Theorem Cramer-Rao Iequality): Let X 1,..., X be a sample with joit pdf f X θ), ad let W X) be ay estimator satisfyig d i) dθ E θw X) [ = W X) θ f X θ) ] dx; The V θ W X) ii) V θ W X) <. E θ d E dθ θw X) ) 2 log f X θ) θ ) 2. 3

4 The RHS above is called the Cramer-Rao Lower Boud. Proof: CB, pg I short, we apply the Cauchy-Schwarz iequality V S) covs, T )/V T ) with S = W X) ad T = θ log f X θ). The choice of T here may seem a bit arbitrary. To get some ituitio, cosider Cramer s derivatio. 1 Start with the followig maipulatio of the equality E θ W X) = W X)fX θ)dx: d dθ E θw X) = d W X)fX θ)dx dθ = W X) θ fx θ)dx = W X) E θ W X)) θ fx θ)dx ote E θ W X) fx θ)dx = 0) θ ) = W X) E θ W X)) log fx θ) fx θ)dx θ Applyig the Cauchy-Schwarz iequality, we have [ d dθ E θw X)] 2 V ar θ W X) E θ [ log fx θ)]2 θ or V ar θ W X) [ d dθ E θw X)] 2 E θ [ θ log fx θ)]2. The LHS of coditio i) above is d dθ W X)fX θ)dx, so by Leibiz rule, this coditio rules out cases where the support of X is depedet o θ. The crucial step i the derivatio of the CR-boud is the iterchage of differetiatio ad itegratio which implies E θ θ log f X θ) = 1 f X θ) = f X θ)dx θ = θ 1 = 0 1 Cramer, Mathematical Methods of Statistics, p. 475ff. f X θ) f X θ)dx θ 1) 4

5 skip) The above derivatio is oteworthy, because θ log f X θ) = 0 is the FOC of maximum likelihood estimatio problem. Alteratively, as i CB, apply coditio i) of CR result, usig W X) = 1.) I the i.i.d. case, this becomes the sample average 1 θ log fx i θ) = 0. i Ad by the LLN: 1 i θ log fx i θ) p E θ0 θ log fx i θ), where θ 0 is the true value of θ 0. This shows that maximum likelihood estimatio of θ is equivalet to estimatio based o the momet coditio E θ0 θ log fx i θ) = 0 which holds oly at the true value θ = θ 0. Thus MLE is cosistet for the true value θ 0, as we ll see later.) However, ote that Eq. 1) holds at all values of θ, ot just θ 0.) [Thik about] What if model is misspecified, i the sese that true desity of X is g x), ad that for all θ Θ, f x θ) g x) that is, there is o value of the parameter θ such that the postulated model f coicides with the true model g)? Does Eq. 1) still hold? What is MLE lookig for? I the iid case, the CR lower boud ca be simplified Corollary : if X 1,..., X i.i.d. fx θ), the V θ W X) d E dθ θw X) ) 2 E θ θ log fx θ)) 2. Up to this poit, Cramer-Rao results ot that operatioal for us to fid a best estimator, because the estimator W X) is o both sides of the iequality. However, for a ubiased estimator, how ca you simplify the expressio further? 5

6 Example: X 1,..., X i.i.d. Nµ, σ 2 ). What is CRLB for a ubiased estimator of µ? Ubiased umerator =1. log fx θ) = log 2π log σ 1 ) 2 x µ 2 σ ) ) x µ 1 log fx θ) = = x µ µ σ σ σ 2 ) 2 E log fx θ) = E X µ) 2 σ 4) = 1 θ σ V X = 1 4 σ. 2 Hece the CRLB = 1 1 σ 2 = σ2. This is the variace of the sample mea, so that the sample mea is a UMVUE for µ. Sometimes we ca simplify the deomiator of the CRLB further: Lemma Iformatio iequality): if fx θ) satisfies ) [ ) ] d *) dθ E θ log fx θ) = log fx θ) fx θ) dx, θ θ θ the ) 2 ) 2 E θ θ log fx θ) = E θ log fx θ). θ2 Rough proof: LHS of *): Usig Eq. 1) above, we get that LHS of *) =0. RHS of *): [ ) ] θ θ log f f dx 2 ) 2 log f 1 f = fdx + dx θ 2 f θ ) 2 =E 2 log f log f + E. θ 2 θ Puttig the LHS ad RHS together yields the desired result. 6

7 The LHS of the above coditio *) is just d log fx θ)) fx θ)dx. As dθ θ before, the crucial step is the iterchage of differetiatio ad itegratio. skip) Also, the iformatio iequality depeds crucially o the equality E θ log fx θ) = θ 0, which depeds o the correct specificatio of the model. Thus iformatio iequality ca be used as basis of specificatio test. How?) Example: for the previous example, cosider CRLB for ubiased estimator of σ 2. We ca use the iformatio iequality, because coditio *) is satisfied for the ormal distributio. Hece: E 1 log fx θ) = σ2 2σ + 1 x µ) 2 ) 2 2 σ 4 log fx θ) = 1 x µ)2 σ 2 σ2 2σ4 σ )) 6 1 log fx θ) = σ2 2σ 1 ) 4 σ 4 σ 2 Hece the CRLB is 2σ4. = 1 2σ 4. Example: X 1,..., X U[0, θ]. Check coditios for CRLB for a ubiased estimator W X) of θ. d EW X) = 1 because it is ubiased) dθ [ W X)f θ X θ) ] dx = ) W X) 1 dx d dθ EW X) = 1 θ 2 Hece, coditio i) of theorem ot satisfied. skip) But whe ca CRLB if it exists) be attaied? Corollary : X 1,..., X i.i.d. fx θ), satisfyig the coditios of CR theorem. Let Lθ X) = i=1 fx θ) deote the likelihood fuctio. Estimator W X) ubiased for θ W X) attais CRLB iff you ca write θ log Lθ X) [ = aθ) W X) ] θ for some fuctio aθ). 7

8 Example: X 1,..., X i.i.d. Nµ, σ 2 ) Cosider estimatig µ: µ log Lθ X) = µ ) log fx θ) i=1 = log 2π log σ 1 ) ) X µ) 2 µ 2 σ 2 i=1 ) X µ = i=1 σ 2 = σ 2 X µ). Hece, CRLB ca be attaied i fact, we showed earlier that CRLB attaied by X ) Loss fuctio optimality Let X f X θ). Cosider a loss fuctio Lθ, W X)), takig values i [0, + ), which pealizes you whe your W X) estimator is far from the true parameter θ. Note that Lθ, W X)) is a radom variable, sice X ad W X)) are radom. Cosider estimators which miimize expected loss: that is mi E θlθ, W X)) mi Rθ, W )) W ) W ) where Rθ, W )) is the risk fuctio. Note: the risk fuctio is ot a radom variable, because X has bee itegrated out.) Loss fuctio optimality is a more geeral criterio tha miimum MSE. I fact, 2, because MSEW X)) = E θ W X) θ) the MSE is actually the risk fuctio associated with the quadratic loss fuctio Lθ, W X)) 2. = W X) θ) Other examples of loss fuctios: Absolute error loss: W X) θ 8

9 Relative quadratic error loss: W X) θ) 2 θ +1 The exercise of miimizig risk takes a give value of θ as give, so that the miimized risk of a estimator depeds o whichever value of θ you are cosiderig. You might be iterested i a estimator which does well regardless of which value of θ you are cosiderig. Aalogous to the focus o the uiform miimal variace.) For this differet problem, you wat to cosider a otio of risk which does ot deped o θ. Two criteria which have bee cosidered are: Average risk: mi W ) Rθ, W ))hθ)dθ. where hθ) is some weightig fuctio across θ. I a Bayesia iterpretatio, hθ) is a prior desity over θ.) Mimax criterio: mi max Rθ, W )). W ) θ Here you choose the estimator W ) to miimize the maximum risk = max θ Rθ, W )), where θ is set to the worse value. So mimax optimizer is the best that ca be achieved i a worst-case sceario. Example: X 1,..., X i.i.d. Nµ, σ 2 ). Sample mea X is: ubiased miimum MSE UMVUE attais CRLB miimizes expected quadratic loss 9

10 2 LARGE SAMPLE PROPERTIES OF ESTIMATORS It ca be difficult to compute MSE, risk fuctios, etc., for some estimators, especially whe estimator does ot resemble a sample average. Large-sample properties: exploit LLN, CLT Cosider { data {X 1, X 2,...} by which we costruct a sequece of estimators W W X 1 ), W X } 2 ),.... W is a radom sequece. Defie: we say that W is cosistet for a parameter θ iff the radom sequece W as coverges i some stochastic sese) to θ. Strog cosistecy obtais whe W θ. p Weak cosistecy obtais whe W θ. For estimators like sample-meas, cosistecy either weak or strog) follows easily usig a LLN. Cosistecy ca be thought of as the large-sample versio of ubiasedess. Defie: a M-estimator is a estimator of θ which a maximizer of a objective fuctio Q θ). Examples: MLE: Q θ) = 1 i=1 log fx i θ) Least squares: Q θ) = i=1 [y i gx i ; θ)] 2. OLS is special case whe gx i ; θ) = α + X iβ. GMM: Q θ) = G θ) W θ)g θ) where G θ) = [ 1 m 1 x i ; θ), 1 i=1 m 2 x i ; θ),..., 1 i=1 m M x i ; θ)], i=1 a M 1 vector of sample momet coditios, ad W is a M M weightig matrix. Notatio: For each θ Θ, let f θ fx 1,..., x,... ; θ) deote the joit desity of the data for the give value of θ. For θ 0 Θ, we deote the limit objective fuctio Q 0 θ) = plim,fθ0 Q θ) at each θ). 10

11 Cosistecy of M-estimators Make the followig assumptios: 1. For each θ 0 Θ, the limitig objective fuctio Q 0 θ) is uiquely maximized at θ 0 idetificatio ) 2. Parameter space Θ is a compact subset of R K. 3. Q 0 θ) is cotiuous i θ 4. Q θ) coverges uiformly i probability to Q 0 θ); that is: p sup Q θ) Q 0 θ) 0. θ Θ Theorem: Cosistecy of M-Estimator) Uder assumptio 1,2,3,4, θ p θ0. Proof: We eed to show: for ay arbitrarily small eighborhood N cotaiig θ 0, P θ N ) 1. For large eough, the uiform covergece coditios that, for all ɛ, δ > 0, ) P sup Q θ) Q 0 θ) < ɛ/2 θ Θ > 1 δ. The evet sup θ Θ Q θ) Q 0 θ) < ɛ/2 implies Similarly, Q θ ) Q 0 θ ) < ɛ/2 Q 0 θ ) > Q θ ) ɛ/2 2) Q θ 0 ) Q 0 θ 0 ) > ɛ/2 Q θ 0 ) > Q 0 θ 0 ) ɛ/2. 3) Sice θ = argmax θ Q θ), Eq. 2) implies Hece, addig Eqs. 3) ad 4), we have So we have show that Q 0 θ ) > Q θ 0 ) ɛ/2. 4) Q 0 θ ) > Q 0 θ 0 ) ɛ. 5) sup Q θ) Q 0 θ) < ɛ/2 = Q 0 θ ) > Q 0 θ 0 ) ɛ θ Θ ) P Q 0 θ ) > Q 0 θ 0 ) ɛ) P sup Q θ) Q 0 θ) < ɛ/2 θ Θ 1. 11

12 Now defie N as ay ope eighborhood of R K, which cotais θ 0, ad N is the complemet of N i R K. The Θ N is compact, so that max θ Θ N Q 0 θ) exists. Set ɛ = Q 0 θ 0 ) max θ Θ N Q 0 θ). The Q 0 θ ) > Q 0 θ 0 ) ɛ Q 0 θ ) > max Q 0 θ) θ Θ N θ N P θ N ) P Q 0 θ ) > Q 0 θ 0 ) ɛ) 1. Sice the argumet above holds for ay arbitrarily small eighborhood of θ 0, we are doe. I geeral, the limit objective fuctio Q 0 θ) = plim Q θ) may ot be that straightforward to determie. But i may cases, Q θ) is a sample average of some sort: Q θ) = 1 qx i θ) eg. least squares, MLE). The by a law of large umbers, we coclude that for all θ) Q 0 θ) = plim 1 qx i θ) = E xi qx i θ) i where E xi deote expectatio with respect to the true but uobserved) distributio of x i. skip) Most of the time, θ 0 ca be iterpreted as a true value. But if model is misspecified, the this iterpretatio does t hold ideed, uder misspecificatio, ot eve clear what the true value is). So a more cautious way to iterpret the cosistecy result is that i θ p argmaxθ Q 0 θ) which holds give the coditios) o matter whether model is correctly specified. skip) Cosider the abstract) mappig from parameter values θ to joit desities of the data f θ = fx 1,..., x,... ; θ). The idetificatio coditio posits: for each θ 1, the limit objective fuctio Q 1 plim fθ1 Q θ) i uiquely optimized at θ 1. This implies that the mappig θ f θ is oe-to-oe. Assume ot: that is, there are two values θ 1 θ 2 such that f θ1 = f θ2 a.s.). 2 The θ 1 = argmi θ Q 1 θ) = argmi θ Q 2 θ) = θ 2. 2 I other words, θ 1 ad θ 2 are observatioally equivalet. 12

13 ** Let s upack the uiform covergece coditio. Sufficiet coditios for this coditios are: 1. Poitwise covergece: For each θ Θ, Q θ) Q 0 θ) = o p 1). 2. Q θ) is stochastically equicotiuous: for every ɛ > 0, η > 0 there exists a sequece of radom variable ɛ, η) ad ɛ, η) such that for all >, P > ɛ) < η ad for each θ there is a ope set N cotaiig θ with sup θ N Q θ) Q θ), >. Note that both ad do ot deped o θ: it is uiform result. This is a i probability versio of the determiistic otio of uiform equicotiuity: we say a sequece of determiistic fuctios R θ) is uiformly equicotiuous if, for every ɛ > 0 there exists δɛ) ad ɛ) such that for all θ sup θ: θ θ <δ R θ) R θ) ɛ, >. To uderstad this more ituitively, cosider a aive argumet for cosistecy. By cotiuity of Q 0, we kow that Q 0 θ) is close to Q 0 θ 0 ) for θ N θ 0 ). By poitwise covergece, we have Q θ) covergig to Q 0 θ) for all θ; hece, eve if Q θ) is ot optimized by θ 0, the optimizer θ = argmax θ Q θ) should ot be far from θ 0. Implicitly, for the last part, we have assumed that Q θ) is equally close to Q 0 θ) for all θ, because the the optimizers of Q ad Q 0 caot be too far apart. However, poitwise covergece is ot eough to esure this equally closeess. At ay give, Q θ 0 ) beig close to Q 0 θ 0 ) does ot imply this at other poits. Uiform covergece esures, roughly speakig, that at ay give, Q ad Q 0 are equally close at all poits θ. As the above discussio shows, uiform covergece is related to a cotiuity coditio o Q ); if Q ) is ot cotiuous, the it ca be maximized far from θ 0. Stochastic equicotiuity is the right otio of cotiuity here sice Q ) ad θ are both radom. Asymptotic ormality for M-estimators Defie the score vector [ Q θ) θq θ) =,..., Q θ) θ 1 θ K. θ= θ] θ= θ 13

14 Similarly, defie the K K Hessia matrix [ θ θ Q θ) ] i,j = 2 Q θ) θ i θ j Note that the Hessia is symmetric. Make the followig assumptios: 1 i, j K. θ= θ, 1. θ = argmax θ Q θ) p θ 0 2. θ 0 iteriorθ) 3. Q θ) is twice cotiuously differetiable i a eighborhood N of θ θ0 Q θ) d N0, Σ) 5. Uiform covergece of Hessia: there exists the matrix Hθ) which is cotiuous at θ 0 ad sup θ N θθ Q θ) Hθ) 0. p 6. Hθ 0 ) is osigular Theorem Asymptotic ormality for M-estimator): Uder assumptios 1,2,3,4,5, where H 0 Hθ 0 ). θ θ 0 ) d N0, H 1 0 ΣH 1 0 ) Proof: sketch) By Assumptios 1,2,3, θ Q θ) = 0 this is FOC of maximizatio problem). The usig mea-value theorem with θ deotig mea value): 0 = θ Q θ) = θ0 Q θ) + θ θ Q θ)θ θ 0 ) θ θ 0 ) = θ0 Q θ) }{{} θ θ Q θ) }{{} p H 0 usig A5) d N0,Σ) usig A4) θ θ 0 ) d Hθ 0 ) 1 N0, Σ) = N0, H 1 0 ΣH 1 0 ). Note: A5 is a uiform covergece assumptio o the sample Hessia. Give previous discussio, it esures that the sample Hessia θθ Q θ) evaluated at θ which is close to θ 0 ) does ot vary far from the limit Hessia Hθ) at θ 0, which is implied by a type of cotiuity of the sample Hessia close to θ 0. 14

15 2.1 Maximum likelihood estimatio The cosistecy of MLE ca follow by applicatio of the theorem above for cosistecy of M-estimators. Essetially, as we oted above,what the cosistecy theorem showed above was that, for ay M-estimator sequece θ : plim θ = argmax θ Q 0 θ). For MLE, there is a distict ad earlier argumet due to Wald 1949), who shows that, i the i.i.d. case, the limitig likelihood fuctio correspodig to Q 0 θ)) is ideed globally maximized at θ 0, the true value. Thus, we ca directly cofirm the idetificatio assumptio of the M-estimator cosistecy theorem. This argumet is of iterest by itself. Argumet: summary of Amemiya, pp ) MLE 1 Defie ˆθ argmax θ i log fx i θ). Let θ 0 deote the true value. By LLN: 1 i log fx i θ) p E θ0 log fx i θ), for all θ ot ecessarily the true θ 0 ). ) ) By Jese s iequality: E θ0 log fx θ) fx θ 0 < log E fx θ) ) θ0 fx θ 0 ) ) But E fx θ) θ0 fx θ 0 = ) fx θ) ) fx θ 0 fx θ ) 0 ) = 1, sice fx θ) is a desity fuctio, for all θ. 3 Hece: ) fx θ) E θ0 log < 0, θ fx θ 0 ) = E θ0 log fx θ) < E θ0 log fx θ 0 ), θ = E θ0 log fx θ) is maximized at the true θ 0. This is the idetificatio assumptio from the M-estimator cosistecy theorem. 3 I this step, ote the importace of assumptio A3 i CB, pg If x has support depedig o θ, the it will ot itegrate to 1 for all θ. 15

16 skip) Aalogously, we also kow that, for δ > 0, µ 1 = E θ0 log fx; θ ) 0 δ) < 0; µ 2 = E θ0 log fx; θ ) 0 + δ) fx; θ 0 ) fx; θ 0 ) i < 0. By the SLLN, we kow that 1 log fx ) i; θ 0 δ) = 1 fx i ; θ 0 ) [log L x; θ 0 δ) log L x : θ 0 )] as µ 1 so that, with probability 1, log L x; θ 0 δ) < log L x; θ 0 ) for large eough. Similarly, for large eough, log L x; θ 0 + δ) < log L x; θ 0 ) with probability 1. Hece, for large, ˆθ argmax θ log L x; θ) θ 0 δ, θ 0 + δ). That is, the MLE ˆθ is strogly cosistet for θ 0. Note that this argumet requires weaker assumptios tha the M-estimator cosistecy theorem above. Now we cosider aother idea, efficiecy, which is a large-sample aalogue of the miimum variace cocept. For the sequece of estimators W, suppose that k)w θ) d N0, σ 2 ) where k) is a polyomial i. The σ 2 is deoted the asymptotic variace of W. I usual cases, k) =. For example, by the CLT, we kow that X µ) d N0, σ 2 ). Hece, σ 2 is the asymptotic variace of the sample mea X. Defiitio : A estimator sequece W is asymptotically efficiet for θ if W θ) d N0, vθ)), where 1 the asymptotic variace vθ) = the E θ0 log fx θ))2 θ CRLB) 16

17 By asymptotic ormality result for M-estimator, we kow what the asymptotic distributio for the MLE should be. However, it turs out give the iformatio iequality, the MLE s asymptotic distributio ca be further simplified. Theorem : Asymptotic efficiecy of MLE Proof: followig Amemiya, pp ) MLE ˆθ satisfies the FOC of the MLE problem: Usig the mea value theorem: 0 = log Lθ X ) θ 0 = log Lθ X ) θ θ=θ0 + 2 log Lθ X ) θ 2 = ) ˆθ θ 0 = log Lθ X ) θ θ=θ0 2 log Lθ X ) = θ=θ θ 2 θ=ˆθmle. θ=θ ˆθMLE 1 1 ) θ 0 i log fx i θ) i θ θ=θ0 2 log fx i θ) θ 2 θ=θ ) Note that, by the LLN, 1 i log fx i θ) θ θ=θ0 p log fx θ) Eθ0 θ θ=θ0 = fxi θ) θ θ=θ0 dx. Usig same argumet as i the iformatio iequality result above, the last term is: f θ dx = fdx = 0. θ Hece, the CLT ca be applied to the umerator of **): ) ) umerator of **) d log fxi θ) 2 N 0, E θ0 θ θ=θ0. By LLN, ad uiform covergece of Hessia term: deomiator of **) p E θ0 2 log fx θ) θ 2 θ=θ0. 17

18 Hece, by Slutsky theorem: ˆθ θ 0 ) d N By the iformatio iequality: so that E log fxi ) 2 θ) θ0 θ θ=θ0 0, [ ] 2. E 2 log fx θ) θ=θ0 θ0 θ 2 ) log fxi θ) 2 E θ0 2 log fx θ) θ θ=θ0 = E θ0 θ 2 ˆθ θ 0 ) d N 0, so that the asymptotic variace is the CRLB. 1 E θ0 log fxi θ) θ θ=θ0 ) 2 θ=θ0 Hece, the asymptotic approximatio for the fiite-sample distributio is ˆθ MLE a N θ 0, 1 1 E θ0 log fxi θ) θ ) 2. θ=θ0 18

Lecture 19: Convergence

Lecture 19: Convergence Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may

More information

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f. Lecture 5 Let us give oe more example of MLE. Example 3. The uiform distributio U[0, ] o the iterval [0, ] has p.d.f. { 1 f(x =, 0 x, 0, otherwise The likelihood fuctio ϕ( = f(x i = 1 I(X 1,..., X [0,

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

Unbiased Estimation. February 7-12, 2008

Unbiased Estimation. February 7-12, 2008 Ubiased Estimatio February 7-2, 2008 We begi with a sample X = (X,..., X ) of radom variables chose accordig to oe of a family of probabilities P θ where θ is elemet from the parameter space Θ. For radom

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

Introductory statistics

Introductory statistics CM9S: Machie Learig for Bioiformatics Lecture - 03/3/06 Itroductory statistics Lecturer: Sriram Sakararama Scribe: Sriram Sakararama We will provide a overview of statistical iferece focussig o the key

More information

Chapter 6 Infinite Series

Chapter 6 Infinite Series Chapter 6 Ifiite Series I the previous chapter we cosidered itegrals which were improper i the sese that the iterval of itegratio was ubouded. I this chapter we are goig to discuss a topic which is somewhat

More information

Lecture 12: September 27

Lecture 12: September 27 36-705: Itermediate Statistics Fall 207 Lecturer: Siva Balakrisha Lecture 2: September 27 Today we will discuss sufficiecy i more detail ad the begi to discuss some geeral strategies for costructig estimators.

More information

6. Sufficient, Complete, and Ancillary Statistics

6. Sufficient, Complete, and Ancillary Statistics Sufficiet, Complete ad Acillary Statistics http://www.math.uah.edu/stat/poit/sufficiet.xhtml 1 of 7 7/16/2009 6:13 AM Virtual Laboratories > 7. Poit Estimatio > 1 2 3 4 5 6 6. Sufficiet, Complete, ad Acillary

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

Sequences and Series of Functions

Sequences and Series of Functions Chapter 6 Sequeces ad Series of Fuctios 6.1. Covergece of a Sequece of Fuctios Poitwise Covergece. Defiitio 6.1. Let, for each N, fuctio f : A R be defied. If, for each x A, the sequece (f (x)) coverges

More information

Lecture 8: Convergence of transformations and law of large numbers

Lecture 8: Convergence of transformations and law of large numbers Lecture 8: Covergece of trasformatios ad law of large umbers Trasformatio ad covergece Trasformatio is a importat tool i statistics. If X coverges to X i some sese, we ofte eed to check whether g(x ) coverges

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS MASSACHUSTTS INSTITUT OF TCHNOLOGY 6.436J/5.085J Fall 2008 Lecture 9 /7/2008 LAWS OF LARG NUMBRS II Cotets. The strog law of large umbers 2. The Cheroff boud TH STRONG LAW OF LARG NUMBRS While the weak

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Empirical Processes: Glivenko Cantelli Theorems

Empirical Processes: Glivenko Cantelli Theorems Empirical Processes: Gliveko Catelli Theorems Mouliath Baerjee Jue 6, 200 Gliveko Catelli classes of fuctios The reader is referred to Chapter.6 of Weller s Torgo otes, Chapter??? of VDVW ad Chapter 8.3

More information

Probability and Statistics

Probability and Statistics ICME Refresher Course: robability ad Statistics Staford Uiversity robability ad Statistics Luyag Che September 20, 2016 1 Basic robability Theory 11 robability Spaces A probability space is a triple (Ω,

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Lecture 10 October Minimaxity and least favorable prior sequences

Lecture 10 October Minimaxity and least favorable prior sequences STATS 300A: Theory of Statistics Fall 205 Lecture 0 October 22 Lecturer: Lester Mackey Scribe: Brya He, Rahul Makhijai Warig: These otes may cotai factual ad/or typographic errors. 0. Miimaxity ad least

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets

More information

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values

Asymptotics. Hypothesis Testing UMP. Asymptotic Tests and p-values of the secod half Biostatistics 6 - Statistical Iferece Lecture 6 Fial Exam & Practice Problems for the Fial Hyu Mi Kag Apil 3rd, 3 Hyu Mi Kag Biostatistics 6 - Lecture 6 Apil 3rd, 3 / 3 Rao-Blackwell

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Problem Set 4 Due Oct, 12

Problem Set 4 Due Oct, 12 EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpeCourseWare http://ocw.mit.edu.00 Ucertaity i Egieerig Fall 2008 For iformatio about citig these materials or our Terms of Use, visit: http://ocw.mit.edu.terms. .00 - Brief Notes # 9 Poit ad Iterval

More information

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes.

Direction: This test is worth 250 points. You are required to complete this test within 50 minutes. Term Test October 3, 003 Name Math 56 Studet Number Directio: This test is worth 50 poits. You are required to complete this test withi 50 miutes. I order to receive full credit, aswer each problem completely

More information

Stat410 Probability and Statistics II (F16)

Stat410 Probability and Statistics II (F16) Some Basic Cocepts of Statistical Iferece (Sec 5.) Suppose we have a rv X that has a pdf/pmf deoted by f(x; θ) or p(x; θ), where θ is called the parameter. I previous lectures, we focus o probability problems

More information

LECTURE 2 LEAST SQUARES CROSS-VALIDATION FOR KERNEL DENSITY ESTIMATION

LECTURE 2 LEAST SQUARES CROSS-VALIDATION FOR KERNEL DENSITY ESTIMATION Jauary 3 07 LECTURE LEAST SQUARES CROSS-VALIDATION FOR ERNEL DENSITY ESTIMATION Noparametric kerel estimatio is extremely sesitive to te coice of badwidt as larger values of result i averagig over more

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Distribution of Random Samples & Limit theorems

Distribution of Random Samples & Limit theorems STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to

More information

An Introduction to Asymptotic Theory

An Introduction to Asymptotic Theory A Itroductio to Asymptotic Theory Pig Yu School of Ecoomics ad Fiace The Uiversity of Hog Kog Pig Yu (HKU) Asymptotic Theory 1 / 20 Five Weapos i Asymptotic Theory Five Weapos i Asymptotic Theory Pig Yu

More information

Statistical Theory MT 2008 Problems 1: Solution sketches

Statistical Theory MT 2008 Problems 1: Solution sketches Statistical Theory MT 008 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. a) Let 0 < θ < ad put fx, θ) = θ)θ x ; x = 0,,,... b) c) where α

More information

Statistical Theory MT 2009 Problems 1: Solution sketches

Statistical Theory MT 2009 Problems 1: Solution sketches Statistical Theory MT 009 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. (a) Let 0 < θ < ad put f(x, θ) = ( θ)θ x ; x = 0,,,... (b) (c) where

More information

Lecture 11 and 12: Basic estimation theory

Lecture 11 and 12: Basic estimation theory Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis

More information

LECTURE 14 NOTES. A sequence of α-level tests {ϕ n (x)} is consistent if

LECTURE 14 NOTES. A sequence of α-level tests {ϕ n (x)} is consistent if LECTURE 14 NOTES 1. Asymptotic power of tests. Defiitio 1.1. A sequece of -level tests {ϕ x)} is cosistet if β θ) := E θ [ ϕ x) ] 1 as, for ay θ Θ 1. Just like cosistecy of a sequece of estimators, Defiitio

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam. Probability ad Statistics FS 07 Secod Sessio Exam 09.0.08 Time Limit: 80 Miutes Name: Studet ID: This exam cotais 9 pages (icludig this cover page) ad 0 questios. A Formulae sheet is provided with the

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Output Analysis and Run-Length Control

Output Analysis and Run-Length Control IEOR E4703: Mote Carlo Simulatio Columbia Uiversity c 2017 by Marti Haugh Output Aalysis ad Ru-Legth Cotrol I these otes we describe how the Cetral Limit Theorem ca be used to costruct approximate (1 α%

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

Kernel density estimator

Kernel density estimator Jauary, 07 NONPARAMETRIC ERNEL DENSITY ESTIMATION I this lecture, we discuss kerel estimatio of probability desity fuctios PDF Noparametric desity estimatio is oe of the cetral problems i statistics I

More information

SOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker

SOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker SOME THEORY AND PRACTICE OF STATISTICS by Howard G. Tucker CHAPTER 9. POINT ESTIMATION 9. Covergece i Probability. The bases of poit estimatio have already bee laid out i previous chapters. I chapter 5

More information

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio

More information

11 THE GMM ESTIMATION

11 THE GMM ESTIMATION Cotets THE GMM ESTIMATION 2. Cosistecy ad Asymptotic Normality..................... 3.2 Regularity Coditios ad Idetificatio..................... 4.3 The GMM Iterpretatio of the OLS Estimatio.................

More information

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday

More information

This section is optional.

This section is optional. 4 Momet Geeratig Fuctios* This sectio is optioal. The momet geeratig fuctio g : R R of a radom variable X is defied as g(t) = E[e tx ]. Propositio 1. We have g () (0) = E[X ] for = 1, 2,... Proof. Therefore

More information

Math 341 Lecture #31 6.5: Power Series

Math 341 Lecture #31 6.5: Power Series Math 341 Lecture #31 6.5: Power Series We ow tur our attetio to a particular kid of series of fuctios, amely, power series, f(x = a x = a 0 + a 1 x + a 2 x 2 + where a R for all N. I terms of a series

More information

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3 MATH 337 Sequeces Dr. Neal, WKU Let X be a metric space with distace fuctio d. We shall defie the geeral cocept of sequece ad limit i a metric space, the apply the results i particular to some special

More information

Fall 2013 MTH431/531 Real analysis Section Notes

Fall 2013 MTH431/531 Real analysis Section Notes Fall 013 MTH431/531 Real aalysis Sectio 8.1-8. Notes Yi Su 013.11.1 1. Defiitio of uiform covergece. We look at a sequece of fuctios f (x) ad study the coverget property. Notice we have two parameters

More information

LECTURE 8: ASYMPTOTICS I

LECTURE 8: ASYMPTOTICS I LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece

More information

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS Jauary 25, 207 INTRODUCTION TO MATHEMATICAL STATISTICS Abstract. A basic itroductio to statistics assumig kowledge of probability theory.. Probability I a typical udergraduate problem i probability, we

More information

Last Lecture. Biostatistics Statistical Inference Lecture 16 Evaluation of Bayes Estimator. Recap - Example. Recap - Bayes Estimator

Last Lecture. Biostatistics Statistical Inference Lecture 16 Evaluation of Bayes Estimator. Recap - Example. Recap - Bayes Estimator Last Lecture Biostatistics 60 - Statistical Iferece Lecture 16 Evaluatio of Bayes Estimator Hyu Mi Kag March 14th, 013 What is a Bayes Estimator? Is a Bayes Estimator the best ubiased estimator? Compared

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013 MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 21 11/27/2013 Fuctioal Law of Large Numbers. Costructio of the Wieer Measure Cotet. 1. Additioal techical results o weak covergece

More information

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn

Review Questions, Chapters 8, 9. f(y) = 0, elsewhere. F (y) = f Y(1) = n ( e y/θ) n 1 1 θ e y/θ = n θ e yn Stat 366 Lab 2 Solutios (September 2, 2006) page TA: Yury Petracheko, CAB 484, yuryp@ualberta.ca, http://www.ualberta.ca/ yuryp/ Review Questios, Chapters 8, 9 8.5 Suppose that Y, Y 2,..., Y deote a radom

More information

EFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS

EFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS EFFECTIVE WLLN, SLLN, AND CLT IN STATISTICAL MODELS Ryszard Zieliński Ist Math Polish Acad Sc POBox 21, 00-956 Warszawa 10, Polad e-mail: rziel@impagovpl ABSTRACT Weak laws of large umbers (W LLN), strog

More information

Lecture 11 October 27

Lecture 11 October 27 STATS 300A: Theory of Statistics Fall 205 Lecture October 27 Lecturer: Lester Mackey Scribe: Viswajith Veugopal, Vivek Bagaria, Steve Yadlowsky Warig: These otes may cotai factual ad/or typographic errors..

More information

Last Lecture. Wald Test

Last Lecture. Wald Test Last Lecture Biostatistics 602 - Statistical Iferece Lecture 22 Hyu Mi Kag April 9th, 2013 Is the exact distributio of LRT statistic typically easy to obtai? How about its asymptotic distributio? For testig

More information

Solutions: Homework 3

Solutions: Homework 3 Solutios: Homework 3 Suppose that the radom variables Y,...,Y satisfy Y i = x i + " i : i =,..., IID where x,...,x R are fixed values ad ",...," Normal(0, )with R + kow. Fid ˆ = MLE( ). IND Solutio: Observe

More information

f n (x) f m (x) < ɛ/3 for all x A. By continuity of f n and f m we can find δ > 0 such that d(x, x 0 ) < δ implies that

f n (x) f m (x) < ɛ/3 for all x A. By continuity of f n and f m we can find δ > 0 such that d(x, x 0 ) < δ implies that Lecture 15 We have see that a sequece of cotiuous fuctios which is uiformly coverget produces a limit fuctio which is also cotiuous. We shall stregthe this result ow. Theorem 1 Let f : X R or (C) be a

More information

Rates of Convergence by Moduli of Continuity

Rates of Convergence by Moduli of Continuity Rates of Covergece by Moduli of Cotiuity Joh Duchi: Notes for Statistics 300b March, 017 1 Itroductio I this ote, we give a presetatio showig the importace, ad relatioship betwee, the modulis of cotiuity

More information

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where

More information

Lecture 10: Universal coding and prediction

Lecture 10: Universal coding and prediction 0-704: Iformatio Processig ad Learig Sprig 0 Lecture 0: Uiversal codig ad predictio Lecturer: Aarti Sigh Scribes: Georg M. Goerg Disclaimer: These otes have ot bee subjected to the usual scrutiy reserved

More information

REAL ANALYSIS II: PROBLEM SET 1 - SOLUTIONS

REAL ANALYSIS II: PROBLEM SET 1 - SOLUTIONS REAL ANALYSIS II: PROBLEM SET 1 - SOLUTIONS 18th Feb, 016 Defiitio (Lipschitz fuctio). A fuctio f : R R is said to be Lipschitz if there exists a positive real umber c such that for ay x, y i the domai

More information

lim za n n = z lim a n n.

lim za n n = z lim a n n. Lecture 6 Sequeces ad Series Defiitio 1 By a sequece i a set A, we mea a mappig f : N A. It is customary to deote a sequece f by {s } where, s := f(). A sequece {z } of (complex) umbers is said to be coverget

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

Lecture 3: MLE and Regression

Lecture 3: MLE and Regression STAT/Q SCI 403: Itroductio to Resamplig Methods Sprig 207 Istructor: Ye-Chi Che Lecture 3: MLE ad Regressio 3. Parameters ad Distributios Some distributios are idexed by their uderlyig parameters. Thus,

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Notes 5 : More on the a.s. convergence of sums

Notes 5 : More on the a.s. convergence of sums Notes 5 : More o the a.s. covergece of sums Math 733-734: Theory of Probability Lecturer: Sebastie Roch Refereces: Dur0, Sectios.5; Wil9, Sectio 4.7, Shi96, Sectio IV.4, Dur0, Sectio.. Radom series. Three-series

More information

ECE 901 Lecture 13: Maximum Likelihood Estimation

ECE 901 Lecture 13: Maximum Likelihood Estimation ECE 90 Lecture 3: Maximum Likelihood Estimatio R. Nowak 5/7/009 The focus of this lecture is to cosider aother approach to learig based o maximum likelihood estimatio. Ulike earlier approaches cosidered

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation ECE 645: Estimatio Theory Sprig 2015 Istructor: Prof. Staley H. Cha Maximum Likelihood Estimatio (LaTeX prepared by Shaobo Fag) April 14, 2015 This lecture ote is based o ECE 645(Sprig 2015) by Prof. Staley

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

MATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED

MATH 472 / SPRING 2013 ASSIGNMENT 2: DUE FEBRUARY 4 FINALIZED MATH 47 / SPRING 013 ASSIGNMENT : DUE FEBRUARY 4 FINALIZED Please iclude a cover sheet that provides a complete setece aswer to each the followig three questios: (a) I your opiio, what were the mai ideas

More information

Detailed proofs of Propositions 3.1 and 3.2

Detailed proofs of Propositions 3.1 and 3.2 Detailed proofs of Propositios 3. ad 3. Proof of Propositio 3. NB: itegratio sets are geerally omitted for itegrals defied over a uit hypercube [0, s with ay s d. We first give four lemmas. The proof of

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory ST525: Advaced Statistical Theory Departmet of Statistics & Applied Probability Tuesday, September 7, 2 ST525: Advaced Statistical Theory Lecture : The law of large umbers The Law of Large Numbers The

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit theorems Throughout this sectio we will assume a probability space (Ω, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Lecture 6 Ecient estimators. Rao-Cramer bound.

Lecture 6 Ecient estimators. Rao-Cramer bound. Lecture 6 Eciet estimators. Rao-Cramer boud. 1 MSE ad Suciecy Let X (X 1,..., X) be a radom sample from distributio f θ. Let θ ˆ δ(x) be a estimator of θ. Let T (X) be a suciet statistic for θ. As we have

More information

Machine Learning Brett Bernstein

Machine Learning Brett Bernstein Machie Learig Brett Berstei Week Lecture: Cocept Check Exercises Starred problems are optioal. Statistical Learig Theory. Suppose A = Y = R ad X is some other set. Furthermore, assume P X Y is a discrete

More information

Empirical Process Theory and Oracle Inequalities

Empirical Process Theory and Oracle Inequalities Stat 928: Statistical Learig Theory Lecture: 10 Empirical Process Theory ad Oracle Iequalities Istructor: Sham Kakade 1 Risk vs Risk See Lecture 0 for a discussio o termiology. 2 The Uio Boud / Boferoi

More information

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >

More information

Asymptotic Results for the Linear Regression Model

Asymptotic Results for the Linear Regression Model Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/5.070J Fall 203 Lecture 3 9//203 Large deviatios Theory. Cramér s Theorem Cotet.. Cramér s Theorem. 2. Rate fuctio ad properties. 3. Chage of measure techique.

More information

Advanced Stochastic Processes.

Advanced Stochastic Processes. Advaced Stochastic Processes. David Gamarik LECTURE 2 Radom variables ad measurable fuctios. Strog Law of Large Numbers (SLLN). Scary stuff cotiued... Outlie of Lecture Radom variables ad measurable fuctios.

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

Application to Random Graphs

Application to Random Graphs A Applicatio to Radom Graphs Brachig processes have a umber of iterestig ad importat applicatios. We shall cosider oe of the most famous of them, the Erdős-Réyi radom graph theory. 1 Defiitio A.1. Let

More information

Notes 19 : Martingale CLT

Notes 19 : Martingale CLT Notes 9 : Martigale CLT Math 733-734: Theory of Probability Lecturer: Sebastie Roch Refereces: [Bil95, Chapter 35], [Roc, Chapter 3]. Sice we have ot ecoutered weak covergece i some time, we first recall

More information

Questions and Answers on Maximum Likelihood

Questions and Answers on Maximum Likelihood Questios ad Aswers o Maximum Likelihood L. Magee Fall, 2008 1. Give: a observatio-specific log likelihood fuctio l i (θ) = l f(y i x i, θ) the log likelihood fuctio l(θ y, X) = l i(θ) a data set (x i,

More information

Product measures, Tonelli s and Fubini s theorems For use in MAT3400/4400, autumn 2014 Nadia S. Larsen. Version of 13 October 2014.

Product measures, Tonelli s and Fubini s theorems For use in MAT3400/4400, autumn 2014 Nadia S. Larsen. Version of 13 October 2014. Product measures, Toelli s ad Fubii s theorems For use i MAT3400/4400, autum 2014 Nadia S. Larse Versio of 13 October 2014. 1. Costructio of the product measure The purpose of these otes is to preset the

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013 Large Deviatios for i.i.d. Radom Variables Cotet. Cheroff boud usig expoetial momet geeratig fuctios. Properties of a momet

More information

2.1. Convergence in distribution and characteristic functions.

2.1. Convergence in distribution and characteristic functions. 3 Chapter 2. Cetral Limit Theorem. Cetral limit theorem, or DeMoivre-Laplace Theorem, which also implies the wea law of large umbers, is the most importat theorem i probability theory ad statistics. For

More information

STAT Homework 1 - Solutions

STAT Homework 1 - Solutions STAT-36700 Homework 1 - Solutios Fall 018 September 11, 018 This cotais solutios for Homework 1. Please ote that we have icluded several additioal commets ad approaches to the problems to give you better

More information

Study the bias (due to the nite dimensional approximation) and variance of the estimators

Study the bias (due to the nite dimensional approximation) and variance of the estimators 2 Series Methods 2. Geeral Approach A model has parameters (; ) where is ite-dimesioal ad is oparametric. (Sometimes, there is o :) We will focus o regressio. The fuctio is approximated by a series a ite

More information

62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 +

62. Power series Definition 16. (Power series) Given a sequence {c n }, the series. c n x n = c 0 + c 1 x + c 2 x 2 + c 3 x 3 + 62. Power series Defiitio 16. (Power series) Give a sequece {c }, the series c x = c 0 + c 1 x + c 2 x 2 + c 3 x 3 + is called a power series i the variable x. The umbers c are called the coefficiets of

More information

Lecture 3 : Random variables and their distributions

Lecture 3 : Random variables and their distributions Lecture 3 : Radom variables ad their distributios 3.1 Radom variables Let (Ω, F) ad (S, S) be two measurable spaces. A map X : Ω S is measurable or a radom variable (deoted r.v.) if X 1 (A) {ω : X(ω) A}

More information

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley Notes O Media ad Quatile Regressio James L. Powell Departmet of Ecoomics Uiversity of Califoria, Berkeley Coditioal Media Restrictios ad Least Absolute Deviatios It is well-kow that the expected value

More information

University of Colorado Denver Dept. Math. & Stat. Sciences Applied Analysis Preliminary Exam 13 January 2012, 10:00 am 2:00 pm. Good luck!

University of Colorado Denver Dept. Math. & Stat. Sciences Applied Analysis Preliminary Exam 13 January 2012, 10:00 am 2:00 pm. Good luck! Uiversity of Colorado Dever Dept. Math. & Stat. Scieces Applied Aalysis Prelimiary Exam 13 Jauary 01, 10:00 am :00 pm Name: The proctor will let you read the followig coditios before the exam begis, ad

More information

MA Advanced Econometrics: Properties of Least Squares Estimators

MA Advanced Econometrics: Properties of Least Squares Estimators MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information