University of Lausanne - École des HEC LECTURE NOTES ADVANCED ECONOMETRICS. Preliminary version, do not quote, cite and reproduce without permission

Size: px
Start display at page:

Download "University of Lausanne - École des HEC LECTURE NOTES ADVANCED ECONOMETRICS. Preliminary version, do not quote, cite and reproduce without permission"

Transcription

1 Uiversity of Lausae - École des HEC LECTURE NOTES ADVANCED ECONOMETRICS Prelimiary versio, do ot quote, cite ad reproduce without permissio Professor : Floria Pelgri CHAPTER 4: THE CONSTRAINED ORDINARY LEAST SQUARES ESTIMATOR

2 Cotets 1 Itroductio Represetatio of costraits Kerel represetatio of a set of liear costraits Image represetatio of a set of liear costraits Examples The costraied multiple liear model ad idetificatio The costraied ordiary least squares estimator Estimatio of β with the kerel represetatio Estimatio of β with the image represetatio Geometrical iterpretatio Projectig oto a costraied subspace Decompositio of the adjusted depedet variable Decompositio of the fitted error terms Statistical properties Ubiasedess properties Efficiecy Parametric model Semi-parametric model Costraied versus ucostraied models Asymptotic theory Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test The Studet test Fisher test Geometrical iterpretatio Heuristic derivatio A more formal derivatio Other derivatios of the Fisher test statistic Implemetatio of the Fisher test Applicatio: Joit sigificace of the explaatory variables, with the exceptio of the costat term Asymptotic tests Ituitio Framework Liear costraits No liear costraits

3 CONTENTS The Wald test Liear costraits Noliear costraits Discussio about the Wald test The likelihood ratio test Liear costraits Noliear costraits Discussio The Lagrage multiplier or score test Liear costraits Noliar costraits Discussio Applicatios Asymptotic tests i the simple liear regressio model k = The liear model with kow variace The liear model with ukow variace Compariso of tests

4 1 Itroductio 3 1 Itroductio The purpose of this chapter is twofold. First, we explai how estimatio ca be coducted i the presece of liear ad/or o liear costraits, I this respect, the costraied ordiary least squares estimator ad the costraied maximum likelihood estimator are preseted i the case of the multiple liear regressio model ucostraied model. Secod, we explai how testig procedures ca be coducted. We cosider the Fisher test ad the trilogy of asymptotic tests Wald test, Lagrage multiplier or score test, likelihood ratio test. For istace, suppose that a applied ecoomist yes, it exists! would like to estimate a productio fuctio usig a Cobb-Douglas specificatio: Y i = AK α i L β i expu i where Y i, K i, L i ad u i are respectively the added value productio of firm sector i, the stock of capital, the labor factor, ad a multiplicative error term. Usig a log-trasformatio, this model ca be writte as: y i = β 0 + β 1 k i + β 2 l i + u i where β 0 = loga, β 1 = α, β 2 = β k t = logk i, y t = logy i, ad l i = logl t I doig so, oe seeks to impose the assumptio of costat returs to scale: β 1 + β 2 = 1. Two methods ca be cosidered: 1. Estimate the ucostraied model without takig ito accout the costrait ad the test the ull hypothesis H 0 : β 1 + β 2 = 1. If the ull hypothesis is ot rejected at a certai cofidece level, say 5%, the the costrait must be imposed ad the costraied model must be estimated. O the other had, if the ull hypothesis is rejected, the the ucostraied model is estimated with least squares method or the maximum likelihood method Chapters 2 ad Estimate the costraied model ad test whether the costraied model is preferable to the ucostraied model for istace, usig a Fisher test. It is worth oticig that the costraied model is give by: i.e. y i = β β 2 l i + β 2 k i + u i y i l i = β 0 + β 2 k i l i + u i where y i l i ad k i l i are the product added value per workers ad the capital per workers, respectively. Therefore, the costraied ordiary least

5 1 Itroductio 4 squares estimator of β 2 is defied to be the solutio of the followig miimizatio program: ˆβ 2,COLS = argmi y i l i β 0 β 2 k i l i 2. β 2 i=1 O the other had, the costraied ordiary least squares estimator of β 0, β 1, β 2 ca also be defied as the solutio of the followig costraied miimizatio problem: ˆβ0,COLS, ˆβ 1,COLS, ˆβ 2,COLS = argmi β 0,β 1,β 2 y i β 0 β 1 k i β 2 l i 2 i=1 s.t. β 1 + β 2 = 1. These two strategies have importat cosequeces. If oe wrogly imposes the costrait, the the ordiary least squares estimator of the costraied model the costraied ordiary least squares estimator is biased ad is ot cosistet. O the other had, if oe truly imposes the costrait, the ordiary least squares estimator of the ucostraied model should be expected to be close to the costraied ordiary least squares estimator. O the other had, costraits ca be used i order to compare ested models. For istace, istead of estimatig a Cobb-Douglas specificatio, oe may assume a Traslog specificatio as follows: 1 1 y i = β 0 + β 1 k i + β 2 l i + β 3 2 log2 K i + β 4 2 log2 L i + β 5 log K i log L i + v i. I particular, the Cobb-Douglas specificatio is ested i the previous specificatio. More specifically, if the ull hypothesis is ot rejected: β 3 = 0, β 4 = 0 ad β 5 = 0 the oe eds up with the Cobb-Douglas specificatio. I cotrast, suppose that the two models are ot ested at least, they have differet explaatory variables. The first model is give by: whereas the secod model writes: y = X 1 β 1 + u 1 1 y = X 2 β 2 + u 2 2 where X 1 M,k1, β 1 M k1,1, X 2 M,k2, ad β 2 M k2,1. If oe seeks to coduct the followig test o-ested assumptios: H 0 : M 0 is true H a : M 1 is true

6 1 Itroductio 5 the the supra-model might be cosidered: y = X 1 b 1 + X 2 b 2 + ε ad a Fisher test might be coducted. To sum up, we cosider estimatig a liear model: = Xβ + u where Y R, X is a matrix of costat ad radom variables, Eu X = 0 1, ad V u X = σ 2 I, With a set of liear costraits: with R M p k ad q M p 1 ; Rβ = q With a set of oliear costraits: h β = 0 p 1 with h β = h 1 β. h p β where h 1,h 2,,h p take their values i R, are differetiable. But we also cosider testig: A set of p liear costraits stated i the ull hypothesis: H 0 : Rβ = q agaist the alterative hypothesis: H a : Rβ q A set of p oliear costraits stated i the ull hypothesis: H 0 : hβ = 0 p 1 agaist the alterative hypothesis: H a : hβ 0 p 1.

7 2 Represetatio of costraits 6 2 Represetatio of costraits The costraits ca be writte i two differet ways: the kerel represetatio ad the image represetatio. Both forms are equivalet. 2.1 Kerel represetatio of a set of liear costraits Defiitio 1. The represetatio of the liear costraits: is called the kerel represetatio. Rβ = q 2.2 Image represetatio of a set of liear costraits Defiitio 2. The kerel represetatio Rβ = q is fully equivalet to:! β R k p / β = β 0 + S β. with β 0 R k, β R k p ad S M k k p. This represetatio is called the image represetatio. 2.3 Examples 1. Cosider the costraied liear multiple regressio model: s.t. Y = X 1 β 1 + X 2 β 2 + X 3 β 3 + X 4 β 4 + X 5 β 5 + ɛ β 2 = 0 β 4 = 0. a The kerel represetatio is give by: β β β 3 0 =. }{{} β 4 0 }{{} R β q 5 }{{} β

8 2 Represetatio of costraits 7 b The image represetatio writes: β 1 0 β 2 β = β 3 β 4 = β 5 0 }{{} β } {{ } S β 1 β 3 β 5 } {{ } β The costraied parameters are β 2 ad β 4 whereas the ucostraied free parameters are β 1, β 3, ad β Cosider the followig costraits: β 2 = β 3 β 3 = β 4 a The kerel represetatio is give by: }{{} R b The image represetatio writes: β 1 β 2 β = β 3 β 4 = β 5 3. Cosider the costraits: β 1 β 2 β 3 β 4 β 5 } {{ } β 0 =. 0 }{{} q β 1 β 3 β 5 β 1 = β 2 = = β K a The kerel represetatio is give by: β 1 β 2 β 3. β k 1 β k = 0 k 1 1

9 3 The costraied multiple liear model ad idetificatio 8 4. Cosider: Therefore, Rβ = q with R M k 1 k, β R k ad q M k 1 1. k-1 idepedet costraits β 3 + β 4 + β 5 = 1 a The kerel represetatio is give by: β = 1. b The image represetatio writes: β 1 β 1 β 2 β = β 3 β 4 = β 2 1 β 4 β 5 β 4 β 5 β = β 0 + S β The costraied multiple liear model ad idetificatio β 1 β 2 β 4 β 5 Defiitio 3. Let Y R, X M,k be a matrix of costat ad radom regressors, R M p,k ad q M p,1. The costraied liear multiple regressio model is defied to be: Y = Xβ + u s.t. Rβ = q with Eu X = 0 1, V u X = σ 2 I, rkx = k ad rkr = p or PrkX = k = 1 with p k.

10 4 The costraied ordiary least squares estimator 9 Remarks : 1. Throughout the chapter, we assume that rkx = k or PrkX = k = 1. I other words, the ucostraied model is idetifiable. 2. Suppose that the ucostraied mutiple liear regressio model is ot idetifiable rkx = k 1 < k. I particular, there is o uique solutio from the ormal equatios. Let β deote a particular solutio of: X X β = X Y where X X is sigular. The set of solutios is give by: β + KerX = { ˆβ R k / ˆβ = β + x, x KerX} where dim KerX = k k 1 ad KerX is the kerel of X matrix. Oe way to circumvet this issue is to impose some idetifyig costraits: Rβ = q where rkr = k k 1. Such idetifyig costraits are ofte used i practise for istace, the ANOVA, CANOVA models, the two-way fixed effects models, etc. Defiitio 4. The multiple liear costraied model is idetifiable if ad oly if β 1, β 2 R k : Xβ 1 = Xβ 2 et Rβ 1 = Rβ 2 β 1 = β 2. 4 The costraied ordiary least squares estimator 4.1 Estimatio of β with the kerel represetatio Defiitio 5. Uder suitable regularity coditios, the costraied ordiary last squares estimator of β, deoted ˆβ COLS, is the solutio of the followig miimizatio program: mi Y Xβ 2 I β s.t. Rβ = q.

11 4 The costraied ordiary least squares estimator 10 Propositio 1. Uder suitable regularity coditios, the costraied ordiary least squares estimator of β is give by: ˆβ COLS = ˆβ OLS X X 1 R [RX X 1 R ] 1 [R ˆβ OLS q] Proof: Uder suitable regularity coditios, the Lagragia is defied to be: LY, X; β = Y Xβ 2 + λ Rβ q = Y Y + β X Xβ Y Xβ β X Y + λ Rβ q where λ M p,1 is the vector of Lagrage multipliers. The first-order coditios with respect to β ad λ are give as follows: L β. = 0 k 1 2X Y X ˆβ COLS + R ˆλ = 0k 1 3 L λ. = 0 p 1 R ˆβ COLS q = 0 p 1 4 Takig Eq. 3, we get: ad thus: ˆβ COLS = X X 1 X Y 1 2 X X 1 R ˆλ ˆβ COLS = ˆβ OLS 1 2 X X 1 R ˆλ. Replacig this expressio i Eq. 4 yields: ad thus: R ˆβ COLS = q R ˆβ OLS 1 2 RX X 1 R ˆλ = q ˆλ = 2[RX X 1 R ] 1 [R ˆβ OLS q] Therefore, takig the expressio of λ ad replacig it i the expressio of ˆβ COLS, we obtai: Remarks: ˆβ COLS = ˆβ OLS X X 1 R [RX X 1 R ] 1 [R ˆβ OLS q] 1. The derivatio of ˆβ COLS makes use of the derivatio of ˆβ OLS the rak coditio of X is thus ecessary. 2. The matrix RX X 1 R is ivertible. Ideed, sice X X 1 is positive defiite ad R is of full row rak, it follows that RX X 1 R is o sigular ad thus ivertible. 3. Secod-order coditios ad qualificatio of costraits hold.

12 4 The costraied ordiary least squares estimator Estimatio of β with the image represetatio Defiitio 6. Uder suitable regularity coditios, the costraied ordiary least squares estimator of β, deoted ˆβ COLS, is defied to be: ˆβ COLS = β 0 + S ˆ β It is strictly equivalet to: = β 0 + S S X XS 1 S X Y Xβ 0. ˆβ COLS = ˆβ OLS X X 1 R [RX X 1 R ] 1 [R ˆβ OLS q]. Proof: The set of liear costraits ca be writte as: Rβ = q! β R k p / β = β 0 + S β. The costraied liear multiple regressio model is the defied by: or Y = Xβ 0 + S β + u Y Xβ 0 = XS β + u The estimator of β is the obtaied as the ordiary least squares estimator of the trasformed model i.e., the projectio of Y Xβ 0 oto XS. Uder suitable regularity coditios, ˆ β is the solutio of the followig miimizatio program: ˆ β = argmi Ỹ X β s 2 I β The ordiary least squares estimator of β is the: ˆ β = X sx s 1 X sỹ = XS XS 1 XS Ỹ = XS XS 1 XS Y Xβ 0 ˆ β = S X XS 1 S X Y Xβ 0. ad ˆβ COLS = β 0 + S ˆ β = β 0 + S S X XS 1 S X Y Xβ 0.

13 4 The costraied ordiary least squares estimator 12 Example : Cosider the multiple liear regressio model the exogeeity assumptio holds ad the error terms are spherical: Y i = β 1 + β 2 x i,2 + β 3 x i, β k x i,k + u i i.e. Y = Xβ + u where X = e, X 2,..., X k ad β R k. Cosider the simple costrait: β 2 = 1. The costraied ordiary least squares estimator of β ca be writte as: ˆβ COLS = β 0 + ˆ βols with β 0 = ad ˆ βols is the ordiary least squares estimator of the followig model: i.e. Y i x i2 = β 1 + β 3 x i3 + β 4 x i β k x ik + u i Y X 2 = XS β + ε where: ad M k,k x 11 x x 1k x XS = 21 x x 1 x 3... x k

14 5 Geometrical iterpretatio 13 5 Geometrical iterpretatio 5.1 Projectig oto a costraied subspace... Let L deoted the subspace spaed by the colum vectors X 1, X 2,..., X k. Let L 0 ad E 0 deote the followig subspaces: L 0 = {Xβ/β R k ad Rβ = q} = {Xβ 0 + XS β/β R k p } give that β = β 0 + S β. E 0 = {XS β/ β R k p } [Isert Figure 1 aroud here] 5.2 Decompositio of the adjusted depedet variable The fitted estimated depedet variable ŶCOLS is defied by: sice ad Ŷ COLS = X ˆβ COLS = X β 0 + Sˆ β = Xβ 0 + P XS Y Xβ 0 with P XS = XS S X XS 1 S X = P XS Y + I P XS Xβ 0 = P XS P X P X Y + I P XS Y 0 with Y 0 = Xβ 0 = P XS Ŷ OLS + I P XS Y 0 P XS P X = XSS X XS 1 S X X X X 1 X ŶCOLS is the result of both: = XSS X XS 1 S X = P XS P X Y = X ˆβ OLS = ŶOLS. The orthogoal projectio of Y usig the projectio matrix P XS P XS Y

15 5 Geometrical iterpretatio 14 The traslatio give by We get: I P XS Xβ 0. û COLS 2 I > û OLS 2 I. I other words, the fit of the restricted least squares solutio caot be better tha that of the urestricted solutio. 5.3 Decompositio of the fitted error terms Propositio 2. The fitted vector of residuals obtaied from the costraied ordiary least squares estimator ca be decomposed as the sum of two orthogoal vectors: û COLS = û OLS + ˆv where ˆv = X ˆβOLS ˆβ COLS = XX t X 1 R t [ RX t X 1 R t] 1 RX t X 1 X t u = P XX t X 1 R tu Proof: Usig the geometrical represetatio, we get: û COLS = û OLS + ˆv where ˆv = X ˆβOLS ˆβ COLS. Usig the defiitio of the costraied ordiary least squares estimator, oe obtais: { ˆv = X X t X R [ t RX t X 1 R t] } 1 R ˆβOLS q with: R ˆβ OLS q = RX t X 1 X t Y q = RX t X 1 X t Xβ + u q = Rβ q + RX t X 1 X t u = RX t X 1 X t u.

16 6 Statistical properties 15 The result follows. Remark: We get û COLS = Y ŶCOLS = Y X ˆβ COLS = Y X ˆβCOLS ˆβ OLS + ˆβ OLS = [Y X ˆβ ] [ OLS + X ˆβOLS ˆβ ] COLS [ = [u OLS ] + X ˆβOLS ˆβ ] COLS with û OLS ad X ˆβOLS ˆβ COLS are idepedet i probability show this result!. 6 Statistical properties 6.1 Ubiasedess properties Propositio 3. If the model is correctly specified β satisfies the set of liear costraits Rβ = q ad the exogeeity assumptio holds, the the costraied ordiary least squares estimator of β, ˆβ COLS, is ubiased: [ ] E ˆβCOLS = β for all β Θ. Proof: Takig the coditioal expectatio with respect to X of the costraied ordiary least squares estimator of β, we get: [ ] [ ] E ˆβCOLS X = E ˆβOLS X [ X X 1 R [RX X 1 R ] 1 E R ˆβ ] OLS q X. We have: E [ R ˆβ ] OLS q X = RX t X 1 X t E [u X] + Rβ q = 0. The first right-had side term equals zero due to the exogeeity assumptio whereas the secod term equals zero[ sice the] model is correctly specified ad the ull assumptio holds. Sice E ˆβOLS X = β, we get [ ] E ˆβCOLS X = β.

17 6 Statistical properties 16 Propositio 4. If the model is icorrectly specified β does ot satisfy the set of liear costraits Rβ = q, the costraied ordiary least squares estimator of β is biased. Proof: I this case, [ E R ˆβ ] OLS q X 0 p 1 ad the result follows. Propositio 5. The variace-covariace matrix of the costraied ordiary least squares estimator of β is defied to be: V ˆβCOLS X = σ 2 X X [ 1 I k R [RX X 1 R ] 1 RX X 1] ad V ˆβCOLS X V ˆβOLS X Proof: The variace-covariace matrix of the costraied ordiary least squares estimator is defied to be: V ˆβCOLS = σ 2 X X [ 1 I R [RX X 1 R ] 1 RX X 1] ad Give that: V ˆβCOLS < V ˆβOLS. V ˆβCOLS = E ˆβ COLS β ˆβ COLS β where: ˆβ COLS β = ˆβ OLS β X X 1 R [RX X 1 R ] 1 [R ˆβ OLS β + Rβ q] = I X X 1 R [RX X 1 R ] 1 R ˆβOLS β X X 1 R [RX X 1 R ] 1 [Rβ q]

18 6 Statistical properties 17 the secod right-had side term is ostochastic ad this does ot eter i the determiatio of the variace-covariace matrix of the costraied ordiary least squares estimator. Therefore, V ˆβCOLS = σ 2 [ I X X 1 R [RX X 1 R ] 1 R ] X X 1 [ I X X 1 R [RX X 1 R ] 1 R ] = σ 2 { X X 1 X X 1 R [RX X 1 R ] 1 RX X 1 X X 1 R [RX X 1 R ] 1 RX X 1 +X X 1 R [RX X 1 R ] 1 RX X 1 R[RX X 1 R ] 1 RX X 1} = σ { 2 X X 1 X X 1 R [RX X 1 R ] 1 RX X 1} The matrix X X 1 R [RX X 1 R ] 1 RX X 1 is positive semi-defiite ad thus: Remarks: V ˆβ COLS V ˆβ OLS 1. The costraied ordiary least squares estimator might be biased, especially if the model is icorrectly specified. 2. The costrait ordiary least squares estimator is more precise tha the ordiary least squares estimator: V ˆβCOLS X V ˆβOLS X 3. There is a trade-off betwee the bias ad the efficiecy: itroducig more costraits i the model improves the precisio of the estimates but the correspodig estimates are more likely biased! The coverse is also true. Propositio 6. A ubiased estimator of σ 2 is defied to be: ˆσ 2 COLS = û COLS 2 I k + p Proof: Takig the orthogoal decompositio of û COLS, we get: û COLS 2 I = û OLS 2 I + ˆv 2 I.

19 6 Statistical properties 18 Therefore, with: Fially, E [ û COLS 2 I X ] = E [ û OLS 2 I X ] + E [ ˆv 2 I X ] = σ 2 k + T rp XX t X 1 R t T r P XX t X 1 R t = p. E [ û COLS 2 I X ] = σ 2 k + p. The result follows. Remark: I cotrast to Chapter 2, the umber of degrees of freedom is k p sice the umber of free parameters is k p ad ot k. 6.2 Efficiecy Parametric model Propositio 7. The costraied ordiary least squares estimator of β, ˆβ COLS, is the maximum likelihood estimator of β i the costraied liear multiple regressio model. Exercise: Show the previous propositio. If the model is correctly specified uder the ull assumptio, the the costraied ordiary least squares estimator is efficiet its variace-covariace matrix equals the FDCR lower boud. Semi-parametric model Propositio 8. Cosider the coditioal static multiple liear regressio model: Y = Xβ 0 + u s.t.rβ = q where Eu i X = 0 ad V u i X = σ 2 for all i, Y is a -dimesioal vector, X is a k matrix of rak k or PrkX = k = 1, R M p k is of full row rak. The costraied ordiary least squares estimator of β 0 defied by: ˆβ COLS = ˆβ OLS X X 1 R [RX X 1 R ] 1 [R ˆβ OLS q] is the best estimator i the class of liear i Y ubiased estimators of β 0 uder the ull hypothesis H 0 : Rβ = q.

20 6 Statistical properties 19 Proof: This is a applicatio of Chapter 2 part II. 6.3 Costraied versus ucostraied models There are trade-offs betwee costraied ad ucostraied models. specifically, More If the costraied model is preferred, the costraied estimator will be more precise tha the ucostraied estimator as for istace the ordiary least squares estimator. At the same time, if oe wrogly imposes the costraits, the costraied estimator will be biased. If the ucostraied model is preferred, the correspodig estimator will be ubiased. At the same time, the estimator will be less precise tha the costraied estimator if the costraits are true i.e. oe wrogly assume that there are o costraits. Note also that multicolliearity is more likely to occur i the ucostraied model tha i the costraied model. All i all, there is "mea-variace" trade-off. It is thus difficult to coclude whether the first or secod strategy is preferable. To further illustrate this trade-off, cosider the followig two examples. Example 1: Omittig oe explaatory variable Omittig a explaatory variable,which is correlated to other regressors, geerally yields biased estimates. I particular, the exogeeity assumptio does ot hold aymore. To see this, suppose that the true data geeratig process is defied to be: Y = X 1 β 1 + X 2 β 2 + u. Suppose that oe wrogly imposes the costrait β 2 = 0, i.e. that oe estimates the followig model: Y = X 1 β 1 + u. Cosequetly, the costraied estimator of β = β 1, β 2 is give by: β 1,COLS = X 1X 1 1 X 1Y β 2,COLS = 0. Sice the true data geeratig process is the ucostraied model, oe gets that: E β1,cols E β COLS = 0

21 6 Statistical properties 20 where: E β 1,COLS = E [ X 1X 1 1 X 1X 1 β 1 + X 2 β 2 + u ] = X 1X 1 1 X 1X 1 β 1 + X 1X 1 1 X 1X 2 β 2 = β 1 + X 1X 1 1 X 1X 2 β 2. I other words, omittig X 2 yields the followig expressio for the bias of the costraied ordiary least squares estimator: X Bias β OCOLS = 1 X 1 1 X 1X 2 β 2 β 2. Therefore, if the costrait does ot hold i the true data geeratig process, the: 1. The bias of the costraied ordiary least squares estimator of β 1 is X 1X 1 1 X 1X 2 β 2. This bias is proportioal to β The bias of the costraied ordiary least squares estimator of β 2 is exactly mius the true ukow value of β The bias of β 1,COLS is higher whe the vector X 2 is close to the hyperplae spaed by the colum vectors of the X matrix for istace, whe the two explaatory variables are highly correlated. 4. If X 1 ad X 2 are orthogoal or perpedicular, the the bias is zero sice this bias depeds o the ier product betwee these two vectors. This result ca be geeralized if X 1 ad X 2 are some distict sets of explaatory variables. Remark: I the geeral case i which the X 2 vector is ot orthogoal to the liear space spaed by the colum vectors of X 1 or the vector X 1, deoted LX 1, the bias of β 1,COLS is obtaied from the regressio of X 2 oto the colum vectors of X 1. Ideed, sice X 1 is of full colum rak, there exists a uique decompositio of X 2 such that: where γ M k 1,1 ad ε LX 1 X 2 = X 1 γ + ε I this respect, γ is the ordiary least squares estimator of γ i the auxiliary model X 2 = X 1 γ + ε, ad γ = X 1X 1 1 X 1X 2. Therefore, the bias of β 1,COLS ca be expressed as: Bias β 1,COLS = γβ 2. Cosequetly, two sources of bias ca be idetified:

22 6 Statistical properties The degree of multicoliearity betwee X 2 ad LX 1 2. The misspecificatio of the costraits. Example 2: Addig a explaatory variable that is ot preset i the true data geeratig process. Suppose that the true data geeratig process is defied to be: Y = X 1 β 1 + X 2 β 2 + u s.t. β 2 = 0. Suppose that oe estimates the ucostraied model: Y = X 1 β 1 + X 2 β 2 + u. The ucostraied estimator of β = β 1, β 2 is ubiased. I doig so, oe applies the Frisch-Waugh theorem: β 1 = X 1M X2 X 1 1 X 1M X2 Y β 2 = X 2M X1 X 2 1 X 2M X1 Y where M X2 = I N X 2X 2 X 2 2 ad M X1 = I N X 1 X 1 X 1 1 X 1. Sice M X1 X 1 = 0, it follows that: ] β1 E [ β = E E β2 = [ β1 0 ] = β The ucostraied estimator of β, β, is ubiased. The same result applied for the ucostraied estimator of σ 2. However, addig the explaatory variable X 2 icreases the variace of the estimator of β 1. Oe ca show that: V β 1 = σ 2 X 1M X2 X 1 1 = σ 2 X 1X X 2 2 X 1X 2X 2 X 1 The variace-covariace matrix of the costraied estimator is give by V β 1,COLS = σ 2 X 1X d où V β 1,COLS < V β 1 sice X 2 X 1X 2X 2 2 X 1 is positive defiite.

23 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test Asymptotic theory The large sample properties of the costraied ordiary least squares estimator of β are applicatios of Chapter II part III. I particular, we ca show that the costraied ordiary least squares estimator of β is strogly weakly cosistet if the model is correctly specified i.e, Rβ 0 = q for the true ukow parameter vector. This result ca be eouced i the case of idepedet ad idetically distributed observatios, idepedet ad heterogeeously distributed observatios, depedet ad idetically distributed observatios, ad depedet ad heterogeeously distributed observatios. Moreover, the asymptotic distributio of the re-ormalized costraied ordiary least squares estimator ca be derived i the same way. For istace, i the case of idepedet ad idetically distributed observatios, we get the followig propositio. Propositio 9. Uder suitable regularity coditios, the samplig distributio of the costraied ordiary least squares estimator of β i a semiparametric costraied liear multiple regressio model with radom regressors is: { a ˆβ COLS N β, σ2 [ EX xi x t ] [ 1 i I k R R [ E X xi x t 1 i] ] } 1 [ R R EX xi x t ] 1 i with y i = x t iβ + u i s.t. Rβ = q. where E u i x i = 0, V u i x i = σ 2, ad E X x i x t i is osigular for all i = 1,,. Proof: See series 7. 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test I this sectio, we cosider the Gaussia parametric multiple liear regressio model ad cosider simple tests oe coefficiet or multiple tests. All i all, we seek to test the ull hypothesis that H 0 : Rβ = q agaist the alterative assumptio that H a : Rβ q. It is worth oticig that other ull or alterative hypothesis could be cosidered. 7.1 The Studet test Cosider the Gaussia liear multiple regressio model: Y = Xβ + u

24 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test 23 where Y R, X M k is oradom for coveiece, ad u N 0, σ 2 I parametric model. Oe seeks to test the sigificace of oe coefficiet give a certai level, say α: H 0 : β j = 0 H a : β j 0. Usig the results of Chapter II part II, the distributio of ˆβ j is give by: ˆβ j N β j, σ 2 m jj where m jj is the j th diagoal elemet of X X 1. Uder the ull hypothesis, oe gets: ˆβ j β 0 j σ m jj = ˆβ j σ m jj N 0, 1. with βj 0 = 0. If σ 2 is kow, the a cofidece set at the α level is give by: P u 1 α/2 ˆβ j σ u α/2 = 1 α m jj where u 1 α/2 = u α/2 is the α/2-percetile of a stadardized ormal distributio. [Isert Figure 3] If σ 2 is ukow, oe makes use of the followig result: k ˆσ2 σ 2 χ2 k ad the fact that ˆβ OLS ad k ˆσ2 are idepedet i probability. The σ 2 test-statistic is the Studet-distributed: T j = ˆβ j ˆσ m jj Γ k. The procedure of the test goes as follows: 1. Defie the critical regio: T j t α 2 2. Determie the critical value t α/2 usig a Studet s table. 3. Determie the realized value of the test-statistic usig the sample: T j,obs 4. Reject H 0 : β j = 0 at the α level if T j,obs > t α/2.

25 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test Fisher test Cosider the multiple liear Gaussia model with suitable regularity coditios ad suppose that oe seeks to test: H 0 : Rβ = q costraied model H a : Rβ q ucostraied model I cotrast to the Studet test, the Fisher test is a joit test. Geometrical iterpretatio Heuristic derivatio... If the ull assumptio H 0 caot be rejected, the ucostraied estimator of β i.e., the ordiary least squares estimator should satisfy approximatively the costraits. I other words, ˆβ OLS ad ˆβ COLS must be close give a metric meaig that ŶOLS 2 I ad ŶCOLS 2 I are approximatively the same, ad thus û OLS 2 I ad û COLS 2 I are approximatively equal. I particular, [Isert Figure 2] 1. û OLS is perpedicular to ŶOLS ŶCOLS i.e., the correspodig radom vectors are idepedet i probability 2. Their ormalized squared orm are Chi-square distributed 3. Their degrees of freedom are respectively k ad k+p k = p. Therefore, the Fisher test statistic ca be defied as follows: i.e., F = ŶOLS ŶCOLS 2 /p I û OLS 2 I / k F = ûcols 2 I û OLS 2 I /p û OLS 2 I / k sice û COLS 2 I = û OLS 2 I + ŶOLS ŶCOLS 2. I

26 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test 25 Defiitio 7. The Fisher test statistic is give by: where F = SSR 0 SSR a SSR a dl a dl 0 dl a F dl 0 dl a, dl a. dl 0 is the umber of degrees of freedom uder the ull hypothesis i.e., dl 0 = k p; dl a is the umber of degrees of freedom uder the alterative hypothesis i.e., dl a = k; SSR 0 is the sum of squared estimated residuals uder the ull hypothesis; SSR a is the sum of squared estimated residuals uder the alterative hypothesis. A more formal derivatio 5, oe gets that: Usig the geometrical iterpretatio of Sectio Ŷ COLS Xβ 0 = P E0 Y Xβ 0. where P E0 = P XS. Therefore, Ŷ ŶCOLS = P X Y Xβ 0 P E0 Y Xβ 0 = P X P E0 Y Xβ 0 = P Y Xβ 0 where P = P X P E0 is the orthogoal space of E 0 i LX. By defiitio sice u N 0, σ 2 I N : Y Xβ 0 EY Xβ0 N, I N σ σ Usig Fisher theorem, oe gets: P Y Xβ 0 σ 2 χ 2 rk P, {}}{ P Y Xβ 2 0 σ degrees of freedom {}}{ eccetricity parameter δ

27 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test 26 where Therefore, I additio, rk P = dimlx dime 0 X = p. Ŷ ŶCOLS σ 2 2 χ 2 p, δ. k σ 2 χ 2 k σ 2 ad, σ 2 ad Ŷ ŶCOLS 2 are idepedet i probability. Therefore, b Y b Y COLS 2 pσ 2 bσ 2 σ 2 F p, k, δ where F p, k, δ is a geeralized Fisher distributio with eccetricity parameter δ ad degrees of freedom p ad k. The eccetricity parameter is zero sice uder the ull hypothesis EY L 0 ad Xβ COLS L 0 imply that EY Xβ MCC L 0 ad P [EY Xβ COLS ] = 0, i.e. δ = 0. The test statistic follows the a Fisher distributio with p ad k degrees of freedom: Ŷ ŶCOLS 2 H F = 0 F p, k p σ 2 Other derivatios of the Fisher test statistic Defiitio 8. Uder suitable regularity coditios, the Fisher test statistic ca also be writte as usig the kerel represetatio of the costraits: F = 1 [ R β p σ 2 OLS q RX X 1 R ] 1 R β OLS q. Remark: A equivalet test statistic is also: F = 1 βols p σ β 2 COLS X X βols β COLS. Exercise: Show these two results.

28 7 Testig the ull hypothesis H 0 : Rβ = q: the Studet ad Fisher test 27 Implemetatio of the Fisher test 1. Determie SSR 0 costraied model ad SSR a ucostraied model; 2. Determie the realizatio of the Fisher test statistic: F obs = SSR 0 SSR a /p SSR a / k where p is the umber of liearly idepedet costraits. 3. Determie the critical regio: W = {F > F α } ad especially the critical value which correspods to the 1 α percetile of a Fisher distributed radom variable F p, k. 4. Reject the ull assumptio H 0 at the level α if F obs > F α. Applicatio: Joit sigificace of the explaatory variables, with the exceptio of the costat term Cosider the followig multiple liear regressio model: Y = e β 1 + Xβ 2 + u where X M,k 1, β 2 M k 1,1, β 1 R ad u N 0, σ 2 I N, ad the followig test: H 0 : β 2 = 0 k 1 1 H a : β 2 0 k 1 1 Uder the ull assumptio, the model is give by: Y = e β 1 + u c I particular, ŶCOLS = Y e. The umber of degrees of freedom is 1. Uder the alterative assumptio, oe has: Ŷ = e β1 + X β 2 The umber of degrees of freedom is k. The Fisher test statistic is the give by: Ŷ ŶCOLS 2 /k 1 F = û 2 / k

29 8 Asymptotic tests 28 i.e. i.e. F = F = Ŷ Y e Y Ŷ 2 /k 1 2 / k Ŷ Y e 2 / Y Y e 2 Y Ŷ 2 / Y Y e 2 k k 1 i.e. F = R2 k 1 R 2 k 1 where R 2 is the coefficiet of determiatio. Sice the Fisher test statistic is a fuctio of the coefficiet of determiatio i this joit sigificace test, it follows that the ull hypothesis is rejected at stadard levels 1%, 5% or 10% if: 1. The umber of observatios is large; 2. The coefficiet of determiatio is large i.e. close to 1. 8 Asymptotic tests The Studet ad Fisher tests are a importat tool i applied ecoometrics, especially i liear models. Moreover, both tests are exact meaig that we ca characterize the distributio of the test statistic uder the assumptio of ormal error terms. However, their priciple caot be geerally exteded i the case of oliear models. I this respect, oe ca use asymptotic tests, especially the trilogy of asymptotic tests: the Wald test, the Lagrage multiplier or score test, ad the likelihood ratio test. These three tests are quite geeral ad ca be applied with liear, oliear equality iequality costraits. Moreover, they are asymptotically equivalet ad cosistet. 1 At the same time, their fiite sample properties may be differet ad their implemetatio is differet: it depeds o the privileged assumptio H 0 or H a or both. 1 A asymptotic test with critical regio W is said to be cosistet if the probability of a type-ii error is asymptotically equal to zero or the power fuctio is asymptotically equal to oe.

30 8 Asymptotic tests Ituitio Cosider the followig simple liear regressio model: y i = b + u i with u i N 0, σ 2 for all i = 1,, ad σ 2 is kow. The log-likelihood fuctio is approximatively give by: ly; b y i b 2 i=1 ad the maximum likelihood estimator of b is: ˆb Ȳ = 1 Y i. The log-likelihood fuctio evaluated at ˆb y is approximatively give by: ly; ˆb y y i ȳ 2. We cosider the followig test: i=1 i=1 [Isert Figure 1] H 0 : b = b H a : b b Uder the ull hypothesis, the costraied maximum likelihood estimator of b is there is o loger a parameter to estimate uder the ull hypothesis!: ˆb0, = b. Uder the alterative hypothesis, the ucostraied maximum likelihood estimator of b is: ˆba, = Ȳ. The three asymptotic tests ca be explaied as follows. - Wald test: Compare ˆb a, ad ˆb 0,. If the differece is small i a sese to precise the we caot reject the ull hypothesis. Note that the distributio of the ucostraied estimator is required. - Likelihood ratio test: Compare ly, ˆb a, y ad ly, ˆb 0, y. If the differece is small the we caot reject the ull hypothesis. Note that the distributio of ly, ˆb 0, y ly, ˆb 0, y is required. - Lagrage multiplier test: Compare the slope of the taget evaluated at ˆb 0, y, ly, ˆb 0, y with zero.

31 8 Asymptotic tests Framework Liear costraits I the case of liear costraits, we will cosider the followig α-size test test: H 0 : Rβ = q H a : Rβ q where R M p k has rak p the liear costraits must be liearly idepedet or o redudat ad q M p 1. No liear costraits I the case of oliear costraits, we will cosider the followig α-size test test: The ull hypothesis is equivalet to: H 0 : hβ = 0 H a : hβ 0. h 1 β = 0 h 2 β = 0. h p β = 0 The fuctios h 1,h 2,,h p take their values i R, are differetiable, ad: [ ] h t rk β β = p for all β Θ o redudat costrait idetificatio assumptio!!! 8.3 The Wald test Liear costraits Propositio 10. Suppose that rkr = p k. The, uder H 0 : Rβ 0 = q, the Wald test is defied by the critical regio: W = { y/t W y χ 2 1 αp }

32 8 Asymptotic tests 31 where the test statistic is give by: T W = [ R ˆβ t [,a q] ˆVasy R ˆβ ] 1 [,a q R ˆβ ],a q where ˆβ,a is the maximum likelihood estimator of β uder the alterative hypothesis. The Wald test has asymptotic level α ad is cosistet. A large value of the test statistic leads to the rejectio of the ull hypothesis. The asymptotic variace-covariace matrix of R ˆβ,a q is give by: ˆV asy R ˆβ,a q = ˆV asy R ˆβ,a = RV asy ˆβ,a R t. The previous propositio ca be restated as follows usig the results of Chapter II part III i the case of the multiple liear regressio model with stochastic regressors ad i.i.d observatios see also Sectio I Chapter II part III, the followig theorem has bee derived i the case of i.i.d. observatios: Theorem 1 White s theorem. Cosider the multiple liear regressio model: for i = 1,, ad β 0 R k. Suppose Y i = x iβ 0 + u i i {x i, u i} is a idepedet ad idetically distributed sequece; ii E x i u i = 0 k 1, E x i,j u i 2 < for j = 1,, k, ad V V 1/2 X u = V is positive defiite; iii E x ij 2 < for j = 1,, k, ad Q E x i x i is positive defiite. The: where D = Q 1 V Q 1 ˆβ,OLS β 0 a.d. N 0 k 1, D I additio, if there exists ˆV symmetric ad positive semi-defiite such that ˆV V p 0 k k, the: where ˆD = X X/ 1 ˆV X X/ 1. ˆD D p 0 Sice the Wald test is based o the evaluatio of the ucostraied estimator uder the ull hypthesis, we eed this result. Note that it ca be geeralized i other cases.

33 8 Asymptotic tests 32 Propositio 11. Suppose that rkr = p k. The, uder H 0 : Rβ 0 = q: i Give that T = RD R t = RQ 1 V Q 1 R t, T 1/2 R ˆβ a.d.,a q N 0, I p where X X Q p 0 k k X X Q = E ; ii The Wald statistic is defied to be: where [ T W = R ˆβ,a q ] t ˆT 1 ˆT = R ˆD X R t X = R X X [ R ˆβ,a q] a χ 2 p 1 X 1 X ˆV R t ˆV = ˆσ 2 "Pseudo-proof": Uder H 0, oe gets: R ˆβ,a = R ˆβ,a β 0 + β 0 q = R ˆβ,a β 0 + Rβ 0 q = R ˆβ,a β 0. Therefore, the limitig distributio of R ˆβ,a q is give by: R ˆβ,a β 0. Sice Chapter II, part III see before ˆβ,a β 0 a.d. N 0, D where D = Q 1 V Q 1, it follows that: R ˆβ,a β 0 a.d. N 0, RD R t.

34 8 Asymptotic tests 33 I this respect, we get: R ˆβ a.d.,a q N 0, RD R t ad T 1/2 R ˆβ a.d.,a q N 0, I p. It is worth otig that a key issue especially i fiite samples is to compute a estimate of the asymptotic variace-covariace matrix. As we explai before, there are at least three differet methods that are asymptotically equivalet which ca lead to differet fiite sample coclusios. Noliear costraits Propositio 12. Suppose that: [ ] h t rk β β = p for all β Θ. The, uder H 0 : gβ = 0 p 1, the Wald test is defied by the critical regio: W = { y/t W y χ 2 1 αp } where the test statistic is give by: T W = [ ] t h ˆβ,a [V asy h ˆβ ] 1 [ ],a h ˆβ,a. where ˆβ,a is the maximum likelihood estimator of β uder the alterative hypothesis ad V asy h ˆβ,a = h ˆβ,a V β t asy ˆβ,a ht ˆβ,a β The Wald test has asymptotic level α ad is cosistet. Remark: The test statistic ca be rewritte as follows: [ ] [ t h 1 h t ] 1 [ ] T W = h ˆβ,a ˆβ,a I β t 1 ˆβ,a ˆβ,a h ˆβ,a. β where I is the average Fisher iformatio matrix for oe observatio.

35 8 Asymptotic tests 34 Proof: Ituitio of the proof : 1 The Wald test requires the maximizatio of the ucostraied problem i.e., the maximizatio problem uder the alterative hypothesis. 3. Therefore, its asymptotic distributio ca be characterized usig the results of Chapter III. Give the asymptotic distributio of ˆβ,a, oe obtais the asymptotic distributio of h ˆβ,a. This is step 1. 2 The key idea of the Wald test is to cosider the distace betwee ˆβ,a ad ˆβ,0 see figure. I other words, the Wald test assesses to what extet there are some gais to switch from the alterative to the ull hypothesis meaig that the ucostraied estimator is evaluated uder the costraits of H 0. This meas that the asymptotic distributio of h ˆβ,a eeds to be reevaluated uder the ull hypothesis. This is step 2. 3 After evaluatig the asymptotic distributio of ˆβ,a uder the ull hypothesis, we ca deduce the quadratic form associated to the Wald test. I doig so, we eed a cosistet estimator of Fisher iformatio matrix evaluated at β 0 the true ukow parameter vector. This is step 3. Step 1 : Determiatio of the asymptotic distributio of h ˆβ,a. Usig the asymptotic distributio of the maximum likelihood estimator, we have: ˆβ,a β 0 d N 0, Iβ 0 1 ad usig a Taylor series expasio aroud β 0 or the delta method h ˆβ,a hβ 0 d N 0, hβ0 Iβ 0 1 ht β 0. β t β Step 2 : Evaluatio of the asymptotic distributio of h ˆβ,a uder H 0. Uder the ull hypothesis, we have hβ 0 = 0 p 1. 3 The Wald test does ot require the maximizatio of the costraied problem i.e. the maximizatio problem uder the ull hypothesis.

36 8 Asymptotic tests 35 I other words, the asymptotic distributio of h ˆβ,a uder H 0 ca be rewritte as: 4 or h ˆβ,a d N 0, hβ0 Iβ 0 1 ht β 0. β t β hβ 0 Iβ 0 1 ht β 0 1/2 h β t β ˆβ d N 0, I p. Step 3 : Determiatio of the quadratic form. Sice ˆβ,a is a cosistet estimator of β 0, we get: hβ 0 Iβ 0 1 ht β 0 1/2 h β t β ˆβ,a d N 0, I p. β= ˆβ,a Takig the quadratic form, we have: [ ] [ t h 1 h t ] 1 [ ] T W = h ˆβ,a ˆβ I ˆβ,a ˆβ,a h ˆβ,a. β t β where ˆβ,a is the maximum likelihood estimator of β uder the alterative hypothesis. Remark: The asymptotic behavior of the Wald test statistic ca also be derived by usig the multivariate mea value theorem. Discussio about the Wald test Oe mai advatage of the Wald test is that the derivatio of the costraied estimator the estimator uder the ull hypothesis is ot required. I that respect, the Wald test measures the extet to which the urestricted estimate of β fails to satisfy the hypothesized restrictios uder the ull hypothesis. Moreover, the Wald test statistic has the advatage with respect to the two 4 The covergece i distributio must be uderstood as follows: h ˆβ,a d,h0 N 0, hβ0 β t Iβ 0 1 ht β 0. β

37 8 Asymptotic tests 36 other asymptotic tests that it does ot rely o a strog distributioal assumptio step 1 of the proof. There are at least two shortcomigs of this testig procedure. O the oe had, it is a test agaist the ull hypothesis : it is ot really a test for a specific alterative hypothesis...i particular, this may leads to problems regardig the power of the test. O the other had, the Wald test statistic is ot ivariat to the formulatio of the restrictios especially i the oliear case!. This lack of ivariace is geerally ot shared by the two other asymptotic tests. 8.4 The likelihood ratio test Liear costraits I the case of the multiple liear regressio model, the likelihood ratio test ca be eouced as follows. Propositio 13. Suppose that rkr = p k. The, uder H 0 : Rβ 0 = q, the likelihood ratio test is defied by the critical regio: ] T LR = 2 [l ˆβ,a l ˆβ,0. with: T LR = ˆσ 2,0 R ˆβ t,a q R X X 1 1 R t R ˆβ,a q. Proof. See Sectio 9. Remarks: 1. I fact, we ca oly prove that: ] 2 [l ˆβ,a l ˆβ,0 ˆσ 2,0 R ˆβ t,a q R X X 1 R t 1 R ˆβ,a q p The secod term is the Wald statistic with ˆV = ˆσ 2,0X X/. The test statistic is Chi-squared asymptotically distributed provided ˆσ 2,0X X/ is a cosistet estimator of V.

38 8 Asymptotic tests 37 Noliear costraits Propositio 14. The likelihood ratio test is defied by the critical regio: W = { y/t LR y χ 2 1 αr } where the test statistic is give by: ] T LR = 2 [l ˆβ,a l ˆβ,0. where ˆβ,a ad ˆβ,0 are respectively the maximum likelihood estimator of β uder the alterative ad the ull hypothesis. The likelihood ratio test has asymptotic level α ad is cosistet. Proof: Ituitio of the proof : The likelihood ratio test treats symmetrically both hypothesis the ull ad the alterative hypothesis. I particular, the likelihood ratio test requires to fid the maximum likelihood estimator uder the ull ad the alterative hypothesis. I other words, oe has to solve the costraied ull hypothesis maximizatio problem ad the ucostraied alterative hypothesis maximizatio problem. At the same time, it is worth otig that the test statistic does ot require to determie ad evaluate the Fisher iformatio matrix i cotrast to the Wald test ad the Lagrage multiplier test fidig a cosistet estimator of the Fisher iformatio matrix ca be quite difficult! 1 The quatities of iterest by defiitio of the likelihood ratio test are the log-likelihood fuctios evaluated uder the ull ad the alterative hypothesis. If we wat to determie the asymptotic distributio of the test statistic, we eed to characterize their behaviors as approaches ifiity. I other words, we eed to start from a Taylor expasio this is exactly the same idea as the asymptotic distributio of the maximum likelihood estimator of the log-likelihood fuctio uder the ull hypothesis respectively, uder the alterative hypothesis. This is step 1. 2 Therefore, a asymptotic approximatio of the likelihood ratio test statistic is give by the differece betwee the two Taylor expasios i the previous step. This is step 2. 3 After simplifyig the result of step 2, a asymptotic quadratic form of the test ca be obtaied this quadratic form is Chi-squared distributed. This is step 3.

39 8 Asymptotic tests 38 Step 1 : Taylor expasio of the log-likelihood fuctio uder H 0 ad H a. The secod-order Taylor expasio of the log-likelihood fuctio aroud β = β 0 uder the ull hypothesis is give by: a l ˆβ,0 l β 0 + l β 0 ˆβ,0 β t ˆβ,0 β 0 2 l β 0 β t 2 a l β 0 + l β 0 ˆβ,0 β 0 β t 2 a l β 0 + l β 0 ˆβ,0 β 0 β t 2 ˆβ,0 β 0 β β t ] 2 l β 0 ˆβ,0 β 0 β β t [ t ˆβ,0 β 0 1 t ˆβ,0 β 0 Iβ 0 ˆβ,0 β 0. The secod-order Taylor expasio of the log-likelihood fuctio aroud β = β 0 uder the alterative hypothesis is give by: a l ˆβ,a l β 0 + l β 0 ˆβ,a β 0 t ˆβ,a β 0 Iβ 0 ˆβ,a β 0. β t 2 Step 2 : Differece of the two Taylor expasios. Takig the differece of the two previous expressios, we get l ˆβ,a l ˆβ,0 a l β 0 ˆβ,a β ˆβ t,0 t ˆβ,a β 0 Iβ 0 ˆβ,a β ˆβ,0 β 0 t I1 β 0 ˆβ,0 β 0. Therefore, a asymptotic approximatio of the likelihood ratio test statistic is give by: T LR 2 l ˆβ,a l ˆβ,0 a 2 l β 0 ˆβ,a β ˆβ t,0 t ˆβ,a β 0 Iβ 0 ˆβ,a β 0 Step 3 : Simplificatio. t + ˆβ,0 β 0 Iβ 0 ˆβ,0 β 0. Usig propositio 1 or the first order Taylor expasio, we get

40 8 Asymptotic tests 39 1 l β 0 β a I 1 β 0 ˆβ,a β 0 ad thus substitutig i the previous expressio T LR 2 l ˆβ,a l ˆβ,0 ˆβ,a ˆβ,0 a t 2 ˆβ,a β 0 Iβ 0 t ˆβ,a β 0 Iβ 0 ˆβ,a β 0 This expressio ca be writte as: a t T LR 2 ˆβ,a β 0 Iβ 0 ˆβ,a ˆβ,0 ˆβ,a β 0 ˆβ,a β 0 t Iβ 0 + ˆβ,0 ˆβ,a + ˆβ,a β 0 t Iβ 0 t + ˆβ,0 β 0 Iβ 0 ˆβ,0 β 0. ˆβ,0 ˆβ,a + ˆβ,a β 0. i.e. a T LR ˆβ,a ˆβ t,0 Iβ 0 ˆβ,a ˆβ,0. with Iβ 0 1/2 ˆβ,a ˆβ,0 2 χ 2 p. Discussio The likelihood ratio test statistic is defied as the ratio betwee the restricted likelihood fuctio the costraied estimator uder the ull hypothesis ad the urestricted likelihood fuctio or is defied by 2 times the differece betwee the restricted log-likelihood fuctio ad the urestricted log-likelihood fuctio. This fuctio the ratio of the two likelihood fuctio must be betwee zero ad oe sice both likelihood fuctios are positive by assumptio ad the restricted likelihood fuctio caot be larger tha the urestricted likelihood fuctio: a restricted optimum is ever superior to a urestricted optimum. Oe mai difficulty of the likelihood ratio test is that the costraied ad ucostraied estimators are required. I complex models, oe of these estimators might be more difficult to derive.

41 8 Asymptotic tests The Lagrage multiplier or score test Liear costraits Usig the expressio of the Lagrage multipliers i Sectio 4 estimatio with the kerel represetatio, oe ca derive the Lagrage multiplier test i the case of the multiple liear regressio model uder geeral assumptios see also Sectio 9. Propositio 15. Suppose that rkr = p k. The, uder H 0 : Rβ 0 = q: i Give that Λ = 4 RQ 1 R t 1 T RQ 1 R t 1 with T = RD R t = RQ 1 V Q 1 R t, Λ 1/2 ˆλ a.d. N 0, I p ii The Lagrage multiplier statistics is defied to be: T LM = ˆλ 1 ˆΛ ˆλ a χ 2 p where X ˆΛ X X = 4 R R t 1 1 X ˆT R R t with ˆT = R ˆD X R t X = R 1 X ˆV 1 X R t ad ˆV is computed from the costraied regressio ad is a cosistet estimator of V uder H 0. "Pseudo-proof": The estimator of λ is defied to be see Sectio 4: ˆλ = 2 RX X 1 R t R ˆβ,a q. Therefore, the limitig distributio of ˆλ is give by: 2 RX X 1 R t R ˆβ,a q or 2 RX X 1 R t R ˆβ,a β 0

42 8 Asymptotic tests 41 sice R ˆβ,a q = R ˆβ,a β 0 uder the ull hypothesis see Sectio Therefore, 2 RX X 1 R t R ˆβ,a β 0 a.d. N 0, 4 RQ 1 R t 1 T RQ 1 R t 1. Fially, where Noliar costraits Λ 1/2 ˆλ a.d. N 0, I p Λ = 4 RQ 1 R t 1 T RQ 1 R t 1. Propositio 16. The score test is defied by the critical regio: W = { y/t S y χ 2 1 αp } where the test statistic is give by: T S = 1 l ˆβ,0 β t 1 l I ˆβ,0 β l l ˆβ,0 V β t asy ˆβ,0 β ˆβ,0 ˆβ,0 where ˆβ,0 is the maximum likelihood estimator of β uder the ull hypothesis. The score test has asymptotic level α ad is cosistet. Propositio 17. The Lagrage multiplier test is defied by the critical regio: W = { y/t LM y χ 2 1 αp } where the test statistic is give by: T LM = 1 ˆλ t h ˆβ,0 1 h t I ˆβ,0 ˆβ,0 ˆλ β t. β where ˆβ,0 is the maximum likelihood estimator of β uder the ull hypothesis. The Lagrage multiplier test has asymptotic level α ad is cosistet.

43 8 Asymptotic tests 42 Remark: The score ad Lagrage multiplier test statistics are equal: Proof: T LM = T S. Ituitio of the proof : The Lagrage multiplier test oly requires the maximizatio of the costraied problem i.e., the maximizatio problem uder the ull hypothesis. 5. The key idea of the Lagrage multiplier test is to cosider the distace betwee the taget at ˆβ,a ad ˆβ,0 see figure or the distace betwee the score vector evaluated at ˆβ,a ad ˆβ,0. I other words, this test assesses to what extet there are some gais to switch from the ull to the alterative hypothesis meaig that we expect the score or the taget evaluated at the ucostraied estimator to be close to the ull vector or zero. Therefore, the test statistic requires to determie the distributio of the Lagrage multiplier vector of the costraied maximizatio problem or to fid a quadratic form with respect to the score fuctio. Before goig through all the details of the proof, we eed the followig Lemma that characterizes the first-order coditios of the maximum likelihood estimator assumig that suitable regularity coditios hold uder the ull hypothesis. Lemma 1. Uder suitable regularity coditios, the maximum likelihood estimator is the solutio of the followig first-order coditios usig the Lagragia fuctio: l ˆβ,0 β + ht ˆβ,0 ˆλ = 0 k 1 β Step 1 : Taylor expasio aroud β 0 of h ˆβ uder the ull ad the alterative hypotheses. We have h ˆβ,a a hβ0 β t ˆβ,a β 0 ad h ˆβ,0 a hβ0 β t ˆβ,0 β 0. 5 The Wald test does ot require the maximizatio of the ucostraied problem i.e. the maximizatio problem uder the alterative hypothesis.

44 8 Asymptotic tests 43 Step 2 : Differece of the first order Taylor expasios. Takig the differece of two previous expressios, we get: h ˆβ,a h ˆβ a hβ 0,0 ˆβ,a β ˆβ t,0. Sice h ˆβ,0 = 0 uder the ull hypothesis costraied estimator, we obtai: h ˆβ,a a hβ0 β t ˆβ,a ˆβ,0. 5 Step 3 : Rewritig of ˆβ,a ˆβ,0 ˆβ,a ˆβ,0 ca be rewritte by usig first-order Taylor expasios of the score fuctio aroud β 0 uder the ull ad alterative hypothesis. Ideed, we have: l β ˆβ,0 which implies: a a a l β β0 + 2 l β β t β0 ˆβ,0 β 0 l [ β β0 1 ] 2 l β β t β0 ˆβ,0 β 0 l β β0 Iβ 0 ˆβ,0 β 0 1 l β ˆβ,0 a 1 l β β0 Iβ 0 ˆβ,0 β 0. 6 I the same respect, we get: 1 l β ˆβ,a a 1 l β β0 Iβ 0 ˆβ,a β 0 7 with ucostraied estimator: 1 l β ˆβ,a = 0 k 1.

45 8 Asymptotic tests 44 Takig the differece Eq. 6 - Eq. 7, we get: which implies: 1 l ˆβ,0 β a Iβ 0 ˆβ,a ˆβ,0 Step 4 : Usig Step 3 i Eq. 5. ˆβ,a ˆβ a,0 Iβ l ˆβ,0. β Replacig the previous result i Eq. 5, we get: Step 5 : Usig Lemma 1 i Eq. 8. Give that Lemma 1: h ˆβ,a a hβ0 Iβ l ˆβ,0. β t β it follows that: l ˆβ,0 β + ht ˆβ,0 ˆλ = 0 k 1 β which is equivalet to: h ˆβ,a a hβ0 Iβ 0 1 ht ˆβ,0 ˆλ β t β h ˆβ,a a hβ0 Iβ 0 1 ht β 0 ˆλ 8 β t β sice ˆβ,0 is a cosistet estimator of β 0 uder the ull hypothesis.

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT OCTOBER 7, 2016 LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT Geometry of LS We ca thik of y ad the colums of X as members of the -dimesioal Euclidea space R Oe ca

More information

11 THE GMM ESTIMATION

11 THE GMM ESTIMATION Cotets THE GMM ESTIMATION 2. Cosistecy ad Asymptotic Normality..................... 3.2 Regularity Coditios ad Idetificatio..................... 4.3 The GMM Iterpretatio of the OLS Estimatio.................

More information

Algebra of Least Squares

Algebra of Least Squares October 19, 2018 Algebra of Least Squares Geometry of Least Squares Recall that out data is like a table [Y X] where Y collects observatios o the depedet variable Y ad X collects observatios o the k-dimesioal

More information

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

1 General linear Model Continued..

1 General linear Model Continued.. Geeral liear Model Cotiued.. We have We kow y = X + u X o radom u v N(0; I ) b = (X 0 X) X 0 y E( b ) = V ar( b ) = (X 0 X) We saw that b = (X 0 X) X 0 u so b is a liear fuctio of a ormally distributed

More information

Solution to Chapter 2 Analytical Exercises

Solution to Chapter 2 Analytical Exercises Nov. 25, 23, Revised Dec. 27, 23 Hayashi Ecoometrics Solutio to Chapter 2 Aalytical Exercises. For ay ε >, So, plim z =. O the other had, which meas that lim E(z =. 2. As show i the hit, Prob( z > ε =

More information

Asymptotic Results for the Linear Regression Model

Asymptotic Results for the Linear Regression Model Asymptotic Results for the Liear Regressio Model C. Fli November 29, 2000 1. Asymptotic Results uder Classical Assumptios The followig results apply to the liear regressio model y = Xβ + ε, where X is

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday

More information

ECON 3150/4150, Spring term Lecture 3

ECON 3150/4150, Spring term Lecture 3 Itroductio Fidig the best fit by regressio Residuals ad R-sq Regressio ad causality Summary ad ext step ECON 3150/4150, Sprig term 2014. Lecture 3 Ragar Nymoe Uiversity of Oslo 21 Jauary 2014 1 / 30 Itroductio

More information

POLS, GLS, FGLS, GMM. Outline of Linear Systems of Equations. Common Coefficients, Panel Data Model. Preliminaries

POLS, GLS, FGLS, GMM. Outline of Linear Systems of Equations. Common Coefficients, Panel Data Model. Preliminaries Outlie of Liear Systems of Equatios POLS, GLS, FGLS, GMM Commo Coefficiets, Pael Data Model Prelimiaries he liear pael data model is a static model because all explaatory variables are dated cotemporaeously

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Rank tests and regression rank scores tests in measurement error models

Rank tests and regression rank scores tests in measurement error models Rak tests ad regressio rak scores tests i measuremet error models J. Jurečková ad A.K.Md.E. Saleh Charles Uiversity i Prague ad Carleto Uiversity i Ottawa Abstract The rak ad regressio rak score tests

More information

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio

More information

Linear Regression Models, OLS, Assumptions and Properties

Linear Regression Models, OLS, Assumptions and Properties Chapter 2 Liear Regressio Models, OLS, Assumptios ad Properties 2.1 The Liear Regressio Model The liear regressio model is the sigle most useful tool i the ecoometricia s kit. The multiple regressio model

More information

x iu i E(x u) 0. In order to obtain a consistent estimator of β, we find the instrumental variable z which satisfies E(z u) = 0. z iu i E(z u) = 0.

x iu i E(x u) 0. In order to obtain a consistent estimator of β, we find the instrumental variable z which satisfies E(z u) = 0. z iu i E(z u) = 0. 27 However, β MM is icosistet whe E(x u) 0, i.e., β MM = (X X) X y = β + (X X) X u = β + ( X X ) ( X u ) \ β. Note as follows: X u = x iu i E(x u) 0. I order to obtai a cosistet estimator of β, we fid

More information

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise)

Lecture 22: Review for Exam 2. 1 Basic Model Assumptions (without Gaussian Noise) Lecture 22: Review for Exam 2 Basic Model Assumptios (without Gaussia Noise) We model oe cotiuous respose variable Y, as a liear fuctio of p umerical predictors, plus oise: Y = β 0 + β X +... β p X p +

More information

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So,

First, note that the LS residuals are orthogonal to the regressors. X Xb X y = 0 ( normal equations ; (k 1) ) So, 0 2. OLS Part II The OLS residuals are orthogoal to the regressors. If the model icludes a itercept, the orthogoality of the residuals ad regressors gives rise to three results, which have limited practical

More information

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2 82 CHAPTER 4. MAXIMUM IKEIHOOD ESTIMATION Defiitio: et X be a radom sample with joit p.m/d.f. f X x θ. The geeralised likelihood ratio test g.l.r.t. of the NH : θ H 0 agaist the alterative AH : θ H 1,

More information

CLRM estimation Pietro Coretto Econometrics

CLRM estimation Pietro Coretto Econometrics Slide Set 4 CLRM estimatio Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Thursday 24 th Jauary, 2019 (h08:41) P. Coretto

More information

[412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION

[412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION [412] A TEST FOR HOMOGENEITY OF THE MARGINAL DISTRIBUTIONS IN A TWO-WAY CLASSIFICATION BY ALAN STUART Divisio of Research Techiques, Lodo School of Ecoomics 1. INTRODUCTION There are several circumstaces

More information

MA Advanced Econometrics: Properties of Least Squares Estimators

MA Advanced Econometrics: Properties of Least Squares Estimators MA Advaced Ecoometrics: Properties of Least Squares Estimators Karl Whela School of Ecoomics, UCD February 5, 20 Karl Whela UCD Least Squares Estimators February 5, 20 / 5 Part I Least Squares: Some Fiite-Sample

More information

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain Assigmet 9 Exercise 5.5 Let X biomial, p, where p 0, 1 is ukow. Obtai cofidece itervals for p i two differet ways: a Sice X / p d N0, p1 p], the variace of the limitig distributio depeds oly o p. Use the

More information

Simple Linear Regression

Simple Linear Regression Chapter 2 Simple Liear Regressio 2.1 Simple liear model The simple liear regressio model shows how oe kow depedet variable is determied by a sigle explaatory variable (regressor). Is is writte as: Y i

More information

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen) Goodess-of-Fit Tests ad Categorical Data Aalysis (Devore Chapter Fourtee) MATH-252-01: Probability ad Statistics II Sprig 2019 Cotets 1 Chi-Squared Tests with Kow Probabilities 1 1.1 Chi-Squared Testig................

More information

A statistical method to determine sample size to estimate characteristic value of soil parameters

A statistical method to determine sample size to estimate characteristic value of soil parameters A statistical method to determie sample size to estimate characteristic value of soil parameters Y. Hojo, B. Setiawa 2 ad M. Suzuki 3 Abstract Sample size is a importat factor to be cosidered i determiig

More information

Introductory statistics

Introductory statistics CM9S: Machie Learig for Bioiformatics Lecture - 03/3/06 Itroductory statistics Lecturer: Sriram Sakararama Scribe: Sriram Sakararama We will provide a overview of statistical iferece focussig o the key

More information

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK)

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) Everythig marked by is ot required by the course syllabus I this lecture, all vector spaces is over the real umber R. All vectors i R is viewed as a colum

More information

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data.

In this section we derive some finite-sample properties of the OLS estimator. b is an estimator of β. It is a function of the random sample data. 17 3. OLS Part III I this sectio we derive some fiite-sample properties of the OLS estimator. 3.1 The Samplig Distributio of the OLS Estimator y = Xβ + ε ; ε ~ N[0, σ 2 I ] b = (X X) 1 X y = f(y) ε is

More information

Linear Regression Demystified

Linear Regression Demystified Liear Regressio Demystified Liear regressio is a importat subject i statistics. I elemetary statistics courses, formulae related to liear regressio are ofte stated without derivatio. This ote iteds to

More information

Classical Linear Regression Model. Normality Assumption Hypothesis Testing Under Normality Maximum Likelihood Estimator Generalized Least Squares

Classical Linear Regression Model. Normality Assumption Hypothesis Testing Under Normality Maximum Likelihood Estimator Generalized Least Squares Classical Liear Regressio Model Normality Assumptio Hypothesis Testig Uder Normality Maximum Likelihood Estimator Geeralized Least Squares Normality Assumptio Assumptio 5 e X ~ N(,s I ) Implicatios of

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett

Lecture Note 8 Point Estimators and Point Estimation Methods. MIT Spring 2006 Herman Bennett Lecture Note 8 Poit Estimators ad Poit Estimatio Methods MIT 14.30 Sprig 2006 Herma Beett Give a parameter with ukow value, the goal of poit estimatio is to use a sample to compute a umber that represets

More information

Lecture 3. Properties of Summary Statistics: Sampling Distribution

Lecture 3. Properties of Summary Statistics: Sampling Distribution Lecture 3 Properties of Summary Statistics: Samplig Distributio Mai Theme How ca we use math to justify that our umerical summaries from the sample are good summaries of the populatio? Lecture Summary

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2. SAMPLE STATISTICS A radom sample x 1,x,,x from a distributio f(x) is a set of idepedetly ad idetically variables with x i f(x) for all i Their joit pdf is f(x 1,x,,x )=f(x 1 )f(x ) f(x )= f(x i ) The sample

More information

Lecture 33: Bootstrap

Lecture 33: Bootstrap Lecture 33: ootstrap Motivatio To evaluate ad compare differet estimators, we eed cosistet estimators of variaces or asymptotic variaces of estimators. This is also importat for hypothesis testig ad cofidece

More information

TAMS24: Notations and Formulas

TAMS24: Notations and Formulas TAMS4: Notatios ad Formulas Basic otatios ad defiitios X: radom variable stokastiska variabel Mea Vätevärde: µ = X = by Xiagfeg Yag kpx k, if X is discrete, xf Xxdx, if X is cotiuous Variace Varias: =

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

10. Comparative Tests among Spatial Regression Models. Here we revisit the example in Section 8.1 of estimating the mean of a normal random

10. Comparative Tests among Spatial Regression Models. Here we revisit the example in Section 8.1 of estimating the mean of a normal random Part III. Areal Data Aalysis 0. Comparative Tests amog Spatial Regressio Models While the otio of relative likelihood values for differet models is somewhat difficult to iterpret directly (as metioed above),

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

LECTURE 14 NOTES. A sequence of α-level tests {ϕ n (x)} is consistent if

LECTURE 14 NOTES. A sequence of α-level tests {ϕ n (x)} is consistent if LECTURE 14 NOTES 1. Asymptotic power of tests. Defiitio 1.1. A sequece of -level tests {ϕ x)} is cosistet if β θ) := E θ [ ϕ x) ] 1 as, for ay θ Θ 1. Just like cosistecy of a sequece of estimators, Defiitio

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 9 Multicolliearity Dr Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Multicolliearity diagostics A importat questio that

More information

The Method of Least Squares. To understand least squares fitting of data.

The Method of Least Squares. To understand least squares fitting of data. The Method of Least Squares KEY WORDS Curve fittig, least square GOAL To uderstad least squares fittig of data To uderstad the least squares solutio of icosistet systems of liear equatios 1 Motivatio Curve

More information

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >

More information

The standard deviation of the mean

The standard deviation of the mean Physics 6C Fall 20 The stadard deviatio of the mea These otes provide some clarificatio o the distictio betwee the stadard deviatio ad the stadard deviatio of the mea.. The sample mea ad variace Cosider

More information

Introduction to Machine Learning DIS10

Introduction to Machine Learning DIS10 CS 189 Fall 017 Itroductio to Machie Learig DIS10 1 Fu with Lagrage Multipliers (a) Miimize the fuctio such that f (x,y) = x + y x + y = 3. Solutio: The Lagragia is: L(x,y,λ) = x + y + λ(x + y 3) Takig

More information

of the matrix is =-85, so it is not positive definite. Thus, the first

of the matrix is =-85, so it is not positive definite. Thus, the first BOSTON COLLEGE Departmet of Ecoomics EC771: Ecoometrics Sprig 4 Prof. Baum, Ms. Uysal Solutio Key for Problem Set 1 1. Are the followig quadratic forms positive for all values of x? (a) y = x 1 8x 1 x

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

Economics 326 Methods of Empirical Research in Economics. Lecture 18: The asymptotic variance of OLS and heteroskedasticity

Economics 326 Methods of Empirical Research in Economics. Lecture 18: The asymptotic variance of OLS and heteroskedasticity Ecoomics 326 Methods of Empirical Research i Ecoomics Lecture 8: The asymptotic variace of OLS ad heteroskedasticity Hiro Kasahara Uiversity of British Columbia December 24, 204 Asymptotic ormality I I

More information

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency Math 152. Rumbos Fall 2009 1 Solutios to Review Problems for Exam #2 1. I the book Experimetatio ad Measuremet, by W. J. Youde ad published by the by the Natioal Sciece Teachers Associatio i 1962, the

More information

Chi-Squared Tests Math 6070, Spring 2006

Chi-Squared Tests Math 6070, Spring 2006 Chi-Squared Tests Math 6070, Sprig 2006 Davar Khoshevisa Uiversity of Utah February XXX, 2006 Cotets MLE for Goodess-of Fit 2 2 The Multiomial Distributio 3 3 Applicatio to Goodess-of-Fit 6 3 Testig for

More information

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector Summary ad Discussio o Simultaeous Aalysis of Lasso ad Datzig Selector STAT732, Sprig 28 Duzhe Wag May 4, 28 Abstract This is a discussio o the work i Bickel, Ritov ad Tsybakov (29). We begi with a short

More information

Problem Set 4 Due Oct, 12

Problem Set 4 Due Oct, 12 EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 016 MODULE : Statistical Iferece Time allowed: Three hours Cadidates should aswer FIVE questios. All questios carry equal marks. The umber

More information

Multivariate Regression: Estimating & Reporting Certainty, Hypotheses Tests

Multivariate Regression: Estimating & Reporting Certainty, Hypotheses Tests Multivariate Regressio: Estimatig & Reportig Certaity, Hypotheses Tests I. The C(N)LRM: A. Core Assumptios:. y=xβ +. E () = 0 3. V () = σ I 4. E ( X) = 0 5. X of full-colum rak. B. Coveiece Assumptios

More information

Statistical Properties of OLS estimators

Statistical Properties of OLS estimators 1 Statistical Properties of OLS estimators Liear Model: Y i = β 0 + β 1 X i + u i OLS estimators: β 0 = Y β 1X β 1 = Best Liear Ubiased Estimator (BLUE) Liear Estimator: β 0 ad β 1 are liear fuctio of

More information

Statistics 511 Additional Materials

Statistics 511 Additional Materials Cofidece Itervals o mu Statistics 511 Additioal Materials This topic officially moves us from probability to statistics. We begi to discuss makig ifereces about the populatio. Oe way to differetiate probability

More information

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS Jauary 25, 207 INTRODUCTION TO MATHEMATICAL STATISTICS Abstract. A basic itroductio to statistics assumig kowledge of probability theory.. Probability I a typical udergraduate problem i probability, we

More information

Chi-squared tests Math 6070, Spring 2014

Chi-squared tests Math 6070, Spring 2014 Chi-squared tests Math 6070, Sprig 204 Davar Khoshevisa Uiversity of Utah March, 204 Cotets MLE for goodess-of fit 2 2 The Multivariate ormal distributio 3 3 Cetral limit theorems 5 4 Applicatio to goodess-of-fit

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Math 61CM - Solutions to homework 3

Math 61CM - Solutions to homework 3 Math 6CM - Solutios to homework 3 Cédric De Groote October 2 th, 208 Problem : Let F be a field, m 0 a fixed oegative iteger ad let V = {a 0 + a x + + a m x m a 0,, a m F} be the vector space cosistig

More information

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley Notes O Media ad Quatile Regressio James L. Powell Departmet of Ecoomics Uiversity of Califoria, Berkeley Coditioal Media Restrictios ad Least Absolute Deviatios It is well-kow that the expected value

More information

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight)

Tests of Hypotheses Based on a Single Sample (Devore Chapter Eight) Tests of Hypotheses Based o a Sigle Sample Devore Chapter Eight MATH-252-01: Probability ad Statistics II Sprig 2018 Cotets 1 Hypothesis Tests illustrated with z-tests 1 1.1 Overview of Hypothesis Testig..........

More information

Regression with an Evaporating Logarithmic Trend

Regression with an Evaporating Logarithmic Trend Regressio with a Evaporatig Logarithmic Tred Peter C. B. Phillips Cowles Foudatio, Yale Uiversity, Uiversity of Aucklad & Uiversity of York ad Yixiao Su Departmet of Ecoomics Yale Uiversity October 5,

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

Logit regression Logit regression

Logit regression Logit regression Logit regressio Logit regressio models the probability of Y= as the cumulative stadard logistic distributio fuctio, evaluated at z = β 0 + β X: Pr(Y = X) = F(β 0 + β X) F is the cumulative logistic distributio

More information

5. Likelihood Ratio Tests

5. Likelihood Ratio Tests 1 of 5 7/29/2009 3:16 PM Virtual Laboratories > 9. Hy pothesis Testig > 1 2 3 4 5 6 7 5. Likelihood Ratio Tests Prelimiaries As usual, our startig poit is a radom experimet with a uderlyig sample space,

More information

Common Large/Small Sample Tests 1/55

Common Large/Small Sample Tests 1/55 Commo Large/Small Sample Tests 1/55 Test of Hypothesis for the Mea (σ Kow) Covert sample result ( x) to a z value Hypothesis Tests for µ Cosider the test H :μ = μ H 1 :μ > μ σ Kow (Assume the populatio

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates Iteratioal Joural of Scieces: Basic ad Applied Research (IJSBAR) ISSN 2307-4531 (Prit & Olie) http://gssrr.org/idex.php?joural=jouralofbasicadapplied ---------------------------------------------------------------------------------------------------------------------------

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

Statistical inference: example 1. Inferential Statistics

Statistical inference: example 1. Inferential Statistics Statistical iferece: example 1 Iferetial Statistics POPULATION SAMPLE A clothig store chai regularly buys from a supplier large quatities of a certai piece of clothig. Each item ca be classified either

More information

Correlation Regression

Correlation Regression Correlatio Regressio While correlatio methods measure the stregth of a liear relatioship betwee two variables, we might wish to go a little further: How much does oe variable chage for a give chage i aother

More information

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Discrete Mathematics for CS Spring 2008 David Wagner Note 22 CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 22 I.I.D. Radom Variables Estimatig the bias of a coi Questio: We wat to estimate the proportio p of Democrats i the US populatio, by takig

More information

Apply change-of-basis formula to rewrite x as a linear combination of eigenvectors v j.

Apply change-of-basis formula to rewrite x as a linear combination of eigenvectors v j. Eigevalue-Eigevector Istructor: Nam Su Wag eigemcd Ay vector i real Euclidea space of dimesio ca be uiquely epressed as a liear combiatio of liearly idepedet vectors (ie, basis) g j, j,,, α g α g α g α

More information

Bayesian Methods: Introduction to Multi-parameter Models

Bayesian Methods: Introduction to Multi-parameter Models Bayesia Methods: Itroductio to Multi-parameter Models Parameter: θ = ( θ, θ) Give Likelihood p(y θ) ad prior p(θ ), the posterior p proportioal to p(y θ) x p(θ ) Margial posterior ( θ, θ y) is Iterested

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

Chapter 13: Tests of Hypothesis Section 13.1 Introduction

Chapter 13: Tests of Hypothesis Section 13.1 Introduction Chapter 13: Tests of Hypothesis Sectio 13.1 Itroductio RECAP: Chapter 1 discussed the Likelihood Ratio Method as a geeral approach to fid good test procedures. Testig for the Normal Mea Example, discussed

More information

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions

Overview. p 2. Chapter 9. Pooled Estimate of. q = 1 p. Notation for Two Proportions. Inferences about Two Proportions. Assumptions Chapter 9 Slide Ifereces from Two Samples 9- Overview 9- Ifereces about Two Proportios 9- Ifereces about Two Meas: Idepedet Samples 9-4 Ifereces about Matched Pairs 9-5 Comparig Variatio i Two Samples

More information

5.1 Review of Singular Value Decomposition (SVD)

5.1 Review of Singular Value Decomposition (SVD) MGMT 69000: Topics i High-dimesioal Data Aalysis Falll 06 Lecture 5: Spectral Clusterig: Overview (cotd) ad Aalysis Lecturer: Jiamig Xu Scribe: Adarsh Barik, Taotao He, September 3, 06 Outlie Review of

More information

Inverse Matrix. A meaning that matrix B is an inverse of matrix A.

Inverse Matrix. A meaning that matrix B is an inverse of matrix A. Iverse Matrix Two square matrices A ad B of dimesios are called iverses to oe aother if the followig holds, AB BA I (11) The otio is dual but we ofte write 1 B A meaig that matrix B is a iverse of matrix

More information

b i u x i U a i j u x i u x j

b i u x i U a i j u x i u x j M ath 5 2 7 Fall 2 0 0 9 L ecture 1 9 N ov. 1 6, 2 0 0 9 ) S ecod- Order Elliptic Equatios: Weak S olutios 1. Defiitios. I this ad the followig two lectures we will study the boudary value problem Here

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

Last Lecture. Wald Test

Last Lecture. Wald Test Last Lecture Biostatistics 602 - Statistical Iferece Lecture 22 Hyu Mi Kag April 9th, 2013 Is the exact distributio of LRT statistic typically easy to obtai? How about its asymptotic distributio? For testig

More information

LECTURE 8: ASYMPTOTICS I

LECTURE 8: ASYMPTOTICS I LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece

More information

Lecture 12: September 27

Lecture 12: September 27 36-705: Itermediate Statistics Fall 207 Lecturer: Siva Balakrisha Lecture 2: September 27 Today we will discuss sufficiecy i more detail ad the begi to discuss some geeral strategies for costructig estimators.

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

6.3 Testing Series With Positive Terms

6.3 Testing Series With Positive Terms 6.3. TESTING SERIES WITH POSITIVE TERMS 307 6.3 Testig Series With Positive Terms 6.3. Review of what is kow up to ow I theory, testig a series a i for covergece amouts to fidig the i= sequece of partial

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit theorems Throughout this sectio we will assume a probability space (Ω, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Matrix Representation of Data in Experiment

Matrix Representation of Data in Experiment Matrix Represetatio of Data i Experimet Cosider a very simple model for resposes y ij : y ij i ij, i 1,; j 1,,..., (ote that for simplicity we are assumig the two () groups are of equal sample size ) Y

More information

Section 14. Simple linear regression.

Section 14. Simple linear regression. Sectio 14 Simple liear regressio. Let us look at the cigarette dataset from [1] (available to dowload from joural s website) ad []. The cigarette dataset cotais measuremets of tar, icotie, weight ad carbo

More information

Lecture 6 Simple alternatives and the Neyman-Pearson lemma

Lecture 6 Simple alternatives and the Neyman-Pearson lemma STATS 00: Itroductio to Statistical Iferece Autum 06 Lecture 6 Simple alteratives ad the Neyma-Pearso lemma Last lecture, we discussed a umber of ways to costruct test statistics for testig a simple ull

More information