Econometria. Estimation and hypotheses testing in the uni-equational linear regression model: cross-section data. Luca Fanelli. University of Bologna

Size: px
Start display at page:

Download "Econometria. Estimation and hypotheses testing in the uni-equational linear regression model: cross-section data. Luca Fanelli. University of Bologna"

Transcription

1 Econometria Estimation and hypotheses testing in the uni-equational linear regression model: cross-section data Luca Fanelli University of Bologna

2 Estimation and hypotheses testing in the uni-equational regression model - Model - Compact representation - OLS - GLS - ML - Constrained estimation - Testing linear hypotheses

3 Estimation: cross-section data Uni-equational linear regression model 8 >< >: y i = x 0 i + u i, i = 1; 2; :::; n classical: generalized: Cov(u i, u j ):=0, 8i 6= j E(u 2 i ):=2 u, 8i Cov(u i, u j ):= i;j zero or not E(u 2 i ):=2 i We know that the generalized model has to be intended as `virtual' unless further information is used ( i;j and 2 i are generally unknown)

4 Compact matrix representation y n1 = X nk k1 + u n1 E(u)= 0 n1, E(uu 0 ):= nn Why do we use this representation? Because it is useful in some cases! It is particularly useful in order to derive the estimators of interest!

5 Write y i = x 0 i + u i, i = 1; 2; :::; n as: y 1 = x u 1, i:=1 y 2 = x u 2, i:=2.... y n = x 0 n + u n, i:=n: Compactly: y 1 y 2. y n = x 0 1 x 0 2. x 0 n u 1 u 2. u n y n1 = X nk k1 + u n1

6 What about the matrix E(uu 0 ):= nn? E(uu 0 ):=E = E u 1 u 2. u n [u 1; u 2 ; : : : u n ] u 2 1 u 1 u 2 u 1 u n u 2 u 1 u u n u 1 u n u 2 u 2 n 1 C A 31 7C 5A

7 = E(u 2 1 ) E(u 1u 2 ) E(u 1 u n ) E(u 2 u 1 ) E(u 2 2 ).... E(u n u 1 ) E(u n u 2 ) E(u 2 n) = E(u 2 1 ) Cov(u 1; u 2 ) Cov(u 1 ; u n ) Cov(u 2 ; u 1 ) E(u 2 2 ).... Cov(u n ; u 1 ) Cov(u n ; u 2 ) E(u 2 n) = ;2 1;n 2; n;1 n;2 2 n It is clear that := ( 2 u I n 6= 2 ui n classical model generalzied model :

8 To sum up, the compact representation of the uniequational linear regression model based on cross-section data is: y n1 = X nk k1 + u n1 where E(u)= 0 n1, E(uu0 ):= nn := ( 2 u I n 6= 2 u I n classical model generalized model :

9 The representations y i = x 0 i + u i, i = 1; 2; :::; n and y n1 are interchangeable. = X nk k1 + u n1 The link between the two is immediately given if one recalls that link is obtained by: X 0 X nx i=1 x i x 0 i ; X 0 y nx i=1 x i y i :

10 We have introduced the classical uni-equation regression model in the following way: y i = x 0 i + u i, i = 1; 2; :::; n E(u i j x i ) = 0 ) E(u i ) = 0 Hp: E(u 2 i j x i) = 2 i Hp: Cov(u i ; u j ) = i;j 8i 6= j: Using the compact representation, we can say even more about the feautes of this model.

11 In particular, given the conditional nature of the model (we are conditioning with respect to regressors) we will be writing y n1 = X nk k1 + u n1 E(u j X) = 0 n1 ) E(u) = 0 n1 E(uu 0 j X):= nn ) E(uu 0 ) = nn so that u and X can be thought of being stocastic independent in the sense that E(u j X) = E(u) E(uu 0 j X) = E(uu 0 ):

12 It is useful to focus on the meaning n 1 vector E(u j X): From the previous set of slides we know: E(u j X):=E 00 u 1. u n 1 C 1 C A j XA := E(u 1 j X). E(u n j X) Now consider e.g. E(u 1 j X) E(u 1 jall stochastic variables within X); E(u 2 j X) E(u 2 jall stochastic variables within X);... E(u n j X) E(u n jall stochastic variables within X):

13 Thus E(u n j X) = 0 means that conditional on the knowledge of all regressors x 1 (i:=1); x 2 (i:=2),..., x n (i:=n) that enter the matrix X, the expected value of u n is zero as the unconditional expectation: E(u n j X) = E(u n ) = 0: This holds for E(u n 1 j X), E(u n 2 j X); :::; E(u 1 j X) and allows us to write that E(u j X)=0 n1 ) E X (E(u j X)) =0 n1 The condition E(u j X)=0 n1 is usually assumed with cross-section data and is stronger than it apparently seems. Why? Example on the blackboard. E(u j X)=0 n1 ) E(u)=0 n1 (law of iterated expectations) The reverse it is not true, E(u)=0 n1 does not imply E(u j X)=0 n1!

14 For regression models based on time series data where lags of y appear among the regressors, the condition E(u j X)=0 n1 doesn't hold (but it holds E(u)=0 n1 ) Example on the blackboard based on: c i = c i z i + u i, i = 1; :::; n:

15 Classical model with cross-section data: OLS estimation y n1 = X nk k1 + u n1 E(u j X)= 0 n1, E(uu0 j X):= 2 ui n: Vector of k + 1 unknown parameters: = ( 0 ; 2 u) 0. Objective function: Q() = 1 2 u nx i=1 (y i x 0 i )2 1 2 u (y X ) 0 (y X ):

16 Given Q() = 1 2 u nx i=1 (y i x 0 i )2 1 2 u (y X) 0 (y X); the OLS estimator of is obtained by solving the problem min Q() i.e. the OLS estimator of is the vector that solves: ^ OLS = arg min Q():

17 First-order conditions: where note that 2 u 2 u 2 u k 1 C A = 0 k1 = C A : The needed derivative rules are in the Appendix 3.A.3 (Ch. 3) of the textbook. Exercise at home.

18 The solution of this problem leads us to ^ OLS := 0 i=1 x i x 0 i 1 A 1 0 i=1 1 x i y i A X 0 X 1 X 0 y : The condition rank(x) = k ) X 0 X is non-singular is crucial to obtain a valid (unique) OLS estimator. The estimator of 2 u is obtained indirectly ^ 2 u = 1 n k nx i=1 ^u 2 i 1 A 1 n k ^u0^u where ^u i = y i x 0 i^ OLS, i = 1; :::; n or, alternatively, ^u = (y X ^OLS ):

19 Are the estimators of and 2 u correct? E(^OLS )=, E(^ 2 u)= 2 u?? (if yes, under which conditions?) Consider that ^ OLS :=(X 0 X) 1 X 0 [X + u] = + (X 0 X) 1 X 0 u and E(^OLS ):=E X E ^OLS j X : Likewise, E(^ 2 u ):=E X E ^ 2 u j X :

20 Estimator of : E ^OLS j X = + (X 0 X) 1 X 0 E (u j X) : Hence, if E (u j X) = 0 n1, one has E ^OLS j X = ) E(^OLS ):=E X E ^OLS j X = E X () =; the OLS estimator is correct. Note that if E (u j X) 6= 0 n1, the estimator is no longer correct (it happens in the regression model with time series data in which regressors include lags of y).

21 Estimator of 2 u: To check whether the estimator ^ 2 u is correct, we start from some computations and considerations: ^u = (y X ^OLS ) = (y X (X 0 X) 1 X 0 y) =(X + u X (X 0 X) 1 X 0 )(X + u) =(X + u X X(X 0 X) 1 X 0 u) =(I n X(X 0 X) 1 X 0 )u The matrix (I n X(X 0 X) 1 X 0 ) will be indicated with the symbol: M XX :

22 M XX :=(I n X(X 0 X) 1 X 0 ) is a `special' matrix (idempotent matrix): - is symmetric: M 0 XX = M XX; - is such that M XX M XX = M XX - rank(m XX )=tr(m XX )=n k: (Properties of trace operator on the blackboard).

23 We have proved that ^u := (y X ^OLS ) = M XX u: E ^ 2 u j X = E 1 n =E 1 n k u0 M XX u j X k ^u0^u j X = n 1 k E u0 M XX u j X = n 1 k E tr u0 M XX u j X = n 1 k E tr M XXuu 0 j X = n 1 k tr E M XXuu 0 j X = n 1 k tr M XXE uu 0 j X = n 1 k tr M XX 2 u I n = n 2 k tr (M XX) = 2 u n k (n k) = 2 u

24 We have proved that E(^ 2 u j X) = 2 u ) E(^ 2 u):=e X E ^ 2 u j X = E X 2 u = 2 u ; the OLS estimator of 2 u is correct.

25 Covariance matrix of ^OLS. V ar ^OLS j X =V ar + (X 0 X) 1 X 0 u j X =V ar (X 0 X) 1 X 0 u j X :=(X 0 X) 1 X 0 2 ui n X(X 0 X) 1 = 2 u(x 0 X) 1 : Recall section Let v be p1 stochastic vector with V ar(v):=v (symmetric positive denite) and A m p non-stochastic matrix. Then (sandwich rule). V ar(av) = A mp V A 0 pp pm Assume now that also A is stochastic, i.e. random variables. its elements are Then the quantity V ar(av j A) means that we can treat A as if it is a non-stochastic matrix! Hence V ar(av j A):=AV ar(v j A)A 0 : End of recall section

26 Note that V ar ^OLS :=EX V ar ^OLS j X = E X 2 u (X 0 X) 1 = 2 u E h (X 0 X) 1i = 2 ue h ( P n i=1 x i x 0 i ) 1i :

27 To sum up: The OLS estimator of in the linear classical model based on cross-section data is correct E(^OLS ) = and has (conditional) covariance matrix: V ar ^OLS j X := 2 u(x 0 X) 1 : The OLS estimator of 2 u E(^ 2 u)= 2 u in the linear classical model based on cross-section data is correct. Usually we are not interested to the variance of the estimator of ^ 2 u but in principle we can also derive it.

28 A famous theorem (Gauss-Markov Theorem) applies to the case in which X does not contain stocastic variables. It says that given the linear model y n1 = X nk k1 + u n1 E(u)=0 n1, E(uu 0 ):= 2 ui n the OLS estimator is, in the class of linear correct estimators, the one whose covariance matrix V ar ^OLS := 2 u (X 0 X) 1 is the `most ecient' (minimum covariance matrix).

29 In our case, X is stochastic (we condition the model with respect to the regressors). A version of the Gauss-Markov theorem where X is stochastic and all variables are conditioned with respect to the elements of X still applies. This means that any estimator dierent from the OLS estimator has (conditional) covariance matrix `larger' than V ar ^OLS j X := 2 u(x 0 X) 1 : This is why the OLS estimator applied to cross-section data (under homoskedasticity) is often called BLUE (Best Linear Unbiased Estimator).

30 Recall section Let 1 and 2 two symmetric positive denite matrices. We say that 1 `is larger' than 2 if the matrix 1 2 is semipositive denite (this means that the eigenvalues of 1 2 are either positive or zero). What are eigenvalues? Given a q q matrix A the eigenvalues are the solutions (that can be complex numbers) to the problem det(i q A) = 0: If the matrix A is symmetric the eigenvalues are real mumbers. If A is symmetric and positive denite the eigenvalues are real and positive. End of recall section

31 Why do we care about the covariance matrix of the estimator ^OLS? Because it is fundamental to make inference! Imagine that ^ OLS j X N(, 2 u(x 0 X) 1 ): ^G j X 2 (n k), ^G:=^2 u 2 (n k): u These distributions hold if u j X N(0, 2 ui n ) (irrespective of whether n is small or large). It is possible to show that the random vector ^ OLS j X is independent on the random variable ^G j X:

32 Observe that given then ^ OLS j X N(, 2 u(x 0 X) 1 ); ^ j j 1=2 j X N(0, 1), u cjj j = 1; :::; k where ^j is the j-th element of ^, j is the j-th element of and c jj is the j-th element on the main diagonal of the matrix (X 0 X) 1 :

33 It then easy to construct a statistial test for H 0 : j :=0 H 1 : j 6= 0. Recall that according to the theory of distributions, if Z N(0; 1) G 2 (q) and Z and G are independent then Z Gq 1=2 t(q):

34 Consider H 0 : j :=0 H 1 : j 6= 0 and the test statistic (t-ratio): ^ j 0 t:= 1=2 = ^ u cjj ^ j s.e.(^j ).

35 Then ^ j 0 ^ j 0 ^ j ^j t:= s.e.(^j ) = 0 1=2 = ^ u cjj u (c jj ) 1=2 ^ u (c jj ) 1=2 u (c jj ) 1=2 = u (c jj ) 1=2 ^ u u Accordingly, = ^ j 0 u (c jj ) 1=2 ^ 2 1=2 = u n k 2 u n k ^ j 0 u (c jj ) 1=2 1=2 ^G n k t j X = ^ j 0 u (c jj ) 1=2 ^G n k 1=2 j X N(0; 1) 1=2 j X t(n 2 (n k) n k Thus, if disturbances are Gaussian we can test the signicance of each regression coecient by running t-tests. These tests are t(n k) distributed. k): Observe that if n large, t(n k) N(0; 1).

36 OLS with `robust' standard errors The standard error s.e.(^j ):=^ u cjj 1=2 provides a measure of the variability of the estimator ^ j : This measure, however, has been obtained under the implicit assumption of homoskedasticity (indeed, to be fully ecient OLS requires homoskedasticity!).

37 Imagine that the `true model' is y n1 = X nk k1 + u n1 E(u j X)= 0 n1, E(uu0 j X):== n that means that we have heteroskedasticity! Assume that an econometrician erraneously believes that there is homoskedasticity and applies the OLS estimator. In this case, the OLS estimator is still correct (proof as exercise!) but will no longer be ecient! Is it possible to improve its eciency?

38 Then V ar ^OLS j X =V ar + (X 0 X) 1 X 0 u j X =V ar (X 0 X) 1 X 0 u j X :=(X 0 X) 1 X 0 X(X 0 X) 1 therefore (X 0 X) 1 X 0 X(X 0 X) 1 6= 2 u(xx) 1 : The matrix (X 0 X) 1 X 0 X(X 0 X) 1 is the `correct' covariance matrix of the OLS estimator in the presence of heteroskedasticity.

39 Let us focus in detail on the structure of the covariance matrix X 0 X: Since is diagonal, it is seen that X 0 X:= nx i=1 2 i x ix 0 i : The US econometrician White has shown that given the residuals ^u i :=y i x 0 i^ OLS, i = 1; 2; :::; n, the quantity 1 n nx i=1 ^u 2 i x ix 0 i is a consistent estimator for 1 n X0 X. In other words, for large n 1 n nx i=1 ^u 2 i x ix 0 i! p 1 n nx i=1 2 i x ix 0 i :

40 The consequence of this result is the following: when the econometrician has the suspect that homoskedasticity might not hold in his/her sample, he/she might take the standard errors not from the covariance matrix ^ 2 u(xx) 1 but rather on the covariance matrix = ( (X 0 X) 1 nx i=1 x i x 0 i ) 1 nx i=1 nx i=1 1 ^u 2 i x ix 0 A i (X 0 X) 1 ^u 2 i x ix 0 i 1 nx A ( i=1 x i x 0 i ) 1 The standard errors are `robust' in the sense that they take into account the heteroskedasticity. The robust standard errors are also known as `White standard errors'. Every econometric package complements OLS estimation with these standard errors.

41 Generalized model: GLS estimation y n1 = X nk k1 + u n1 E(u j X)= 0 n1, E(uu0 j X)= 6= 2 ui n: Vector of unknown parameters: = ( 0 ; vech()) 0. It is impossible to estimate on the basis of n observations. We need some assumptions. Hp: is known.

42 Objective function: Q() = (y X ) 0 1 (y X ) nx nx i=1 j=1 ij(y i x 0 i )(y j x 0 j ) ij elements of GLS estimator of (given ) is obtained by solving the problem min Q() i.e. the GLS estimator of is the vector that solves: ^ GLS = arg min Q():

43 The solution of this problem leads us to ^ GLS := X 0 1 X 1 X 0 1 y : The condition rank(x) = k ) X 0 1 X is non-singular is crucial to obtain a valid (unique) GLS estimator. It can be noticed that the GLS estimator of is `virtual' in the sense that it requires the knowledge of otherwise it can not be computed.

44 ^ GLS = + X 0 1 X 1 X 0 1 u E(^GLS j X) = + X 0 1 X 1 X 0 1 E(u j X) = : V ar(^gls j X) = V ar(+ X 0 1 X X) 1 X 0 1 u j =V ar( X 0 1 X 1 X 0 1 u j X) = X 0 1 X 1 The GLS estimator is correct under the assumption E(u j X) = 0 n1 and has conditional covariance matrix V ar(^gls j X) = X 0 1 X 1 :

45 A famous theorem (Aitken Theorem) applies to the case in which X does not contain stocastic variables. It says that given the generalized linear model y n1 = X nk k1 + u n1 E(u)= 0 n1, E(uu0 ):= 6= 2 ui n the GLS estimator is, in the class of linear correct estimators, the one whose covariance matrix V ar ^GLS :=(X 0 1 X) 1 is the `most ecient' (minimum covariance matrix).

46 In our case, X is stochastic (we condition the model with respect to the regressors). A version of the Aitken theorem where X is stochastic and all variables are conditioned with respect to the elements of X still applies. This means that any estimator dierent from the GLS estimator has (conditional) covariance matrix `larger' than V ar ^GLS j X :=(X 0 1 X) 1 :

47 Feasible GLS estimators There are cases in which the GLS is `feasible', i.e. it can be calculated by using the information in the data. Imagine the model is: y i = x 1;i + 2 x 2;i + u i, i = 1; :::; n E(u i j x i ) = 0 ) E(u i ) = 0 Hp: E(u 2 i j x i) = 2 i = h (x 1i) 2, h > 0 Hp: Cov(u i ; u j ) = 0, 8i 6= j: We can transform the model by dividing both sides by 1 i, obtaining y i i = 1 i i 1 x 1;i + 1 i 2 x 2;i + 1 i u i

48 which is equivalent to y i (h) 1=2 x 1i = which is equivalent to where 1 (h) 1=2 x 1i (h) 1=2 x 1i 1 x 1;i 1 1 (h) 1=2 2 x 2;i + x 1i (h) 1=2 u i x 1i y i 1 = x 2;i 2 + u i x 1i x 1i x 1i x 1i y i = x 1;i + 2x 2;i + u i y i := y i x 1i, x 1;i := 1 x 1i, x 2;i :=x 2;i x 1i, u i := u i x 1i. Now, E( u i 2 j xi ) = E( u i x 1i 2 j xi ) = 1 x1i 2 E(u 2 i j x i ) = 1 x1i 2 h (x1i ) 2 = h 8i:

49 The transformed model y i = x 1;i + 2x 2;i + u i is homoskedatic because E( u i 2 j xi ):=h=const 8i: OLS can be applied to estimate 1, 0 and 2 eciently. This transformation is equivalent to the following idea: Hp : 2 i :=h (x 1i) 2, h > x1; ) :=h 0 x1; x1;n :=hv ^ GLS := X 0 (hv ) 1 X 1 X 0 (hv ) 1 y = X 0 V 1 X 1 X 0 V 1 y it is feasible because the matrix V is known!

50 Maximum Likelihood (ML) estimation Before discussing the ML estimation of the linear regression model, we overwiew this crucial estimation method. Statistical model = ( stochastic distribution sampling scheme ) f(data; ) = L() likelihood function. The g 1 vector of unknown parameters belongs to the (open) space, and the `true' value of, 0, is an interior point of.

51 The ML method is parametric because it requires the knowledge of the stochastic distribution of the variables, other than the sampling scheme. Model (classical to simplify): y i = x 0 i + u i Our parameters are: :=( 0, 2 u ) Key hypothesis (normality): y i j x i N(x 0 i, 2 u), i = 1; :::; n Recall that E(y i j x i )=x 0 i ; This implies u i j x i N(0, 2 u), i = 1; :::; n:

52 Thus, for each xed j 6= i u i u j! j x i; x j N 0 0!, " 2 u u #! : More generally, we can write so that where u j X:= 0 u 1 j X. u n j X f(u j X; ):= 1 C A N 0 n1, 2 ui n ny i=1 f(u i j X; ) 1 1 f(u i j X; ):= u (2) 1=2e 2 2 uu 2 i = 1 u (2) 1=2e u(y i x 0 i )2 :

53 Then L() = f(data; ) = ny i=1 f(u i j X; ): It will be convenient to focus on the log-likelihood function: log L() = nx i=1 log f(u i j X; ) = C n 2 log 2 u u nx i=1 (y i x 0 i )2 C n 2 log 2 u u (y X) 0 (y X): The ML estimator of is obtained by solving the problem i.e. max log L() ^ML = arg max log L():

54 As is known, in order to obtain ^ML it is necessary to solve the rst-order conditions: s n () log = log log log g 1 C A = 0 g1 that means that ^ML is such that s n (^ML ) = 0 g1 : The vector s n () is known as the score (gradient) of the likelihood function. It is not always possible to solve the rst order conditions analytically; in many curcumstances (e.g. nonlinear restrictions) numerical optimization procedures are required.

55 Note that s n () log P n i=1 log f(u i j x i ; = nx log f(u i j x i ; = nx i=1 s i (): Each component s i () of the score depends on the data, hence it is a random variable! To be sure that ^ML is a maximum (and not a minimum), it is further necessary that the Hessian matrix H n () = = @ 2 log 2 2 log 2 log 2 log 2 log 2 0s n() log 0 2 log 2 log 2 log 2 log 2 g be (semi)negative denite at the point ^ML, i.e. H n (^ML ) 0: 1 C A gg

56 We dene (Fisher) Information Matrix the quantity: I n () = log = E (H n ()) and it can be shown that under a set of regularity conditions, including the hypothesis of correct speci- cation of the model, one has the equivalence: I n () = log = E = E s n ()s n () 0 " # log # 0 1 A : The Asymptotic Information Matrix is given by the quantity 1 I 1 () = lim n!1 n I 1 n() = lim n!1 n E( H n()): The inverse of the matrix I 1 () represents a lower bound for any estimator of : this means that any estimator of will have covariance matrix `greater' than or at most equal to I1 1 ():

57 A crucial requirement for ML estimation is that log L( 0 ) > log L() for each 2 n f 0 g condition that ensures the existence of a global maximum. However, also situations of the time: log L( 0 ) > log L() for 2 N 0 where N 0 is a neighborhood of 0 are potentially ne, local maximum.

58 Properties of ML estimator Crucial Assumption: the model is correctly specied (the underlying statistical model is correct). 1. In general, E(^ML ) 6= 0 namely the ML is not correct! 2. Under general conditions the ML estimator is consistent: ^ML! p 0 : 3. Under general conditions the ML estimator is Asymptotically Gaussian: n 1=2 ^ML 0!D N(0 g1 ; V^): This property suggests that one can do standard inference in large samples!

59 4. Under general conditions the ML estimator estimator is asymptotically ecient, in particular V^ = [I 1 ()] 1 : The properties are asymptotic properties and make the ML estimator `optimal'.

60 max log L() ) max C n log 2 u u (y X) 0 (y X)! log 2 u 1 C A = 0 k1 0! k The needed derivative rules are in the Appendix 3.A.3 (Ch. 3) of the textbook.

61 The solution of this problem leads us to ^ ML := 0 i=1 x i x 0 i 1 A ^ 2 u = 1 n 1 0 i=1 0 1 ^u 2 i i=1 1 x i y i A X 0 X A 1 n ^u0^u 1 X 0 y where ^u i = y i x 0 i^ ML, i = 1; :::; n or, alternatively, ^u = (y X ^ML ): Recall that by contruction, s n (^ML ; ^ 2 u) = 0 k1 and H n (^ML ; ^ 2 u) is negative semidenite.

62 Constrained estimation Given y = X + u with u j X N(0 n1 ; 2 ui n ) general (linear) restrictions on can be represented either in implicit form or in explicit form. Implicit form: or, alternatively R qk k1 := r q1 R qk k1 where q is the # of restrictions; r q1 := 0 q1

63 R qk k1 r q1 := 0 q1 R is a known q k matrix whose rows select the elements of that must be restricted; r is a known q 1 vector. EXAMPLES: BLACKBOARD!

64 Explicit form: where a = (k := H k1 ka q), ' a1 + h k1 H is a known k a selection matrix; ' is the a 1 vector that contains the elements of which are not subject to restrictions (free parameters); h is a known k 1 selection vector. EXAMPLES: BLACKBOARD!

65 One can write the linear contraints on either in implicit or in explicit form. The two methods are alternative but equivalent. For certain problem it is convenient the implicit form representation, for others it is convenient the explicit form representation. It is therefore clear that there must exist a connection between the matrices R and H and r and h:

66 In particular, take the explicit form :=H' + h and multiply both sides by R, obtaining R:=RH' + Rh: Then, since the right-hand-side of the expression above must be equal to r, it must hold RH:=0 qa Rh:=r: This means that the (rows of the) matrix R lie in the null column space of the matrix H (sp(r) sp(h? )).

67 Recall: Let A a n m matrix, n > m, with (column) rank m; A = [a 1 : a 2 : :::: : a m ]. sp(a) = fall linear combination of a 1 ; a 2 ; :::; a m g R n sp(a? ) = n v, v 0 a i = 0; for each i = 1; :::; m o R n all the vectors in sp(a? ) are said to lie in the null column space of the matrix A: Let A? be a n(n m) full column rank matrix that satises A 0? A = 0 (n m)m A 0 A? = 0 m(n m) : A? is called orthgonal complement of A. The n n matrix [A : A? ] has rank n and forms a basis of R n.

68 Constrained estimation means that one stimates the model under H 0, i.e. imposing the restriction implied by the null hypothesis. Suppose that H 0 :R:=r, (H 0 ::=H' + h) has been accepted by the data. Then we wish to re-estimate the null model, i.e. the econometric model that embodies the linear restriction. In this case, using the restrictions in explicit form is convenient.

69 Indeed, the null model can be written as and re-arranged as y = X[H' + h] + u where y n1 = X na ' a1 + u n1 y :=y Xh X :=XH: We obtain a transformed dynamic linear regression model whose parameters are no longer in = ( 0 ; 2 u) 0 but in = (' 0 ; 2 u )0 : Recall that ' is the a1 vector, a = k q, that contains the unrestricted parameters, i.e. those parameters which are not aected by H 0.

70 We know very well how to estimate ' and 2 u and their properties: ^' = (X 0 X ) 1 X 0 y = (H 0 X 0 XH) 1 H 0 X 0 (y Xh) where ^' = ^' CML or ^' = ^' CLS ; ~ 2 u = 1 n a (y X ^') 0 (y X ^'):

71 The constrained ML or LS estimators of is obtained indirectly as ~ = ^CML = H ^' CML + h ~ = ^CLS = H ^' CLS + h and inherit the same properties as ^' CML.

72 Obviously, the asymptotic covariance matrix of the restricted estimator ~ will be V ar(~ j X):=V ar(h ^'CLS + h j X) =HV ar(^' CLS j X)H 0 =H(X 0 X ) 1 H 0 =H(H 0 (X 0 X)H) 1 H 0 : It can be proved that H(H 0 (X 0 X)H) 1 H 0 2 u(x 0 X) 1 regardless of whether H 0 is true or not! This means that imposing restrictions on has the effect of reducing the variability of the estimator (thus the standard errors of the single coecients). Of course, this operation is `correct' only when H 0 is uspported by the data.

73 In the `traditional' textbook econometrics the derivation of constrained estimator is usually obtained by exploiting the restriction in implicit form, by solving either or min max ( ( Q() subject to R r = 0 log L() subject to R r:=0 :

74 In order to solve these problems it is necessary to apply a technique based on Lagrange multipliers. For instance, in the case of ML estimation, the problem above amounts to solving where max () (; ) = log L() + 0 (R r) 1q q1 is the Lagrangean function and is the vector of Lagrange multipliers. Each element of gives a weight to the corresponding linear restriciton. By using the restrictions in explicit form we stick to the standard framework without the need of using constrained estimation techniques. More specically, we have turned a restricted estimation problem into an unrestricted estimation problem (' is estimated unrestrictedly!)

75 Testing problem Given y t = x 0 i + u i, i = 1; :::; n or objective is testing H 0 : R:=r, (H 0 : :=H' + h) vs H 1 : R 6= r, (H 1 : 6= H' + h). A testing problem requires solving a probabilistic decision rule that allows the researcher establishing whether the evidence provided by the data is closer to H 0 or to H 1.

76 This decision is based on the following ingredients: a Pre-xed nominal type I error (or signicance level) = Pr(rejecting H 0 j H 0 ): The cases we typically address are such that is actually an asymptotic error in the sense that Pr(reject H 0 j H 0 ) can be evaluated asymptotically (for large n); b Test statistic ^G n = G n (^) which is a function of ^ and whose distribution is known under H 0 asymptotically (later on we see that ^G n = G n (^) usually belongs to one of three families); c Decision rule of the type: reject H 0 if G n (^) > cv(1 ), where cv (1 ) is the 100*(1 ) percentile taken from the (asymptotic) distribution of ^G n = G n (^) under H 0 ;

77 d To understand whether the test does a good job one should be able to evaluate the function 1 () = lim n!1 Pr( reject H 0 j H 1 ) = Pr(G 1 (^) cv(1 ) j H 1 ) known as asymptotic power function. A`desired' test should be such that once one xes the type I error (usually 5% or 10%), the test is consistent meaning that 1 () = lim n!1 Pr(reject H 0 j H 1 ) = Pr(G 1 (^) cv(1 ) j H 1 )! 1:

78 When estimation is perfomed with ML we can classify the test statistic ^G n = G n (^) into three families, known as Wald, Lagrange Multipliers (LM) and Likelihood Ratio (LR) tests. However, the existence of these three families is not solely conned to ML estimation but can be extended to other classes of estimators as well. General philosophy: Wald These type of tests are based on the idea of checking whether the unrestricted (ML) estimator of (i.e. the estimator obtained without imposing any constraint) is `close in a statistical sense' to H 0, namely whether R^ r 0q1 which is what one would expect if H 0 is true. Computing a Wald-type test requires estimating the regression model one time without any restriction.

79 LM These type of tests are based on the the estimation of the null model (i.e. the regression model under H 0 ), hence are based on the constrained estimator of. Let ^ be the unconstrained (obtained under H 1 ) estimator of and ~ its constrained (obtained under H 0 ) counterpart. The idea is that if H 0 is true, the score of the unrestricted model, evaluated at the constrained estimator (recall that s n (^) = 0k1 ) should be `close to zero in a statistical sense', i.e. s n (~ ) 0k1 : Computing a LM test requires estimating the regression model one time under the restrictions implied by H 0 (however, it also requires that we know the structure of the score of the unrestricted model!).

80 LR The type of tests are based on the idea of estimating the linear regression model both under H 0 (~ ) and without restrictions (^); estimation is carried out two times. If H 0 is true, the distance between the likelihood functions of the null and unrestricted models should not be too large (note that one always has log L(~ ) log L(^)).

81 Recall: Let v a p 1 stochastic vector such that v N(0 p1 ; V ), where V is a p p covariance matrix (hence symmetric positive denite). Then v 0 V 1 v 2 (p): The same result holds if ` ' is replaced with `! D '. Note also that kvk 2 = (v 0 v) 1=2 is the Euclidean norm of the vector v (a measure of its length in the space R p, i.e. a measure of the distance of the vector v from the vector 0 p1 ). One can generalize this measure by dening the norm kvk A = (v 0 Av) 1=2, where A is a symmetric positive denite matrix; this norm measures the distance of v from 0 p1 `weighted' by the elements of the matrix A. This means that the random variable (quadratic form) v 0 V 1 v measures a weighted distance of v from 0 p1.

82 Wald test Let ^ be either the OLS or the ML estimator of in the classical regression model with corss-section data. Standard Assumptions. We start from the assumption (Gaussian) which implies that u j X N(0 n ; 2 ui n ): ^ j X N(; 2 u(x 0 X) 1 ): From the properties of the multivariate Gaussian distribution we obtain R(^ ) q1 j X N(0 q ; 2 ur(x 0 X) 1 R 0 ):

83 The result above implies that under H 0 :R:=r : R^ q1 r j X N(0 q1 ; 2 u R(X0 X) 1 R 0 ): Interpretation: the left-hand side can be regarded as a measure of the distance of the q 1 (stochastic) vector R^ from the known vector r (or the distance of R^ r from the vector 0q1 ); moreover, we have a symmetric positive denite matrix ( 2 ur(x 0 X) 1 R 0 ) with respect to which we can weight this distance.

84 Imagine rst that 2 u is known (unrealistic). We dene the test statistic (a quadratic form from Gaussian) W n = h (R^ r) i 0 n 2 u R(X 0 X) 1 R 0o 1 h (R^ r) i : From the properties of multivariate Gaussian distribution: W n j X 2 (q): Thus, xed, one rejects H 0 if W n cv (1 ), where cv (1 ) is the 100*(1-) percentile taken from the 2 (q) distribution (Pr[ 2 (q) > cv (1 ) ] = ). One accepts H 0 otherwise.

85 Since 2 u is generally unknown, we replace it with the estimator ^ 2 u. Please recall that we know that under our assumptions: ^G j X 2 (n k), ^G:=^2 u 2 (n k): u Recall also that from the theory of distributions, if G 1 2 (q 1 ) G 2 2 (q 2 ) and G 1 and G 2 are independent then G1 q 1 G2 q 2 F (q 1 ; q 2 ):

86 Now dene the test statistic W n = h (R^ r) i 0 n^ 2 u R(X 0 X) 1 R 0o 1 h (R^ r) i =q h (R^ r) i 0 n R(X 0 X) 1 R 0o 1 h (R^ r) i = q^ 2 u h 0fR(X (R^ r)i 0 X) 1 R 0 g 1h i (R^ r) = = ^ 2 u 2 u q 2 u n n h 0fR(X (R^ r)i 0 X) 1 R 0 g 1h i (R^ r) q 2 u ^G n k k k It can be proved that the two quadratic forms above and below the fraction are independent.

87 Accordingly, W n j X = h 0fR(X (R^ r)i 0 X) 1 R 0 g 1h i (R^ r) q 2 u ^G n k j X 2 (q) q 2 (n k) n k F (q; n k): Thus, xed, one rejects H 0 if W n cv (1 ), where cv (1 ) is the 100*(1-) percentile taken from the F (q; n k) distribution (Pr[ 2 (q) > cv (1 ) ] = ). One accepts H 0 otherwise. General linear hypotheses on can be tested by using the F-distribution. This is true if u j X N(0 n ; 2 u I n) irrespective of whether n is small or large. Note that F (q; n k)! 2 (q)=q for n! 1:

88 LM and LR tests will be reviewed when we deal with the regression model based on time series data.

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance V (u i x i ) = σ 2 is common to all observations i = 1,..., In many applications, we may suspect

More information

ECONOMETRIC MODELS. The concept of Data Generating Process (DGP) and its relationships with the analysis of specication.

ECONOMETRIC MODELS. The concept of Data Generating Process (DGP) and its relationships with the analysis of specication. ECONOMETRIC MODELS The concept of Data Generating Process (DGP) and its relationships with the analysis of specication. Luca Fanelli University of Bologna luca.fanelli@unibo.it The concept of Data Generating

More information

the error term could vary over the observations, in ways that are related

the error term could vary over the observations, in ways that are related Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance Var(u i x i ) = σ 2 is common to all observations i = 1,..., n In many applications, we may

More information

Heteroskedasticity and Autocorrelation

Heteroskedasticity and Autocorrelation Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Testing Linear Restrictions: cont.

Testing Linear Restrictions: cont. Testing Linear Restrictions: cont. The F-statistic is closely connected with the R of the regression. In fact, if we are testing q linear restriction, can write the F-stastic as F = (R u R r)=q ( R u)=(n

More information

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley Models, Testing, and Correction of Heteroskedasticity James L. Powell Department of Economics University of California, Berkeley Aitken s GLS and Weighted LS The Generalized Classical Regression Model

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

Maximum Likelihood (ML) Estimation

Maximum Likelihood (ML) Estimation Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear

More information

GMM estimation of spatial panels

GMM estimation of spatial panels MRA Munich ersonal ReEc Archive GMM estimation of spatial panels Francesco Moscone and Elisa Tosetti Brunel University 7. April 009 Online at http://mpra.ub.uni-muenchen.de/637/ MRA aper No. 637, posted

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Simple Linear Regression: The Model

Simple Linear Regression: The Model Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random

More information

The outline for Unit 3

The outline for Unit 3 The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information

Spatial Regression. 9. Specification Tests (1) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Spatial Regression. 9. Specification Tests (1) Luc Anselin.   Copyright 2017 by Luc Anselin, All Rights Reserved Spatial Regression 9. Specification Tests (1) Luc Anselin http://spatial.uchicago.edu 1 basic concepts types of tests Moran s I classic ML-based tests LM tests 2 Basic Concepts 3 The Logic of Specification

More information

ECONOMETRICS. Bruce E. Hansen. c2000, 2001, 2002, 2003, University of Wisconsin

ECONOMETRICS. Bruce E. Hansen. c2000, 2001, 2002, 2003, University of Wisconsin ECONOMETRICS Bruce E. Hansen c2000, 200, 2002, 2003, 2004 University of Wisconsin www.ssc.wisc.edu/~bhansen Revised: January 2004 Comments Welcome This manuscript may be printed and reproduced for individual

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

The Linear Regression Model

The Linear Regression Model The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general

More information

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56 Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The

More information

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE Chapter 6. Panel Data Joan Llull Quantitative Statistical Methods II Barcelona GSE Introduction Chapter 6. Panel Data 2 Panel data The term panel data refers to data sets with repeated observations over

More information

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015 EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,

More information

Heteroskedasticity-Robust Inference in Finite Samples

Heteroskedasticity-Robust Inference in Finite Samples Heteroskedasticity-Robust Inference in Finite Samples Jerry Hausman and Christopher Palmer Massachusetts Institute of Technology December 011 Abstract Since the advent of heteroskedasticity-robust standard

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

Exercises Chapter 4 Statistical Hypothesis Testing

Exercises Chapter 4 Statistical Hypothesis Testing Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

Reduced rank regression in cointegrated models

Reduced rank regression in cointegrated models Journal of Econometrics 06 (2002) 203 26 www.elsevier.com/locate/econbase Reduced rank regression in cointegrated models.w. Anderson Department of Statistics, Stanford University, Stanford, CA 94305-4065,

More information

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria SOLUTION TO FINAL EXAM Friday, April 12, 2013. From 9:00-12:00 (3 hours) INSTRUCTIONS:

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

The Equality of OLS and GLS Estimators in the

The Equality of OLS and GLS Estimators in the The Equality of OLS and GLS Estimators in the Linear Regression Model When the Disturbances are Spatially Correlated Butte Gotu 1 Department of Statistics, University of Dortmund Vogelpothsweg 87, 44221

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 11, 2012 Outline Heteroskedasticity

More information

p(z)

p(z) Chapter Statistics. Introduction This lecture is a quick review of basic statistical concepts; probabilities, mean, variance, covariance, correlation, linear regression, probability density functions and

More information

Scaled and adjusted restricted tests in. multi-sample analysis of moment structures. Albert Satorra. Universitat Pompeu Fabra.

Scaled and adjusted restricted tests in. multi-sample analysis of moment structures. Albert Satorra. Universitat Pompeu Fabra. Scaled and adjusted restricted tests in multi-sample analysis of moment structures Albert Satorra Universitat Pompeu Fabra July 15, 1999 The author is grateful to Peter Bentler and Bengt Muthen for their

More information

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Introduction For pedagogical reasons, OLS is presented initially under strong simplifying assumptions. One of these is homoskedastic errors,

More information

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin Sensitivity Analysis in LP and SDP Using Interior-Point Methods E. Alper Yldrm School of Operations Research and Industrial Engineering Cornell University Ithaca, NY joint with Michael J. Todd INFORMS

More information

Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley

Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley Zellner s Seemingly Unrelated Regressions Model James L. Powell Department of Economics University of California, Berkeley Overview The seemingly unrelated regressions (SUR) model, proposed by Zellner,

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations. Exercises for the course of Econometrics Introduction 1. () A researcher is using data for a sample of 30 observations to investigate the relationship between some dependent variable y i and independent

More information

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u Interval estimation and hypothesis tests So far our focus has been on estimation of the parameter vector β in the linear model y i = β 1 x 1i + β 2 x 2i +... + β K x Ki + u i = x iβ + u i for i = 1, 2,...,

More information

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econ 510 B. Brown Spring 2014 Final Exam Answers Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity

More information

Robust linear optimization under general norms

Robust linear optimization under general norms Operations Research Letters 3 (004) 50 56 Operations Research Letters www.elsevier.com/locate/dsw Robust linear optimization under general norms Dimitris Bertsimas a; ;, Dessislava Pachamanova b, Melvyn

More information

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of Wooldridge, Introductory Econometrics, d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of homoskedasticity of the regression error term: that its

More information

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

1 The Multiple Regression Model: Freeing Up the Classical Assumptions 1 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions were crucial for many of the derivations of the previous chapters. Derivation of the OLS estimator

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 17, 2012 Outline Heteroskedasticity

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Chapter 1. GMM: Basic Concepts

Chapter 1. GMM: Basic Concepts Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Introduction to Econometrics. Heteroskedasticity

Introduction to Econometrics. Heteroskedasticity Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random

More information

Practical Econometrics. for. Finance and Economics. (Econometrics 2)

Practical Econometrics. for. Finance and Economics. (Econometrics 2) Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics

More information

New Introduction to Multiple Time Series Analysis

New Introduction to Multiple Time Series Analysis Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2

More information

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables,

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables, University of Illinois Fall 998 Department of Economics Roger Koenker Economics 472 Lecture Introduction to Dynamic Simultaneous Equation Models In this lecture we will introduce some simple dynamic simultaneous

More information

Notes on Time Series Modeling

Notes on Time Series Modeling Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g

More information

8. Hypothesis Testing

8. Hypothesis Testing FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements

More information

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0:

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0: Linear Regression 1 Single Explanatory Variable Assume (y is not necessarily normal) where Examples: y» F; Ey = + x Vary = ¾ 2 ) y = + x + u Eu = 0 Varu = ¾ 2 Exu = 0: 1. School performance as a function

More information

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

Linear Models 1. Isfahan University of Technology Fall Semester, 2014 Linear Models 1 Isfahan University of Technology Fall Semester, 2014 References: [1] G. A. F., Seber and A. J. Lee (2003). Linear Regression Analysis (2nd ed.). Hoboken, NJ: Wiley. [2] A. C. Rencher and

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Appendix A: The time series behavior of employment growth

Appendix A: The time series behavior of employment growth Unpublished appendices from The Relationship between Firm Size and Firm Growth in the U.S. Manufacturing Sector Bronwyn H. Hall Journal of Industrial Economics 35 (June 987): 583-606. Appendix A: The time

More information

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Tak Wai Chau February 20, 2014 Abstract This paper investigates the nite sample performance of a minimum distance estimator

More information

1. The Multivariate Classical Linear Regression Model

1. The Multivariate Classical Linear Regression Model Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The

More information

LECTURE 5 HYPOTHESIS TESTING

LECTURE 5 HYPOTHESIS TESTING October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.

More information

Matrix Differential Calculus with Applications in Statistics and Econometrics

Matrix Differential Calculus with Applications in Statistics and Econometrics Matrix Differential Calculus with Applications in Statistics and Econometrics Revised Edition JAN. R. MAGNUS CentERjor Economic Research, Tilburg University and HEINZ NEUDECKER Cesaro, Schagen JOHN WILEY

More information

Discrete Dependent Variable Models

Discrete Dependent Variable Models Discrete Dependent Variable Models James J. Heckman University of Chicago This draft, April 10, 2006 Here s the general approach of this lecture: Economic model Decision rule (e.g. utility maximization)

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

Likelihood Ratio Based Test for the Exogeneity and the Relevance of Instrumental Variables

Likelihood Ratio Based Test for the Exogeneity and the Relevance of Instrumental Variables Likelihood Ratio Based est for the Exogeneity and the Relevance of Instrumental Variables Dukpa Kim y Yoonseok Lee z September [under revision] Abstract his paper develops a test for the exogeneity and

More information

Econometrics II - EXAM Answer each question in separate sheets in three hours

Econometrics II - EXAM Answer each question in separate sheets in three hours Econometrics II - EXAM Answer each question in separate sheets in three hours. Let u and u be jointly Gaussian and independent of z in all the equations. a Investigate the identification of the following

More information

Econometrics I. Ricardo Mora

Econometrics I. Ricardo Mora Econometrics I Department of Economics Universidad Carlos III de Madrid Master in Industrial Economics and Markets Outline Motivation 1 Motivation 2 3 4 Motivation The Analogy Principle The () is a framework

More information

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in

More information

Financial Econometrics

Financial Econometrics Material : solution Class : Teacher(s) : zacharias psaradakis, marian vavra Example 1.1: Consider the linear regression model y Xβ + u, (1) where y is a (n 1) vector of observations on the dependent variable,

More information

A Course on Advanced Econometrics

A Course on Advanced Econometrics A Course on Advanced Econometrics Yongmiao Hong The Ernest S. Liu Professor of Economics & International Studies Cornell University Course Introduction: Modern economies are full of uncertainties and risk.

More information

1. The OLS Estimator. 1.1 Population model and notation

1. The OLS Estimator. 1.1 Population model and notation 1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

Lecture 1: OLS derivations and inference

Lecture 1: OLS derivations and inference Lecture 1: OLS derivations and inference Econometric Methods Warsaw School of Economics (1) OLS 1 / 43 Outline 1 Introduction Course information Econometrics: a reminder Preliminary data exploration 2

More information

Heteroskedasticity. Part VII. Heteroskedasticity

Heteroskedasticity. Part VII. Heteroskedasticity Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least

More information

The Standard Linear Model: Hypothesis Testing

The Standard Linear Model: Hypothesis Testing Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 25: The Standard Linear Model: Hypothesis Testing Relevant textbook passages: Larsen Marx [4]:

More information

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error

More information

Analysis of Cross-Sectional Data

Analysis of Cross-Sectional Data Analysis of Cross-Sectional Data Kevin Sheppard http://www.kevinsheppard.com Oxford MFE This version: November 8, 2017 November 13 14, 2017 Outline Econometric models Specification that can be analyzed

More information

Tests forspatial Lag Dependence Based onmethodof Moments Estimation

Tests forspatial Lag Dependence Based onmethodof Moments Estimation 1 Tests forspatial Lag Dependence Based onmethodof Moments Estimation by Luz A. Saavedra* Department of Economics University of South Florida 4202 East Fowler Ave. BSN 3111 Tampa, FL 33620-5524 lsaavedr@coba.usf.edu

More information

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle

More information

Y t = ΦD t + Π 1 Y t Π p Y t p + ε t, D t = deterministic terms

Y t = ΦD t + Π 1 Y t Π p Y t p + ε t, D t = deterministic terms VAR Models and Cointegration The Granger representation theorem links cointegration to error correction models. In a series of important papers and in a marvelous textbook, Soren Johansen firmly roots

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

(c) i) In ation (INFL) is regressed on the unemployment rate (UNR):

(c) i) In ation (INFL) is regressed on the unemployment rate (UNR): BRUNEL UNIVERSITY Master of Science Degree examination Test Exam Paper 005-006 EC500: Modelling Financial Decisions and Markets EC5030: Introduction to Quantitative methods Model Answers. COMPULSORY (a)

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

Reliability of inference (1 of 2 lectures)

Reliability of inference (1 of 2 lectures) Reliability of inference (1 of 2 lectures) Ragnar Nymoen University of Oslo 5 March 2013 1 / 19 This lecture (#13 and 14): I The optimality of the OLS estimators and tests depend on the assumptions of

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

4.8 Instrumental Variables

4.8 Instrumental Variables 4.8. INSTRUMENTAL VARIABLES 35 4.8 Instrumental Variables A major complication that is emphasized in microeconometrics is the possibility of inconsistent parameter estimation due to endogenous regressors.

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information