CHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model

Size: px
Start display at page:

Download "CHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model"

Transcription

1 CHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model Prof. Alan Wan 1 / 57

2 Table of contents 1. Assumptions in the Linear Regression Model 2 / 57

3 Assumptions The validity and properties of least squares estimation depend very much on the validity of the classical assumptions underlying the regression model. As we shall see, many of these assumptions are rarely appropriate when dealing with data for business. However, they represent a useful starting point dealing with the inferential aspects of the regression and for the development of more advanced techniques. The assumptions are as follows: 3 / 57

4 Assumptions 1. The regression model is linear in the unknown parameters. 4 / 57

5 Assumptions 1. The regression model is linear in the unknown parameters. 2. The elements in X are non-stochastic, meaning that the values of X are fixed in repeated samples (i.e., when repeating the experiment, choose exactly the same set of X values on each occasion so that they remain unchanged). Notice, however, this does not imply that the values of Y also remain unchanged from sample to sample. The Y values depend also on the uncontrollable values of ɛ, which vary from one sample to another. Y as well as ɛ are therefore stochastic, meaning that their values are determined by some chance mechanism and hence subject to a probability distribution. 4 / 57

6 Assumptions 1. The regression model is linear in the unknown parameters. 2. The elements in X are non-stochastic, meaning that the values of X are fixed in repeated samples (i.e., when repeating the experiment, choose exactly the same set of X values on each occasion so that they remain unchanged). Notice, however, this does not imply that the values of Y also remain unchanged from sample to sample. The Y values depend also on the uncontrollable values of ɛ, which vary from one sample to another. Y as well as ɛ are therefore stochastic, meaning that their values are determined by some chance mechanism and hence subject to a probability distribution. Essentially this means our regression analysis is conditional on the given values of the regressors. It is possible to weaken the assumption to one of stochastic X distributed independently of the disturbance term. 4 / 57

7 Assumptions 3. Zero mean value of the disturbance ɛ i, i.e., E(ɛ i ) = 0, i, or in matrix terms, ɛ 1 0 ɛ 2 E(ɛ) = E. = 0. ɛ n 0 = 0 leading to E(Y ) = X β The zero mean of the disturbances implies that no relevant regressors have been omitted from the model. 5 / 57

8 Assumptions 4. The variance-covariance matrix of ɛ is a scalar matrix. That is, ɛ 1 E(ɛɛ ɛ 2 [ ] ) = E( ɛ1 ɛ 2 ɛ n ). = ɛ n E(ɛ 2 1 ) E(ɛ 1ɛ 2 ) E(ɛ 1 ɛ n ) σ E(ɛ 2 ɛ 1 ) E(ɛ 2 2 ) E(ɛ 2ɛ n ) = 0 σ E(ɛ n ɛ 1 ) E(ɛ n ɛ 2 ) E(ɛ 2 n) 0 0 σ 2 = σ 2 I. 6 / 57

9 Assumptions This variance covariance matrix embodies two assumptions: var(ɛ i ) = σ 2 i. This assumption is termed homoscedasticity (the converse is heteroscedasticity). 7 / 57

10 Assumptions This variance covariance matrix embodies two assumptions: var(ɛ i ) = σ 2 i. This assumption is termed homoscedasticity (the converse is heteroscedasticity). cov(ɛ i ɛ j ) = 0 i j. This assumption is termed pairwise uncorrelatedness (the coverse is serial correlation or autocorrelation). 7 / 57

11 Assumptions 5. ρ(x ) = rank(x ) = k < n. In other words, the explanatory variables do not form a linear dependent set as X is n k. We say that X has full column rank. If this conditions fails, then X X cannot be inverted and O.L.S. estimation becomes infeasible. This problem is known as perfect multicollinearity. 8 / 57

12 Assumptions 5. ρ(x ) = rank(x ) = k < n. In other words, the explanatory variables do not form a linear dependent set as X is n k. We say that X has full column rank. If this conditions fails, then X X cannot be inverted and O.L.S. estimation becomes infeasible. This problem is known as perfect multicollinearity. 6. As n, n i=1 (x ij x j ) 2 /n Q j, where Q j is finite, j = 1,, k. 8 / 57

13 Properties of O.L.S. When some or all of the above assumptions are satisfied, the O.L.S. estimator b of β possesses the following properties. Note that not every property requires all of the above assumptions to be fulfilled. Properties of the O.L.S. estimator: 9 / 57

14 Properties of O.L.S. When some or all of the above assumptions are satisfied, the O.L.S. estimator b of β possesses the following properties. Note that not every property requires all of the above assumptions to be fulfilled. Properties of the O.L.S. estimator: b is a linear estimator in the sense that it is a linear combination of the observations of Y : b = (X X ) 1 X Y c 11 c 12 c 1n y 1 c 21 c 22 c 2n y 2 = c k1 c k2 c kn y n 9 / 57

15 Properties of O.L.S. Unbiasedness E(b) = E((X X ) 1 X Y ) = E(β + (X X ) 1 X ɛ) = β + (X X ) 1 X E(ɛ) = β 10 / 57

16 Properties of O.L.S. Unbiasedness E(b) = E((X X ) 1 X Y ) = E(β + (X X ) 1 X ɛ) = β + (X X ) 1 X E(ɛ) = β Thus, b is an unbiased estimator of β. That is, in repeated samples, b has an average value identical to β, the parameter b tries to estimate. 10 / 57

17 Properties of O.L.S. Variance-Covariance matrix: COV (b) = E((b E(b))(b E(b)) ) = E((b β)(b β) ) = E((X X ) 1 X ɛɛ X (X X ) 1 ) = (X X ) 1 X E(ɛɛ )X (X X ) 1 = σ 2 (X X ) 1 11 / 57

18 Properties of O.L.S. Variance-Covariance matrix: COV (b) = E((b E(b))(b E(b)) ) = E((b β)(b β) ) = E((X X ) 1 X ɛɛ X (X X ) 1 ) = (X X ) 1 X E(ɛɛ )X (X X ) 1 = σ 2 (X X ) 1 Main diagonal elements are the variances of b j s, j = 1,, k; off-diagonal elements are covariances. For the special case of a simple linear regression, [ ] [ b1 σ 2 ( 1 n Cov( ) = + x 2 n i=1 (x ) σ 2 x ] i x) 2 n i=1 (x i x) 2 b 2 σ 2 x σ 2 n i=1 (x i x) 2 n i=1 (x i x) 2 11 / 57

19 Properties of O.L.S. b is the best linear unbiased (B.L.U.) estimator of β. Refer to the Gauss-Markov theorem. The B.L.U. properties implies that each b j, j = 1,, k, has the smallest variance among the class of all linear unbiased estimators of β j. More discussion in class. 12 / 57

20 Properties of O.L.S. b is the best linear unbiased (B.L.U.) estimator of β. Refer to the Gauss-Markov theorem. The B.L.U. properties implies that each b j, j = 1,, k, has the smallest variance among the class of all linear unbiased estimators of β j. More discussion in class. b is the minimum variance unbiased (M.V.U.) estimator of β, meaning that b j has a variance no larger than that of any unbiased estimator of β j, linear or non-linear. More discussion in class. 12 / 57

21 Properties of O.L.S. b is the best linear unbiased (B.L.U.) estimator of β. Refer to the Gauss-Markov theorem. The B.L.U. properties implies that each b j, j = 1,, k, has the smallest variance among the class of all linear unbiased estimators of β j. More discussion in class. b is the minimum variance unbiased (M.V.U.) estimator of β, meaning that b j has a variance no larger than that of any unbiased estimator of β j, linear or non-linear. More discussion in class. b is a consistent estimator of β, meaning that when n becomes sufficiently large, the probability of b j = β j converges to 1, j = 1,, k. We say that b converges in probability to the true value of β. More discussion in class. 12 / 57

22 Matters of Inference If one assumes additionally that ɛ MVN(0, σ 2 I ), then Y MVN(X β, σ 2 I ) b MVN(β, σ 2 (X X ) 1 ) Using properties of the sampling distribution of b, inference about the population parameters in β can be drawn. 13 / 57

23 Matters of Inference However, we need an estimator of σ 2, the variance around the regression line. This estimator is given by s 2 = e e n n k = i=1 (y i ŷ i ) 2, n k where n k is the model s degrees of freedom (d.o.f.) - the number of logically independent pieces of information in the data. 14 / 57

24 Matters of Inference However, we need an estimator of σ 2, the variance around the regression line. This estimator is given by s 2 = e e n n k = i=1 (y i ŷ i ) 2, n k where n k is the model s degrees of freedom (d.o.f.) - the number of logically independent pieces of information in the data. It can be shown that e e/σ 2 χ 2 (n k), or (n k)s 2 /σ 2 χ 2 (n k). 14 / 57

25 Matters of Inference However, we need an estimator of σ 2, the variance around the regression line. This estimator is given by s 2 = e e n n k = i=1 (y i ŷ i ) 2, n k where n k is the model s degrees of freedom (d.o.f.) - the number of logically independent pieces of information in the data. It can be shown that e e/σ 2 χ 2 (n k), or (n k)s 2 /σ 2 χ 2 (n k). Using the properties of the Chi-square distribution, it can be shown that E(s 2 ) = σ 2, i.e., s 2 is an unbiased estimator of σ / 57

26 Matters of Inference The quantities b j, j = 1,, k, are simply point estimates (single numbers). Often it is more desirable to state a range of values in which the parameter is thought to lie rather than a single number. These ranges are called confidence interval (C.I.) estimates. 15 / 57

27 Matters of Inference The quantities b j, j = 1,, k, are simply point estimates (single numbers). Often it is more desirable to state a range of values in which the parameter is thought to lie rather than a single number. These ranges are called confidence interval (C.I.) estimates. Although the interval estimate is less precise, the confidence that the true population parameters falls between the interval limits is increased. The interval should be precise enough to be practically useful. 15 / 57

28 Matters of Inference Consider a coefficient β j in β. The interval L β j U is a 100(1 α) % confidence interval for β j in the sense that, prior to sampling, P(L β j U) = 1 α This definition states that the C.I. with confidence coefficient 1 α is an interval estimate such that the probability is 1 α that the calculated limits include β j for any random trial. That is, in many random samples of size n, 100(1 α) percent of the interval estimates will include β j. 16 / 57

29 Matters of Inference Recall that b N(β, σ 2 (X X ) 1 ). Hence b j β j σc jj N(0, 1), where c 2 jj is the jj th element of (X X ) / 57

30 Matters of Inference Recall that b N(β, σ 2 (X X ) 1 ). Hence b j β j σc jj N(0, 1), where c 2 jj is the jj th element of (X X ) 1. Hence P(z (α/2) b j β j σc jj z (1 α/2) ) = 1 α Recognising that z(α/2) = z(1 α/2) and after some manipulations, we can write P(b j z (1 α/2) σc jj β j b j + z (1 α/2) σc jj ) = 1 α 17 / 57

31 Matters of Inference However, σ 2 is typically unknown. Replacing σ 2 by s 2 results in b j β j sc jj t (n k). 18 / 57

32 Matters of Inference However, σ 2 is typically unknown. Replacing σ 2 by s 2 results in b j β j sc jj t (n k). If Z N(0, 1) and W χ 2 (n k) and Z and W are Z independently distributed, then t (n k). More W /(n k) discussion on t distribution in class. 18 / 57

33 Matters of Inference However, σ 2 is typically unknown. Replacing σ 2 by s 2 results in b j β j sc jj t (n k). If Z N(0, 1) and W χ 2 (n k) and Z and W are Z independently distributed, then t (n k). More W /(n k) discussion on t distribution in class. Hence the confidence interval becomes P(b j t (1 α/2,n k) sc jj β j b j + t (1 α/2,n k) sc jj ) = 1 α 18 / 57

34 Matters of Inference Hypothesis tests about β j can also be performed. The most common test about a coefficient in a regression is: H 0 : β j = β j vs. H 1 : β j β j at a significance level α, the probability of rejecting H 0 when H 0 is correct, the so-called Type I error. 19 / 57

35 Matters of Inference Hypothesis tests about β j can also be performed. The most common test about a coefficient in a regression is: H 0 : β j = β j vs. H 1 : β j β j at a significance level α, the probability of rejecting H 0 when H 0 is correct, the so-called Type I error. To test this hypothesis, a t statistic is used: t = b j β j sc jj If H 0 is true then t has a t distribution with n k degrees of freedom. 19 / 57

36 Matters of Inference If H 0 is true, then t is expected to lie not too far from the centre of the distribution. The decision rule is: Reject H 0 if t > t (1 α/2,n k) or t < t (1 α/2,n k) Do not reject otherwise. 20 / 57

37 Matters of Inference If H 0 is true, then t is expected to lie not too far from the centre of the distribution. The decision rule is: Reject H 0 if t > t (1 α/2,n k) or t < t (1 α/2,n k) Do not reject otherwise. Testing this hypothesis is equivalent to asking if βj lies in the 100(1 α) percent C.I. of β j. It is a common practice to test the hypothesis of H 0 : β j = 0. Failure to reject this hypothesis would imply that β j is not significantly different from zero, or equivalently, X j has no significant impact on the behaviour of Y, at level of significance α. 20 / 57

38 Matters of Inference Return to Example 1.3, and consider the estimation of β 2. Back in Chapter 1, we already computed b 2 = The output shows that e e = 25 i=1 e2 i = Note that d.o.f. = 25 3 = 22. Hence s 2 = e e/22 = / 57

39 Matters of Inference Return to Example 1.3, and consider the estimation of β 2. Back in Chapter 1, we already computed b 2 = The output shows that e e = 25 i=1 e2 i = Note that d.o.f. = 25 3 = 22. Hence s 2 = e e/22 = From (X X ) 1, c22 2 = Hence s.e.(b 2 ) = = / 57

40 Matters of Inference Return to Example 1.3, and consider the estimation of β 2. Back in Chapter 1, we already computed b 2 = The output shows that e e = 25 i=1 e2 i = Note that d.o.f. = 25 3 = 22. Hence s 2 = e e/22 = From (X X ) 1, c22 2 = Hence s.e.(b 2 ) = = Set α = From the t distribution table, t (1 0.05/2,22) = Hence the 95 percent C.I. for β 2 is P( (2.074)(0.1721) β (2.074)(0.1721)) = 0.95 or P( β ) = / 57

41 Matters of Inference This C.I. contains 0, meaning that if we test H 0 : β 2 = 0 vs. H 1 : β 2 0, we would not be able to reject H 0 at α = This is indeed the case. 22 / 57

42 Matters of Inference This C.I. contains 0, meaning that if we test H 0 : β 2 = 0 vs. H 1 : β 2 0, we would not be able to reject H 0 at α = This is indeed the case. Note that for testing H 0, t = = 1.928, which lies to the left of t (1 0.05/2,22) = Hence H 0 cannot be rejected at α = / 57

43 Matters of Inference This C.I. contains 0, meaning that if we test H 0 : β 2 = 0 vs. H 1 : β 2 0, we would not be able to reject H 0 at α = This is indeed the case. Note that for testing H 0, t = = 1.928, which lies to the left of t (1 0.05/2,22) = Hence H 0 cannot be rejected at α = Alternatively, one can base the decision on the p-value, which is the probability of obtaining a value of t at least as extreme as the actual computed value if H 0 is true. In our example, the p-value is , meaning that P(t > or t < 1.928) = / 57

44 Matters of Inference The p value can be viewed as the minimum level of significance chosen for the test to result in a rejection of H 0. Thus, a decision rule using p-value may be stated as: Reject H 0 if p-value< α Do not reject H 0 if p-value α 23 / 57

45 Matters of Inference The p value can be viewed as the minimum level of significance chosen for the test to result in a rejection of H 0. Thus, a decision rule using p-value may be stated as: Reject H 0 if p-value< α Do not reject H 0 if p-value α In the last example, the p-value is Hence we cannot reject H 0 at α = On the other hand, H 0 can be rejected at any significance level at or higher than / 57

46 Matters of Inference The p value can be viewed as the minimum level of significance chosen for the test to result in a rejection of H 0. Thus, a decision rule using p-value may be stated as: Reject H 0 if p-value< α Do not reject H 0 if p-value α In the last example, the p-value is Hence we cannot reject H 0 at α = On the other hand, H 0 can be rejected at any significance level at or higher than Similarly, we can test H 0 : β 3 = 0 and conclude that H 0 is rejected at α = / 57

47 Matters of Inference Altogether, it means allowing for a 5% Type 1 risk, disposable income is not significant for explaining consumption but the total value of assets is significant. Note that if we conclude that β j = 0, it does not necessarily follow that X j is unrelated to Y. It simply means that, when the other explanatory variables are included in the model, the marginal contribution of X j to further improving the model s fit is negligible. 24 / 57

48 Matters of Inference Altogether, it means allowing for a 5% Type 1 risk, disposable income is not significant for explaining consumption but the total value of assets is significant. Note that if we conclude that β j = 0, it does not necessarily follow that X j is unrelated to Y. It simply means that, when the other explanatory variables are included in the model, the marginal contribution of X j to further improving the model s fit is negligible. Sometimes it also makes sense to conduct a hypothesis test for the intercept coefficient. This should be done only when there are data that span X = 0 or at least near X = 0, and the difference between Y equaling zero and not equaling zero when X = 0 is scientifically plausible and interesting. 24 / 57

49 Type 1 and Type 2 errors Rejecting H 0 when it is true is called a Type 1 error. Recall that if H 0 is true the probability that it will be (incorrectly) rejected is P(t > t (1 α/2,n k) ) + P(t < t (1 α/2,n k) ) = α. This is the significance level; by choosing α, we effectively determine the probability that the test will incorrectly reject a true hypothesis. 25 / 57

50 Type 1 and Type 2 errors Rejecting H 0 when it is true is called a Type 1 error. Recall that if H 0 is true the probability that it will be (incorrectly) rejected is P(t > t (1 α/2,n k) ) + P(t < t (1 α/2,n k) ) = α. This is the significance level; by choosing α, we effectively determine the probability that the test will incorrectly reject a true hypothesis. If H 0 is false and it is not rejected then a Type 2 error has been committed. While we can fix P(Type 1 error), the same control of Type 2 error is not possible. See the following diagram for an illustration for testing H 0 : β 2 = 50 with σ 2 known and var(b 2 ) = Suppose that β 2 is either 50 or Note that Type 2 error probability depends on the true value of β 2 which is unknown in practice. 25 / 57

51 Type 1 and Type 2 errors 26 / 57

52 Type 1 and Type 2 errors Most elementary texts define the power of the test as the probability of rejecting a false H 0, i.e., the probability of doing the right thing in the face of an incorrect H 0. By this definition, the power is equal to 1 minus the Type 2 error probability. Sometimes the power is simply defined as the probability of rejecting H 0. By this definition, α, the significance level, is a point on the power curve. 27 / 57

53 Type 1 and Type 2 errors Let H 0 : β 2 = β2. P(Type 2 error β 2 β2 ) = P(Not rejecting H 0 β 2 β2 ). Power(β 2 ) = P(rejecting H 0 β 2 ) A test is unbiased if Power(β 2 β 2 β2 ) P(Type 1 error). For a test where H 0 corresponds to a point in the parameter space (e.g., a two-sided t test), the significance level is a point on the power curve. For a test where H 0 corresponds to a region in the parameter space (e.g., a one-sided t test), the significance level is the maximum probability of committing a Type 1 error within the region defined by H 0, and P(Type 1 error) has a range of values with α being the maximum of the range. 28 / 57

54 Partitioning of Total Sum of Squares Analysis of variance (ANOVA) is a useful and flexible way of analysing the fit of the regression. To motivate, consider y i = ŷ i + e i y i ȳ = ŷ i ȳ + e i n (y i ȳ) 2 = n (ŷ i ȳ) 2 + i=1 i=1 n ei i=1 n (ŷ i ȳ)e i i=1 29 / 57

55 Partitioning of Total Sum of Squares Note that n (y i ȳ) 2 = Total Sum of Squares (TSS) i=1 n (ŷ i ȳ) 2 = Regression Sum of Squares (RSS) i=1 n ei 2 = Error Sum of Squares (ESS) i=1 n (ŷ i ȳ)e i = 0 (provided that there is an intercept) i=1 30 / 57

56 Coefficient of Determination Thus, or TSS = RSS + ESS R 2 = RSS TSS = 1 ESS TSS, which is the coefficient of determination. It measures the model s goodness of fit : the proportion of variability of the sample Y values that has been explained by the regression. Obviously, 0 R / 57

57 Partitioning of Degrees of Freedom TSS has n 1 d.o.f. because there are n deviations y i ȳ that enter into TSS, but one constraint on the deviations, namely, n i=1 (y i ȳ) = 0. So there are n 1 d.o.f. in the deviations. 32 / 57

58 Partitioning of Degrees of Freedom TSS has n 1 d.o.f. because there are n deviations y i ȳ that enter into TSS, but one constraint on the deviations, namely, n i=1 (y i ȳ) = 0. So there are n 1 d.o.f. in the deviations. ESS has n k d.o.f. because there are n residuals but k d.o.f. are lost due to k constraints on the e i s associated with estimating the β s. 32 / 57

59 Partitioning of Degrees of Freedom TSS has n 1 d.o.f. because there are n deviations y i ȳ that enter into TSS, but one constraint on the deviations, namely, n i=1 (y i ȳ) = 0. So there are n 1 d.o.f. in the deviations. ESS has n k d.o.f. because there are n residuals but k d.o.f. are lost due to k constraints on the e i s associated with estimating the β s. RSS has k 1 d.o.f. because the regression function contains k parameters but the deviations ŷ i ȳ are subject to the constraint that n i=1 (ŷ i ȳ) = / 57

60 Partitioning of Degrees of Freedom TSS has n 1 d.o.f. because there are n deviations y i ȳ that enter into TSS, but one constraint on the deviations, namely, n i=1 (y i ȳ) = 0. So there are n 1 d.o.f. in the deviations. ESS has n k d.o.f. because there are n residuals but k d.o.f. are lost due to k constraints on the e i s associated with estimating the β s. RSS has k 1 d.o.f. because the regression function contains k parameters but the deviations ŷ i ȳ are subject to the constraint that n i=1 (ŷ i ȳ) = 0. The d.o.f. add up: (n 1) = (n k) + (k 1) 32 / 57

61 Mean Squares A sum of squares divided by its associated d.o.f. is called a mean square. Mean Square Regression (MSR) = RSS/(k 1) Mean Square Error (MSE) = ESS/(n k) 33 / 57

62 Mean Squares A sum of squares divided by its associated d.o.f. is called a mean square. Mean Square Regression (MSR) = RSS/(k 1) Mean Square Error (MSE) = ESS/(n k) In Example 1.3, n=25, k=3, RSS= , ESS= , TSS= Hence R 2 = / = MSR = /2 = MSE = /22 = / 57

63 Overall Significance of the Model Frequently, one may wish to test whether or not there is a relationship between Y and the regression model constructed. It is a test of H 0 : β 2 = β 3 = = β k = 0 vs. H 1 : at least one of β js, (j = 2,, k), is non-zero. The test statistic is F = MSR MSE = RSS/(k 1) ESS/(n k) distributed as F (k 1,n k) if H 0 is true. 34 / 57

64 Overall Significance of the Model The decision rule is: To reject H 0 if F > F 1 α,k 1,n k or p-value< α; Not to reject H 0 otherwise. 35 / 57

65 Overall Significance of the Model The decision rule is: To reject H 0 if F > F 1 α,k 1,n k or p-value< α; Not to reject H 0 otherwise. F F (k 1,n k) because under H 0, RSS/σ 2 χ 2 (k 1), ESS/σ 2 χ 2 (n k), and RSS and ESS are distributed independently. 35 / 57

66 Overall Significance of the Model The decision rule is: To reject H 0 if F > F 1 α,k 1,n k or p-value< α; Not to reject H 0 otherwise. F F (k 1,n k) because under H 0, RSS/σ 2 χ 2 (k 1), ESS/σ 2 χ 2 (n k), and RSS and ESS are distributed independently. Refer to Example 1.3, F = = F (0.95,2,22) = Hence we reject H 0 convincingly at significance level We cannot reject H 0 only if α is set to E-08 or lower, as indicated by the test statistic s p-value. 35 / 57

67 Overall Significance of the Model Why do we perform an F test in addition to t tests? What can we learn from the F test? 36 / 57

68 Overall Significance of the Model Why do we perform an F test in addition to t tests? What can we learn from the F test? In the intercept only model, all of the fitted values equal the mean of the response variable. Therefore, if the overall F test is significant, the regression model predicts the response better than than the mean of the response. 36 / 57

69 Overall Significance of the Model Why do we perform an F test in addition to t tests? What can we learn from the F test? In the intercept only model, all of the fitted values equal the mean of the response variable. Therefore, if the overall F test is significant, the regression model predicts the response better than than the mean of the response. While R 2 provides an estimate of the strength of the relationship, it does not provide a formal hypothesis test for this relationship. If the overall F test is significant, one can conclude that the R 2 is significantly different from zero. In fact, the F statistic can be written as F = R 2 /(k 1) (1 R 2 )/(n k) 36 / 57

70 Overall Significance of the Model If the overall F test is significant, but few or none of the t tests are significant then it is an indication that multicollinearity might be a problem for the data. More on multicollinearity in Chapter / 57

71 Overall Significance of the Model If the overall F test is significant, but few or none of the t tests are significant then it is an indication that multicollinearity might be a problem for the data. More on multicollinearity in Chapter 3. Notice that for a simple linear regression model, the null hypothesis for the overall F test is simply β 2 = 0, which is precisely the same null for the t test of β 2 = 0. In fact, when k = 1, F (1,n k) = t 2 (n k). In Example 1.1, for testing H 0 : β 2 = 0, F = = (9.717) 2 = t 2, p-values are exactly the same. 37 / 57

72 Overall Significance of the Model If the overall F test is significant, but few or none of the t tests are significant then it is an indication that multicollinearity might be a problem for the data. More on multicollinearity in Chapter 3. Notice that for a simple linear regression model, the null hypothesis for the overall F test is simply β 2 = 0, which is precisely the same null for the t test of β 2 = 0. In fact, when k = 1, F (1,n k) = t 2 (n k). In Example 1.1, for testing H 0 : β 2 = 0, F = = (9.717) 2 = t 2, p-values are exactly the same. In Example 1.2, for testing H 0 : β 2 = 0, F = = (13.937) 2 = t 2, p-values are exactly the same. 37 / 57

73 F test for linear restrictions In fact, the usefulness of the F test is not limited to testing overall significance. The F test can be used for testing any linear equality restrictions on β. 38 / 57

74 F test for linear restrictions In fact, the usefulness of the F test is not limited to testing overall significance. The F test can be used for testing any linear equality restrictions on β. The general formula for the F statistic is F = (e e r e e ur )/m e e ur /(n k) = (R 2 ur R 2 r )/m (1 R 2 ur )/(n k) F (m,n k) H 0, where the subscripts ur and r correspond to the unrestricted and restricted models respectively, and m is the number of restrictions under H / 57

75 F test for linear restrictions e e r is the ESS associated with the restricted model (i.e., the model that imposes the restrictions implied by H 0 ; e e ur is the ESS associated with the unrestricted model (i.e., the model that ignores the restrictions). Rr 2 and Rur 2 are defined analogously. 39 / 57

76 F test for linear restrictions e e r is the ESS associated with the restricted model (i.e., the model that imposes the restrictions implied by H 0 ; e e ur is the ESS associated with the unrestricted model (i.e., the model that ignores the restrictions). Rr 2 and Rur 2 are defined analogously. The F statistic for testing H 0 : β 2 = β 3 = = β k = 0 is a special case of (1), because under H 0, m = k 1, e e r = TSS (the restricted model has no explanatory power) and correspondingly, Rr 2 = / 57

77 F test for linear restrictions Example 2.1 One model of production that is widely used in economics is the Cobb-Douglas production function: y i = β 1x β 2 2i x β 3 3i exp(ɛ i), where y i =output; x 2i =labour input; x 3i =capital input. Or, in log-transformed terms, lny i = lnβ 1 + β 2 lnx 2i + β 3 lnx 3i + ɛ i, = β 1 + β 2 lnx 2i + β 3 lnx 3i + ɛ i, 40 / 57

78 F test for linear restrictions To illustrate, we use annual data for the agricultural sector of Taiwan for Results obtained using SAS: The REG Procedure Model: MODEL1 Dependent Variable: lny Number of Observations Read 15 Number of Observations Used 15 Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model <.0001 Error Corrected Total Root MSE R-Square Dependent Mean Adj R-Sq Coeff Var Parameter Estimates Parameter Standard Variable DF Estimate Error t Value Pr > t Intercept lnx lnx / 57

79 F test for linear restrictions β 2 is the elasticity of output with respect to the labour input; it measures the percentage change in output due to a one percent change in labour input; β 3 is interpreted analogously. 42 / 57

80 F test for linear restrictions β 2 is the elasticity of output with respect to the labour input; it measures the percentage change in output due to a one percent change in labour input; β 3 is interpreted analogously. The sum β 2 + β 3 gives information on returns to scale, that is, the response of output to a proportional change in the inputs. In particular, if this sum is 1, then there are constant returns to scale, that is, doubling the inputs will double the outputs. Hence one may be interested in testing H 0 : β 2 + β 3 = / 57

81 F test for linear restrictions The restricted model is one that imposes the restriction β 2 + β 3 = 1 onto the coefficients in the minimisation of the SSE. The least squares estimator (referred to as restricted least squares (R.L.S.)) is obtained by minimising the objective function φ = (y Xb ) (y Xb ) 2λ (Rb r), where R is a m k matrix of constants, r is a m 1 vector of constants and b is the R.L.S. estimator. 43 / 57

82 F test for linear restrictions The restricted model is one that imposes the restriction β 2 + β 3 = 1 onto the coefficients in the minimisation of the SSE. The least squares estimator (referred to as restricted least squares (R.L.S.)) is obtained by minimising the objective function φ = (y Xb ) (y Xb ) 2λ (Rb r), where R is a m k matrix of constants, r is a m 1 vector of constants and b is the R.L.S. estimator. For this example, R = [0 1 1] and r = / 57

83 F test for linear restrictions The SAS results are as follows: The REG Procedure Model: MODEL1 Dependent Variable: lny NOTE: Restrictions have been applied to parameter estimates. Number of Observations Read 15 Number of Observations Used 15 Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model <.0001 Error Corrected Total Root MSE R-Square Dependent Mean Adj R-Sq Coeff Var Parameter Estimates Parameter Standard Variable DF Estimate Error t Value Pr > t Intercept lnx lnx < / 57

84 F test for linear restrictions Using the F test procedure, to test H 0 : β 2 + β 3 = 1 vs. H 1 : otherwise F = (Rur 2 Rr 2 )/m (1 Rur 2 )/(n k) = ( )/1 ( )/12 = 4.34 At α = 0.05, F (0.95,1,12) = Hence we cannot reject H 0 at 0.05 level of significance and conclude that the returns to scale is constant. 45 / 57

85 F test for linear restrictions SAS can perform the test automatically. The result from the following output concurs with the result based on our calculations. The p-value indicates that H 0 can be rejected only when α is set to at least The REG Procedure Model: MODEL1 Test 1 Results for Dependent Variable lny Mean Source DF Square F Value Pr > F Numerator Denominator / 57

86 Coefficient of variation The coefficient variation is obtained by dividing the standard error of the regression by the mean of y i values and multiplying by / 57

87 Coefficient of variation The coefficient variation is obtained by dividing the standard error of the regression by the mean of y i values and multiplying by 100. It expresses the standard error of the regression in unit free values. Thus the coefficients of variation for two different regressions can be compared more readily than the standard errors because the influence of the units of the data has been removed. The SAS program for Example 2.1 is as follows. 47 / 57

88 SAS Program for Example 2.1 data example21; input y x2 x3; ods html close; ods listing; lny=log(y); lnx2=log(x2); lnx3=log(x3); cards; ; proc reg data=example21; model lny=lnx2 lnx3; test lnx2+lnx3=1; run; proc reg data=example21; model lny=lnx2 lnx3; restrict lnx2+lnx3=1; run; ods html close; ods html; run; 48 / 57

89 Adjusted Coefficient of Determination Some statisticians have suggested to modify R 2 to recognise the number of independent variables in the model. The reason is that R 2 can generally be made larger if additional explanatory variables are added to the model. A measure that recognises the number of explanatory variables in the model is called the adjusted coefficient of determination: R 2 a = 1 ESS/(n k) TSS/(n 1) 49 / 57

90 Adjusted Coefficient of Determination Some statisticians have suggested to modify R 2 to recognise the number of independent variables in the model. The reason is that R 2 can generally be made larger if additional explanatory variables are added to the model. A measure that recognises the number of explanatory variables in the model is called the adjusted coefficient of determination: R 2 a = 1 ESS/(n k) TSS/(n 1) For the unrestricted model of Example 2.1, R 2 a = / /14 = Hence the adjustment has only a small effect, as R 2 a is almost the same as R / 57

91 Inference on Prediction A point prediction is obtained by inserting the given X values into the regression equation, giving ŷ f = b 1 + b 2 x f 2 + b 3 x f b k x fk Let g = (1, x f 2, x f 3,, x fk ). Then ŷ f = g b. Note that var(g b) = g var(b)g. If we assume normality for the disturbance term, it follows that g b g β N(0, 1) var(g b) 50 / 57

92 Inference on Prediction When the unknown σ 2 in var(b) is replaced by s 2, the usual shift to the t distribution occurs, giving ŷ f E(y f ) s g (X X ) 1 g t (n k), from which a 100(1 α) percent confidence (or prediction) interval for E(y f ) is ŷ f ± t (1 α/2,n k) s g (X X ) 1 g (1) 51 / 57

93 Inference on Prediction Returning to Example 1.3, the estimated regression equation is: ŷ i = x i x i3 A family with annual disposable income of $50,000 and liquid assets worth $100,000 is predicted to spend ŷ f = (50) (100) = thousand dollars on non-durable goods and services in a year. 52 / 57

94 Inference on Prediction For this example, (X X ) 1 = Hence g (X X ) 1 g = [ ] = / 57

95 Inference on Prediction s = and t (0.975,22) = Thus, the 95% prediction interval for E(y f ) is ± 2.074(38.436) or to / 57

96 Inference on Prediction s = and t (0.975,22) = Thus, the 95% prediction interval for E(y f ) is ± 2.074(38.436) or to Sometimes one may wish to obtain a prediction interval for y f rather than E(y f ). The two differ only by the disturbance term ɛ f, which is unpredictable with a mean of 0, so the point prediction remains the same. 54 / 57

97 Inference on Prediction s = and t (0.975,22) = Thus, the 95% prediction interval for E(y f ) is ± 2.074(38.436) or to Sometimes one may wish to obtain a prediction interval for y f rather than E(y f ). The two differ only by the disturbance term ɛ f, which is unpredictable with a mean of 0, so the point prediction remains the same. However, the uncertainty of the prediction increases due to the presence of ɛ f. Now, y f = g β + ɛ f. Therefore, 54 / 57

98 Inference on Prediction e f = y f ŷ f = ɛ f g (b β). Squaring both sides and taking expectations gives var(e f ) = σ 2 + g var(b)g = σ 2 (1 + g (X X ) 1 g) from which we can derive the following t statistic: ŷ f y f s 1 + g (X X ) 1 g t (n k) 55 / 57

99 Inference on Prediction which leads to the 100(1 α) percent confidence interval for y f : ŷ f ± t (1 α/2,n k) s 1 + g (X X ) 1 g 56 / 57

100 Inference on Prediction which leads to the 100(1 α) percent confidence interval for y f : ŷ f ± t (1 α/2,n k) s 1 + g (X X ) 1 g Comparison with (1) shows that the only difference is an increase of 1 inside the square root term. Thus, for the data in Example 1.3, the prediction interval for y f is: ± 2.074(38.436) or to One can obtain these outputs directly using SAS by adding the following options to PROC REG: /p CLM CLI; 56 / 57

101 Inference on Prediction The REG Procedure Model: MODEL1 Dependent Variable: y Output Statistics Dependent Predicted Std Error Obs Variable Value Mean Predict 95% CL Mean 95% CL Predict Residual Sum of Residuals 0 Sum of Squared Residuals Predicted Residual SS (PRESS) / 57

CHAPTER 4: Forecasting by Regression

CHAPTER 4: Forecasting by Regression CHAPTER 4: Forecasting by Regression Prof. Alan Wan 1 / 57 Table of contents 1. Revision of Linear Regression 3.1 First-order Autocorrelation and the Durbin-Watson Test 3.2 Correction for Autocorrelation

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

Ordinary Least Squares Regression

Ordinary Least Squares Regression Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Lecture 2 Multiple Regression and Tests

Lecture 2 Multiple Regression and Tests Lecture 2 and s Dr.ssa Rossella Iraci Capuccinello 2017-18 Simple Regression Model The random variable of interest, y, depends on a single factor, x 1i, and this is an exogenous variable. The true but

More information

Lecture 3: Multiple Regression

Lecture 3: Multiple Regression Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u

More information

The OLS Estimation of a basic gravity model. Dr. Selim Raihan Executive Director, SANEM Professor, Department of Economics, University of Dhaka

The OLS Estimation of a basic gravity model. Dr. Selim Raihan Executive Director, SANEM Professor, Department of Economics, University of Dhaka The OLS Estimation of a basic gravity model Dr. Selim Raihan Executive Director, SANEM Professor, Department of Economics, University of Dhaka Contents I. Regression Analysis II. Ordinary Least Square

More information

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Q = (Y i β 0 β 1 X i1 β 2 X i2 β p 1 X i.p 1 ) 2, which in matrix notation is Q = (Y Xβ) (Y

More information

1. The Multivariate Classical Linear Regression Model

1. The Multivariate Classical Linear Regression Model Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Microeconometria Day # 5 L. Cembalo. Regressione con due variabili e ipotesi dell OLS

Microeconometria Day # 5 L. Cembalo. Regressione con due variabili e ipotesi dell OLS Microeconometria Day # 5 L. Cembalo Regressione con due variabili e ipotesi dell OLS Multiple regression model Classical hypothesis of a regression model: Assumption 1: Linear regression model.the regression

More information

Introduction to Estimation Methods for Time Series models. Lecture 1

Introduction to Estimation Methods for Time Series models. Lecture 1 Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation

More information

ECO220Y Simple Regression: Testing the Slope

ECO220Y Simple Regression: Testing the Slope ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x

More information

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47 ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with

More information

14 Multiple Linear Regression

14 Multiple Linear Regression B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 14 Multiple Linear Regression 14.1 The multiple linear regression model In simple linear regression, the response variable y is expressed in

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

CHAPTER 3: Multicollinearity and Model Selection

CHAPTER 3: Multicollinearity and Model Selection CHAPTER 3: Multicollinearity and Model Selection Prof. Alan Wan 1 / 89 Table of contents 1. Multicollinearity 1.1 What is Multicollinearity? 1.2 Consequences and Identification of Multicollinearity 1.3

More information

Multiple Linear Regression

Multiple Linear Regression Chapter 3 Multiple Linear Regression 3.1 Introduction Multiple linear regression is in some ways a relatively straightforward extension of simple linear regression that allows for more than one independent

More information

Greene, Econometric Analysis (7th ed, 2012)

Greene, Econometric Analysis (7th ed, 2012) EC771: Econometrics, Spring 2012 Greene, Econometric Analysis (7th ed, 2012) Chapters 2 3: Classical Linear Regression The classical linear regression model is the single most useful tool in econometrics.

More information

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

EXST Regression Techniques Page 1. We can also test the hypothesis H : œ 0 versus H : EXST704 - Regression Techniques Page 1 Using F tests instead of t-tests We can also test the hypothesis H :" œ 0 versus H :" Á 0 with an F test.! " " " F œ MSRegression MSError This test is mathematically

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

Chapter 2 Inferences in Simple Linear Regression

Chapter 2 Inferences in Simple Linear Regression STAT 525 SPRING 2018 Chapter 2 Inferences in Simple Linear Regression Professor Min Zhang Testing for Linear Relationship Term β 1 X i defines linear relationship Will then test H 0 : β 1 = 0 Test requires

More information

Inferences for Regression

Inferences for Regression Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In

More information

STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002

STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002 Time allowed: 3 HOURS. STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002 This is an open book exam: all course notes and the text are allowed, and you are expected to use your own calculator.

More information

Econometrics Homework 1

Econometrics Homework 1 Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z

More information

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

Chapter 2 The Simple Linear Regression Model: Specification and Estimation Chapter The Simple Linear Regression Model: Specification and Estimation Page 1 Chapter Contents.1 An Economic Model. An Econometric Model.3 Estimating the Regression Parameters.4 Assessing the Least Squares

More information

2 Prediction and Analysis of Variance

2 Prediction and Analysis of Variance 2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering

More information

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators Thilo Klein University of Cambridge Judge Business School Session 4: Linear regression,

More information

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H. ACE 564 Spring 2006 Lecture 8 Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information by Professor Scott H. Irwin Readings: Griffiths, Hill and Judge. "Collinear Economic Variables,

More information

Formal Statement of Simple Linear Regression Model

Formal Statement of Simple Linear Regression Model Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

The regression model with one fixed regressor cont d

The regression model with one fixed regressor cont d The regression model with one fixed regressor cont d 3150/4150 Lecture 4 Ragnar Nymoen 27 January 2012 The model with transformed variables Regression with transformed variables I References HGL Ch 2.8

More information

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

STOR 455 STATISTICAL METHODS I

STOR 455 STATISTICAL METHODS I STOR 455 STATISTICAL METHODS I Jan Hannig Mul9variate Regression Y=X β + ε X is a regression matrix, β is a vector of parameters and ε are independent N(0,σ) Es9mated parameters b=(x X) - 1 X Y Predicted

More information

Heteroskedasticity and Autocorrelation

Heteroskedasticity and Autocorrelation Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity

More information

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION Answer all parts. Closed book, calculators allowed. It is important to show all working,

More information

13 Simple Linear Regression

13 Simple Linear Regression B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 3 Simple Linear Regression 3. An industrial example A study was undertaken to determine the effect of stirring rate on the amount of impurity

More information

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u Interval estimation and hypothesis tests So far our focus has been on estimation of the parameter vector β in the linear model y i = β 1 x 1i + β 2 x 2i +... + β K x Ki + u i = x iβ + u i for i = 1, 2,...,

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

The Classical Linear Regression Model

The Classical Linear Regression Model The Classical Linear Regression Model ME104: Linear Regression Analysis Kenneth Benoit August 14, 2012 CLRM: Basic Assumptions 1. Specification: Relationship between X and Y in the population is linear:

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Lecture 3: Inference in SLR

Lecture 3: Inference in SLR Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals

More information

ECON 4230 Intermediate Econometric Theory Exam

ECON 4230 Intermediate Econometric Theory Exam ECON 4230 Intermediate Econometric Theory Exam Multiple Choice (20 pts). Circle the best answer. 1. The Classical assumption of mean zero errors is satisfied if the regression model a) is linear in the

More information

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation 1/30 Outline Basic Econometrics in Transportation Autocorrelation Amir Samimi What is the nature of autocorrelation? What are the theoretical and practical consequences of autocorrelation? Since the assumption

More information

Simple linear regression

Simple linear regression Simple linear regression Biometry 755 Spring 2008 Simple linear regression p. 1/40 Overview of regression analysis Evaluate relationship between one or more independent variables (X 1,...,X k ) and a single

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

General Linear Model (Chapter 4)

General Linear Model (Chapter 4) General Linear Model (Chapter 4) Outcome variable is considered continuous Simple linear regression Scatterplots OLS is BLUE under basic assumptions MSE estimates residual variance testing regression coefficients

More information

Lab 07 Introduction to Econometrics

Lab 07 Introduction to Econometrics Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand

More information

Multiple Regression Analysis

Multiple Regression Analysis Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple

More information

LECTURE 5 HYPOTHESIS TESTING

LECTURE 5 HYPOTHESIS TESTING October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression ST 430/514 Recall: a regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates).

More information

3. The F Test for Comparing Reduced vs. Full Models. opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics / 43

3. The F Test for Comparing Reduced vs. Full Models. opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics / 43 3. The F Test for Comparing Reduced vs. Full Models opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics 510 1 / 43 Assume the Gauss-Markov Model with normal errors: y = Xβ + ɛ, ɛ N(0, σ

More information

STAT 3A03 Applied Regression With SAS Fall 2017

STAT 3A03 Applied Regression With SAS Fall 2017 STAT 3A03 Applied Regression With SAS Fall 2017 Assignment 2 Solution Set Q. 1 I will add subscripts relating to the question part to the parameters and their estimates as well as the errors and residuals.

More information

Sociology 6Z03 Review II

Sociology 6Z03 Review II Sociology 6Z03 Review II John Fox McMaster University Fall 2016 John Fox (McMaster University) Sociology 6Z03 Review II Fall 2016 1 / 35 Outline: Review II Probability Part I Sampling Distributions Probability

More information

Statistical Inference. Part IV. Statistical Inference

Statistical Inference. Part IV. Statistical Inference Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Empirical Economic Research, Part II

Empirical Economic Research, Part II Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction

More information

The Linear Regression Model

The Linear Regression Model The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general

More information

Inference for Regression

Inference for Regression Inference for Regression Section 9.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 13b - 3339 Cathy Poliak, Ph.D. cathy@math.uh.edu

More information

2 Regression Analysis

2 Regression Analysis FORK 1002 Preparatory Course in Statistics: 2 Regression Analysis Genaro Sucarrat (BI) http://www.sucarrat.net/ Contents: 1 Bivariate Correlation Analysis 2 Simple Regression 3 Estimation and Fit 4 T -Test:

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

Chapter 1 Linear Regression with One Predictor

Chapter 1 Linear Regression with One Predictor STAT 525 FALL 2018 Chapter 1 Linear Regression with One Predictor Professor Min Zhang Goals of Regression Analysis Serve three purposes Describes an association between X and Y In some applications, the

More information

Linear Regression Models

Linear Regression Models Linear Regression Models November 13, 2018 1 / 89 1 Basic framework Model specification and assumptions Parameter estimation: least squares method Coefficient of determination R 2 Properties of the least

More information

ST505/S697R: Fall Homework 2 Solution.

ST505/S697R: Fall Homework 2 Solution. ST505/S69R: Fall 2012. Homework 2 Solution. 1. 1a; problem 1.22 Below is the summary information (edited) from the regression (using R output); code at end of solution as is code and output for SAS. a)

More information

Topic 14: Inference in Multiple Regression

Topic 14: Inference in Multiple Regression Topic 14: Inference in Multiple Regression Outline Review multiple linear regression Inference of regression coefficients Application to book example Inference of mean Application to book example Inference

More information

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables. Regression Analysis BUS 735: Business Decision Making and Research 1 Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn how to estimate

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory

More information

Table 1: Fish Biomass data set on 26 streams

Table 1: Fish Biomass data set on 26 streams Math 221: Multiple Regression S. K. Hyde Chapter 27 (Moore, 5th Ed.) The following data set contains observations on the fish biomass of 26 streams. The potential regressors from which we wish to explain

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

Instrumental Variables, Simultaneous and Systems of Equations

Instrumental Variables, Simultaneous and Systems of Equations Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated

More information

Multiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor)

Multiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor) 1 Multiple Regression Analysis: Estimation Simple linear regression model: an intercept and one explanatory variable (regressor) Y i = β 0 + β 1 X i + u i, i = 1,2,, n Multiple linear regression model:

More information

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e., 1 Motivation Auto correlation 2 Autocorrelation occurs when what happens today has an impact on what happens tomorrow, and perhaps further into the future This is a phenomena mainly found in time-series

More information

Correlation and the Analysis of Variance Approach to Simple Linear Regression

Correlation and the Analysis of Variance Approach to Simple Linear Regression Correlation and the Analysis of Variance Approach to Simple Linear Regression Biometry 755 Spring 2009 Correlation and the Analysis of Variance Approach to Simple Linear Regression p. 1/35 Correlation

More information

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK 1 ECONOMETRICS STUDY PACK MAY/JUNE 2016 Question 1 (a) (i) Describing economic reality (ii) Testing hypothesis about economic theory (iii) Forecasting future

More information

The Simple Linear Regression Model

The Simple Linear Regression Model The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate

More information

1 Independent Practice: Hypothesis tests for one parameter:

1 Independent Practice: Hypothesis tests for one parameter: 1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)

More information

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA March 6, 2017 KC Border Linear Regression II March 6, 2017 1 / 44 1 OLS estimator 2 Restricted regression 3 Errors in variables 4

More information

MORE ON SIMPLE REGRESSION: OVERVIEW

MORE ON SIMPLE REGRESSION: OVERVIEW FI=NOT0106 NOTICE. Unless otherwise indicated, all materials on this page and linked pages at the blue.temple.edu address and at the astro.temple.edu address are the sole property of Ralph B. Taylor and

More information

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

Quantitative Methods for Economics, Finance and Management (A86050 F86050) Quantitative Methods for Economics, Finance and Management (A86050 F86050) Matteo Manera matteo.manera@unimib.it Marzio Galeotti marzio.galeotti@unimi.it 1 This material is taken and adapted from Guy Judge

More information

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58 Inference ME104: Linear Regression Analysis Kenneth Benoit August 15, 2012 August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58 Stata output resvisited. reg votes1st spend_total incumb minister

More information

Regression #8: Loose Ends

Regression #8: Loose Ends Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity 1/25 Outline Basic Econometrics in Transportation Heteroscedasticity What is the nature of heteroscedasticity? What are its consequences? How does one detect it? What are the remedial measures? Amir Samimi

More information

Two-Variable Regression Model: The Problem of Estimation

Two-Variable Regression Model: The Problem of Estimation Two-Variable Regression Model: The Problem of Estimation Introducing the Ordinary Least Squares Estimator Jamie Monogan University of Georgia Intermediate Political Methodology Jamie Monogan (UGA) Two-Variable

More information

Lecture 1 Linear Regression with One Predictor Variable.p2

Lecture 1 Linear Regression with One Predictor Variable.p2 Lecture Linear Regression with One Predictor Variablep - Basics - Meaning of regression parameters p - β - the slope of the regression line -it indicates the change in mean of the probability distn of

More information

Handout 11: Measurement Error

Handout 11: Measurement Error Handout 11: Measurement Error In which you learn to recognise the consequences for OLS estimation whenever some of the variables you use are not measured as accurately as you might expect. A (potential)

More information

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time Autocorrelation Given the model Y t = b 0 + b 1 X t + u t Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time This could be caused

More information

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s

More information

Correlation & Simple Regression

Correlation & Simple Regression Chapter 11 Correlation & Simple Regression The previous chapter dealt with inference for two categorical variables. In this chapter, we would like to examine the relationship between two quantitative variables.

More information

Brief Suggested Solutions

Brief Suggested Solutions DEPARTMENT OF ECONOMICS UNIVERSITY OF VICTORIA ECONOMICS 366: ECONOMETRICS II SPRING TERM 5: ASSIGNMENT TWO Brief Suggested Solutions Question One: Consider the classical T-observation, K-regressor linear

More information

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like.

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like. Measurement Error Often a data set will contain imperfect measures of the data we would ideally like. Aggregate Data: (GDP, Consumption, Investment are only best guesses of theoretical counterparts and

More information

Topic 3: Inference and Prediction

Topic 3: Inference and Prediction Topic 3: Inference and Prediction We ll be concerned here with testing more general hypotheses than those seen to date. Also concerned with constructing interval predictions from our regression model.

More information