Chapter 7. Testing Linear Restrictions on Regression Coefficients

Size: px
Start display at page:

Download "Chapter 7. Testing Linear Restrictions on Regression Coefficients"

Transcription

1 Chapter 7 Testing Linear Restrictions on Regression Coefficients 1.F-tests versus t-tests In the previous chapter we discussed several applications of the t-distribution to testing hypotheses in the linear regression model. In this chapter the range of possible hypotheses is extended beyond the scope of the t-distribution and accordingly we will introduce the F-distribution. Hypotheses can be thought of as restrictions on the population parameters. In Chapter 6, all the hypotheses that we considered implied a single restriction e.g. H 0 : 1 = 0. We will refer to these as Single Restriction Hypotheses (SRH). In this chapter we introduce the idea of Multiple Restriction Hypotheses (MRH). To put he discussion in context consider the general linear regression model: Y =. + ß 1 X 1 + ß 2 X ß K X K + u [7.1] In this model there are H = K + 1 coefficients on the right hand side. If the model is estimated with n observations, then the number of degrees of freedom is (n - H). In the previous chapter it was shown that an SRH of the form H 0 : ß k = ß 0 can be tested by comparing the sample t-statistic (b k - ß 0 )/SE(b k ) = t(n - H) with the appropriate critical value of the t-distribution with (n - H) degrees of freedom. In many cases researchers need to test more complex hypotheses such as H 0 : ß 1 = ß 2 = 0 [7.2a] The hypothesis in [7.2a] not only involves two coefficients, but it involves two separate restrictions namely that ß 1 = 0 and ß 2 = 0. Since both restrictions are to be tested together, the hypothesis in [7.2a] is referred to as a Multiple Restriction Hypotheses (MRH). The key feature that distinguishes an MRH from an SRH hypothesis is not the number of parameters in the hypothesis, but rather the number of separate restrictions. Each equals sign in an MRH represents a single restriction. The MRH in [7.2a] has p = 2 restrictions, where p represents the number of restrictions. The simplest way to count the restrictions in an MRH is to count the number of equals signs. Other examples of MRHs are

2 Chapter 7 2 H 0 : ß 1 = ß 2 = ß 3 = 0 [7.2b] and H 0 : ß 1 = ß 2 and ß 3 = ß 4 [7.2c] In [7.2b] there are three equals signs, so p = 3 in this case. In example [7.2c], the MRH involves four parameters, but there are only two equals signs so p = 2 in this case. The main reason for distinguishing between single and multiple restriction hypotheses involving population regression coefficients is that an MRH can be tested only by using the F-distribution; the t- distribution cannot be used to test an MRH. 2. Hypotheses as Restrictions The most useful way to approach the testing of compound hypotheses is to focus on the fact that the hypothesis consists of p linear restrictions on the population regression coefficients. To illustrate these ideas consider the five-parameter model (H = 5): Y =. + ß 1 X 1 + ß 2 X 2 + ß 3 X 3 + ß 4 X 4 + u [7.3] Equation [7.3] is referred to as the unrestricted model. Hypothesis [7.2a] restricts two of the slope coefficients to be zero. To impose these restrictions, substitute the equations from the null hypothesis into the unrestricted model. The result is the so-called restricted model, which in this case is: Y =. + ß 3 X 3 + ß 4 X 4 + u [7.3a] Notice that in the restricted model [7.5a] there are three parameters to estimate. The five parameters in the unrestricted model have been reduced to three, H - p = 5-2 = 3. The reduction in the number of parameters is equal to the number of restrictions. Now consider the restrictions implied by the hypothesis in [7.2b]. When these three restrictions are imposed on the unrestricted model, the result is a restricted model with H - p = 5-3 = 2 parameters: Y =. + ß 4 X 4 + u [7.3b] Finally, consider the hypothesis [7.2c] which involves 4 coefficients but implies just two restrictions. To impose these restrictions we will eliminate the parameters ß 2 and ß 4 :

3 Chapter 7 3 Y =. + ß 1 X 1 + ß 1 X 2 + ß 3 X 3 + ß 3 X 4 + u Notice that the restricted model has H - p = 5-2 = 3 parameters. However, it is not in the usual form because there are four right hand side variables and a constant term, but only three free parameters. To estimate the restricted model for hypothesis [2.3c], variables with identical coefficients must be collected together: Y =. + ß 1 Z + ß 3 W + u [7.3c] where Z = X 1 + X 2 and W = X 3 + X 4. The three parameters that appear in the restricted model can be estimated by regressing Y on Z, W and a constant. The variables Z and W are calculated by adding together the appropriate pairs of X-variables. These examples illustrate the general principle that when p linear restrictions that are substituted into the unrestricted model with H parameters, the result is a restricted model with (H - p) parameters. Whereas estimating the unrestricted involves regressing Y on H variables (which include the constant term) the restricted model is estimated by regressing Y on (H - p) variables. As shown in the next section, the test statistic used to test H 0 is based on a comparison of how well the unrestricted and restricted models fit the sample data. 3. The F-statistic When the unrestricted model [7.1] is estimated using the available sample of n observations, there will be a certain Sum of Squared Residuals (SSR u ) associated with this unrestricted model. 2 SSR u = e i where e i = Y i - a - b 1 X 1i b K X Ki To test an MRH such as the ones described by [7.2a], [7.2b] and [7.2c], the appropriate restricted model is fitted to exactly the same sample. The result will be a restricted Sum of Squared Residuals, SSR r. One point that we can be absolutely sure of is that the unrestricted model must fit the data at least as well as the restricted model. In other words, the unrestricted Sum of Squared Residuals must be less than or equal to the restricted Sum of Squared Residuals: SSR u SSR r i.e., SSR r - SSR u 0 When the unrestricted model is estimated, the H parameters are chosen by the least squares procedure to

4 Chapter 7 4 ensure that the Sum of Squared Residuals is as small as possible. When restrictions are imposed on the way in which the parameters are chosen it is not possible that the Sum of Squared Residuals will be reduced still further. To illustrate, consider the unrestricted model [7.3] and the restricted model [7.3a]. Suppose [7.3a] is estimated first and the estimated coefficients are a * *, b 3 and b 4*. The restricted Sum of Squared Residuals is SSR r. When the unrestricted model [7.3] is estimated, it is obvious that least squares could again achieve SSR r by simply choosing a = a * * *, b 3 = b 3 and b 4 = b 4 while b 1 = b 2 = 0. In fact, unrestricted least squares will generally reduce the Sum of Squared Residuals further by selecting non zero values of b 1 and b 2. Notice that unrestricted least squares can always reproduce the fit that is achieved by restricted least squares. In almost every sample, unrestricted least squares will fit the data better than the restricted model. What would it mean if in a particular sample SSR u = SSR r? This can happen only if the unrestricted parameters (chosen by unrestricted least squares) satisfy exactly the restrictions in the hypothesis. In the previous example, this would mean that unrestricted least squares is free to choose any values for b 3 and b 4, but the best values to choose just happen to be zero, as specified in the null hypothesis. In this unlikely event, we would have to conclude that the sample evidence is perfectly consistent with the hypothesis that the population parameters ß 3 and ß 4 are both zero. After all, when least squares is free to estimate these parameters, the estimates are zero. Clearly, imposing the restrictions in the hypothesis has no effect on the Sum of Squared Residuals. Normally we think of the unrestricted model as the base case. Now if the restrictions in H 0 result in a big increase in the Sum of Squared Residuals (compared to the unrestricted model), there is a conflict between the sample evidence and the hypothesis. In short, if the restrictions raise the Sum of Squared Residuals by a sufficient amount, then the hypothesis is rejected. The question of whether the restrictions raise the residual sum squares enough to cause H 0 to be rejected is answered by the F-statistic: F(p,n-H) = {(SSR r - SSR u )/SSR u }*(n-h)/p [7.4] The first part of the F-statistic is {(SSR r - SSR u )/SSR u }, which is the change in the Sum of Squared Residuals divided by the unrestricted Sum of Squared Residuals. In other words, the first part is the percentage increase in the Sum of Squared Residuals that results from the restrictions in H 0. If this is a large percentage increase, it indicates the sample information contradicts the hypothesis. The second part of the F-statistic

5 Chapter 7 5 is the ratio (n-h)/p, i.e., the degrees of freedom in the unrestricted model divided by the number of restrictions in H 0. The F-statistic is characterized by two degrees of freedom, p and (n-h). The first is p, often referred to as the degrees of freedom in the numerator of the F-statistic, since it is the p restrictions that are responsible for the quantity (SSR r - SSR u ). Notice that while the numerator of [7.4] has p is the degrees of freedom, the value of p actually appears in the denominator of [7.4]. The second quantity is (n-h), which are the degrees of freedom in denominator term, namely the unrestricted residuals, SSR u. The critical value of the F-statistic depends on three factors. First, is the level of significance or the size of the test. Typically, tests are conducted at the 5% or 1% level of significance. Second, the critical value depends on the degrees of freedom p and (n-h). The critical F-value for a test at the 5% level of significance is written: F.05 (p, n-h) The hypothesis H 0 is rejected if: F(p,n-H) > F.05 (p, n-h) 4. An Equivalent Form of the F-statistic The F-statistic can be expressed in an alternative form by noting that the estimate of the variance of the population deviations u 1 u 2 from the unrestricted model is s 2 2 = {1/(n-H)}e i = {1/(n-H)}SSR u The F-statistic in [7.4] can therefore be written as F(p,n-H) = {(SSR r - SSR u )}/(p*s 2 ) [7.5] 2 It is important to note that [7.5] uses the estimate of 1 u based on the unrestricted model. The intuitive explanation for this is that the restrictions in H 0 may be false and if they are the restricted model will provide 2 a biased estimate of 1 u2. The unrestricted model gives an unbiased estimate of 1 u whether the restrictions are true or false (assuming as we are that the unrestricted model is a valid starting point). 5 Two Examples: A Time Series Saving Equation

6 Chapter 7 6 The dependent variable in equation [7.6] is RSAV which represents per capita real savings in Canada. RSAV has been computed from quarterly National Accounts data over a period that starts in the first quarter of 1951 (1951:1) up to the 3 rd quarter in 1993 (1993:3) which amounts to 171 quarterly observations. In [7.6] quarterly per capita household saving is explained by real personal disposable income per capita (RPDI) and an annual interest rate (R4Y). It is expected that the coefficients on both RPDI and R4Y will be positive - higher incomes and higher interest rates encourage households to save more of their income. In addition, the four quarterly dummy variables Q1 to Q4 allow RSAV to exhibit quarterly seasonal variations. See Chapter 5 for a more detailed discussion of this model and the data. RSAV =. 1 Q Q Q Q4 + ß 1 *RPDI + ß 2 *R4Y + u [7.6] The first hypothesis that we will test is: H 0 : ß 1 = 0 against the alternative that ß 1 g 0. The restricted model is obtained by substituting the restriction from H 0 into [7.6]. The restricted model is therefore: RSAV =. 1 Q Q Q Q4 + ß 2 *R4Y + u [7.7] Notice that a single restriction reduces the number of right hand side parameters from 6 to 5 i.e., a reduction of one. Table 7.1 reports the results of estimating two regressions that correspond to the unrestricted [7.6] and restricted [7.7] models. The F-statistic for this test is given by equation [7.4] and is calculated using the restricted and unrestricted residual sums of squares that are reported in Table 7.1. Notice that there are 171 observations, H u = 6 and p = 1. F(1, 165) = {( * *10 5 )/ *10 5 }*(165/1) F(1, 165) = {( )/ }*165 = The critical F-value for a test at the 5% level of significance can be located in the table of critical values. The appropriate number is found by consulting the relevant row and column where row = (n - H u ) and column = p. Since there is no row for (n - H u ) = 165 we consult the next row up, which applies to 120 degrees of freedom. The critical value is F * (1,120) = 3.92 and since the computed F-statistic is much greater than F *, the null hypothesis is rejected. The sample evidence contradicts the null hypothesis that income has no effect on saving. To put this in a positive way, the sample evidence supports the view that saving is positively related to income. The hypothesis H 0 : ß 1 = 0 is an SRH not an MRH and could just have easily been tested using the

7 Chapter 7 7 t-test. Notice that TSP automatically reports the t-statistics for all the separate hypotheses that the population each regression coefficient is zero. For the null hypothesis that the true income coefficient is zero, the t- statistic is and the critical value for a two-tailed test at the 5% level of significance and 120 degrees of freedom is Thus according to the t-test, the hypothesis is rejected. In fact, the t-test and the F-test are equivalent tests because F(1, n-h u ) = t 2 (n-h u ). Notice that the square of the sample t-statistic is = , which is the sample F-statistic. Also the square of the critical t-value is = 3.92, which is the critical F-value. This functional relationship ensures that the t-test and F-test approaches to testing simple hypotheses are bound to result in identical conclusions. In no case will they produce conflicting evidence.

8 Chapter 7 8 Table 7.1 Dependent variable: RSAV Current sample: 1951:1 to 1993:3 Number of observations: 171 Mean of dependent variable = Std. dev. of dependent var. = The Unrestricted Model: Sum of squared residuals = Variance of residuals = Std. error of regression = R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q RPDI R4Y The Restricted Model Sum of squared residuals = E+07 Variance of residuals = Std. error of regression = R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q R4Y Example 2: The second example tests the hypothesis that the seasonal dummy variables all have the same coefficient. If this hypothesis is true then there is no seasonal variation in saving beyond the seasonality that may be transmitted to saving from any seasonal variations in income and the interest rate. In other words,. Specifically, in the context of equation [7.6] the hypothesis is H 0 :. 1 =. 2 =. 3 =. 4

9 Chapter 7 9 There are three "=" signs in the null hypothesis (p = 3) and so this MRH must be tested with an F-statistic - there is no valid t-test for this case. The unrestricted or alternative model is shown in equation [7.6]. The null hypothesis is that the seasonal intercepts are all identical. Let. be the common value of the coefficients on Q1, Q2, Q3 and Q4. When the restrictions are substituted into [7.6], the four seasonal factors collapse to a constant term with coefficient.:. 1 Q Q Q Q4 =.(Q1 + Q2 + Q3 + Q4) =.C where C is a vector of 1's i.e. the constant term. (see the Chapter 5 for more details on the relationship between the Q s and C) So the restricted model is RSAV =.C + ß 1 *RPDI + ß 2 *R4Y + u [7.8] where C is a vector of units ("1s"). Notice that if. is substituted for the seasonal.'s in [7.6], the common. can be factored out:.(q1 + Q2 + Q3 + Q4) =.C, since the vector sum of the seasonal dummy variables is the constant vector C which has 1 in every position. Table 7.2 shows the results of estimating the restricted model [7.8]. To test the null hypothesis, the key information in Table 7.2 is the sum of squared residuals. From equation [7.4] the F-statistic is F(3, 165) = {( * *10 5 )/ *10 5 }*(165/3) F(3, 165) = The critical value F * (3,120) at the 5% level of significance is 2.68, so the null hypothesis is overwhelmingly rejected. That is, the hypothesis that there is no seasonal variation in real per capita saving beyond that which can be explained by seasonal variations in RPDI and R4Y is rejected. Again, it is useful to express the evidence in a positive way. The evidence in the sample supports the contention that RSAV has a seasonal component that cannot be explained by the any seasonal variations in RPDI and R4Y. Suppose the hypothesis had not been rejected. In this case, the sample evidence would not have been able to reject model [7.8]. Notice that model [7.8] could be consistent with some seasonal pattern in RSAV. But in [7.8], all of the seasonal pattern in RSAV has to come from seasonal variations in RPDI or R4Y. In [7.8], there can be no independent seasonality in RSAV. However, the sample evidence has strongly rejected [7.8]. Finally, it is instructive to look at the contents of Table 7.2 in more detail. First, note that the R-

10 Chapter 7 10 squared statistic is 0.49 and this much lower than the R-squared for the unrestricted model. Clearly, the restrictions result in a much poorer fit (the sum of squared residual rises considerably. Also, the coefficient on R4Y in Table 7.2 has a t-statistic of This low value means that in the context of [7.8], we would not reject the hypothesis that the interest rate has no effect on real saving. (If someone claims interest rates have no effect on saving, the evidence in [7.2] cannot reject this claim.) But is this really what the data imply? The answer is no because [7.2] is itself rejected by the data. This illustrates how a badly specified model such as [7.8] can give misleading inferences. Model [7.8] has been rejected in favour of [7.6] and when [7.6] is estimated the claim that the interest rate has no effect on real saving is rejected (interest rates do appear to influence saving.) Dependent variable: RSAV Current sample: 1951:1 to 1993:3 Number of observations: 171 Table 7.2 Mean of dependent variable = Std. dev. of dependent var. = Sum of squared residuals = E+07 Variance of residuals = Std. error of regression = R-squared = Estimated Standard Variable Coefficient Error t-statistic C RPDI R4Y Example 3: Restrictions that Modify the Dependent Variable The hypotheses that we have considered so far have all led to restricted models that alter the right hand side of the equation. In some cases the restrictions alter the dependent variable as well and we will consider such an example in this section. For illustrative purposes consider the model in [7.9] y. 1 X 1 2 X 2 3 X 3 u [7.9]

11 Chapter 7 11 The null hypothesis is H 0 : and When the restrictions in the null are substituted into [7.9] to obtain the restricted model we get an equation in which two coefficients on the right hand side are predetermined. Since it is not possible to estimate a model in that form, the predetermined part is brought to the left hand side and so redefines the dependent variable: y. 1.5X 1 1.0X 2 3 X 3 u (y 1.5X 1 1.0X 2 ) 3 X 3 u W 3 X 3 u [7.10] where W (y 1.5X 1 1.0X 2 ) Equation [7.10] is the restricted model. It is derived by substituting the restrictions implied by the null hypothesis into the unrestricted model. In order to estimate the restricted model in [7.10], it is necessary to calculate the modified dependent variable, W. Now lets turn to the empirical example. The price of a single-detached house depends on the characteristics of the house and the location of the house. Table 7.3 reports the results of a study of 2446 house sales. Each transaction represents the sale of a single house for a certain price on a certain date. Some characteristics of these houses have been recorded in the sample. SAGAR is a dummy variable that takes the value 1 if the house has a single attached garage, SAGER = 0 otherwise. DAGAR signals the presence of a double attached garage. SIZE and LSIZE are the size of the house and lot in square feet. AGE is the age of the house in years. BATHP is the number of bathroom pieces (for example, two full bathrooms each with basin, shower, tub and toilet means BATHP = 8). POOLIG and POOLAG are dummy variables that record the presence of in-ground and above-ground pools respectively. TIME is a time trend variable that records the date of the market transaction. The full set of results in Table 7.3 will not be discussed here. We will focus on the value of single and double garages, which are the subject of our the hypotheses. First, note from Table 7.3 that the

12 Chapter 7 12 coefficient on SAGAR implies a single attached garage is estimated to raise the house price by $6,292 while a double garage is valued at $16,935. (Prices and values of characteristics apply to the early 1980s in the City of Guelph). We will consider two hypotheses concerning the value of garages. Hypotheses should be framed before looking at the data, so for the purposes of our example, assume that these hypotheses were suggested by two real estate agents who derived them from their general knowledge of the housing market, not the statistical results in Table 7.3. The first hypothesis states that single car garages and double car garages raise house prices by 6000 and dollars respectively. These values are very close to the unrestricted least squares estimates in Table 7.3, so it will not be surprising if this hypothesis is not rejected by the data i.e., consistent with the data.. The second hypothesis claims that single and double car garages raise house prices by 5000 and dollars respectively. These values are further from the unrestricted estimates and in fact, as we will see, far enough away that the hypothesis is rejected at the 5% level of significance. The procedure is to substitute the restrictions into the model and to estimate the resulting restricted model. Since the restrictions impose particular non-zero coefficient values on SAGAR and DAGAR, this is a case in which the dependent variable is modified.

13 Chapter 7 13 Current sample: 1 to 2446 Dependent variable: PRICE Current sample: 1 to 2446 Number of observations: 2446 Table 7.3 Mean of dependent variable = Std. dev. of dependent var. = Sum of squared residuals = E+12 Variance of residuals = E+09 Std. error of regression = R-squared = Adjusted R-squared = Estimated Standard Variable Coefficient Error t-statistic C SAGAR DAGAR SIZE LSIZE AGE BATHP POOLAG POOLIG TIME In the TSP program (see Appendix A), the new dependent variables are W1 and W2. The program calculates the F-statistics for the two tests and the values are printed in Table 7.4 along with the two restricted models. The critical values of the F-statistic is obtained from the appropriate statistical table.

14 Chapter 7 14 Table 7.4 Dependent variable: W1 Number of observations: 2446 Mean of dependent variable = Std. dev. of dependent var. = Sum of squared residuals = E+12 Std. error of regression = R-squared = Estimated Standard Variable Coefficient Error t-statistic C SIZE LSIZE AGE BATHP POOLAG POOLIG TIME F = At the 5% level of significance, the critical F(2, ) is 3.00 so the null hypothesis is not rejected. Dependent variable: W2 Number of observations: 2446 Mean of dependent variable = Std. dev. of dependent var. = Sum of squared residuals = E+12 Std. error of regression = R-squared = Estimated Standard Variable Coefficient Error t-statistic C SIZE LSIZE AGE BATHP POOLAG POOLIG TIME F = At the 5% level of significance, the critical F(2, ) is 3.00 so the null hypothesis is rejected

15 Chapter Structural Change It is often important to investigate whether regression parameters are stable over time. Did the North American Free Trade Agreement cause a shift in Canada's import function (the relationship between imports and national income)? Do advertising expenditures shift demand curves and can advertising brand names reduce own-price elasticities to the advantage of firms)? Is the demand for money function stable? If it is not, then a demand function estimated over the recent past may not produce good forecasts in the future. Economists often talk about "testing for structural change." What they really mean is that they want to test a hypothesis that parameters are stable (the restricted case) against an alternative in which parameters change from one period to another. The alternative or unrestricted model always has more parameters than the restricted model and it is the restricted model that the data may reject. In the previous examples equation [7.6] represented the unrestricted model. We simply accepted that over the period 1951:1 to 1993:3 the parameters of the saving function were stable or constant. That assumption will now be put to the test. To do so we will specify an unrestricted model in which the parameters are allowed to change at the end of 1972:1 In the new unrestricted model, all six parameters are allowed to take different values in the two subperiods (1951:1 to 1972:1) and (1972:2 to 1993:3). On way to estimate the new unrestricted model is to break the sample into two part and estimate [7.6] using the two sub-samples. The results are shown in Table 7.5

16 Chapter 7 16 Table 7.5 Dependent variable: RSAV Current sample: 1951:1 to 1972:1 Number of observations: 85 Mean of dependent variable = Sum of squared residuals = , R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q RPDI R4Y Current sample: 1972:2 to 1993:3 Number of observations: 86 Mean of dependent variable = Sum of squared residuals = , R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q RPDI R4Y On the one hand, the seasonal pattern in real saving does not seem to have changed substantially between the two subperiods. Both before and after 1972:1 savings are highest in the third quarter and lowest in the fourth quarter (given a certain level of disposable income and a certain interest rate). On the other hand, the estimated interest rate coefficient is negative in the first sub-period and positive in the second. The marginal propensity to save also seems to change dramatically between the two subperiods. The question is: are these differences statistically significant? Do the data reject parameter constancy? The question boils down to whether a model with fixed coefficients over the entire sample can explain the data as well as a model that allows the coefficients to be different in the two sub-samples. To be specific, the unrestricted model is RSAV =. 11 Q Q Q Q4 + ß 11 *RPDI + ß 21 *R4Y + u 1951:1-1972:1 RSAV =. 12 Q Q Q Q4 + ß 12 *RPDI + ß 22 *R4Y + u 1972:2-1993:3

17 Chapter 7 17 Notice that in the unrestricted model there are H u = 12 parameters. In the context of the above model, the null hypothesis is: H 0 :. 11 =. 12,. 21 =. 22,. 31 =. 32,. 41 =. 42, ß 11 = ß 12, ß 21 = ß 22 Notice that there are 6 "=" signs in the null hypothesis, that is p = 6 restrictions. When these restrictions are imposed, the model reduces to [7.6], which applies to the entire sample period. The sum of squared residuals for the unrestricted model over the entire sample range is the total of the sum of squared residuals for both subperiods. From Table 7.5, that total is ( )*10 5 = *10 5. In the present case, the estimates of the restricted model are reported in the top half of Table 7.1 There, the sum of squared residuals are *10 5. After cancelling the common factors 10 5, the F-statistic for this test is therefore: F(6, ) = {( )/ }*(159/12) = The critical F-value, F * (6, 120) is Since the calculated F-statistic is well above the critical value, the hypothesis of parameter stability is rejected in favour of the alternative hypothesis that the parameters have changed over time. Having rejected parameter stability, the next question is one of detail. Did all the parameters shift, or did just a few shift? The evidence in Table 7.5 suggests that the interest rate and income coefficients are very different between the two subperiods, but that the basic seasonal pattern in saving did not change at least in qualitative terms. The unrestricted model continues to be the pair of linear equations fitted to the subsamples 1951:1-1972:1 and 1972:2-1993:3 (which has 12 coefficients.) Consider first the hypothesis that the coefficients on income and the interest rate remained constant over the entire period while the quarterly intercepts shifted (given the evidence in Table 7.5, it would not be surprising if this is rejected). The hypothesis is: H 0 : ß 11 = ß 12, ß 21 = ß 22 [7.13] This hypothesis contains two restrictions (p = 2). The alternative model is that all 6 parameters shifted. The unrestricted model has already been estimated (see Table 7.5). But how is the restricted model to be estimated? The restricted model allows the seasonal dummy coefficients to shift between the two subperiods, but constrains the other two coefficients to be constant. Appendix B lists a TSP program that estimates the restricted and unrestricted models for this example. The first olsq statement estimates the unrestricted model

18 Chapter 7 18 in one step (instead of two separate regressions). Notice that RSAV is regressed on 12 variables. Q11, Q21, Q31 and Q41 are the seasonal dummies in the first half of the sample. Q12, Q22, Q32 and Q42 are the seasonal dummies in the second half of the sample. Similarly the coefficients on RPDI1 and RPDI2 are the marginal propensities to save in the first and second sub-samples respectively. The results from the first olsq statement are presented in Table 7.6. Notice that the results in Table 7.6 are identical to those from the separate subsample regressions that are reported in Table 7.5. (Compare the sum of squared residuals) Table 7.6 Dependent variable: RSAV Current sample: 1951:1 to 1993:3 Number of observations: 171 Mean of dependent variable = Sum of squared residuals = R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q Q Q Q Q RPDI RPDI R4Y R4Y The second olsq statement in Appendix B estimates the restricted model for this hypothesis. Notice that in the restricted regression there are two sets of seasonal dummy variables, but single coefficients on RPDI and R4Y. The results of the restricted regression are presented in Table 7.7. The restricted sum of squared residuals is *10 5. The F-statistic for the hypothesis [7.13] is F(2, 159) = {( )/ }*(159/2) = The critical F-value for a test at the 5% level of significance is 3.07, so the null hypothesis is rejected.

19 Chapter Dependent variable: RSAV Current sample: 1951:1 to 1993:3 Number of observations: 171 Mean of dependent variable = Sum of squared residuals = , R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q Q Q Q Q RPDI R4Y The final hypothesis test is that the seasonal coefficients are stable over the entire period (but the coefficients on R4Y and RPDI change in 1972). The unrestricted model again has 12 coefficients (6 in each of the two subsamples.) The null hypothesis, which has p = 4 restrictions, is H 0 :. 11 =. 12,. 21 =. 22,. 31 =. 32,. 41 =. 42 [7.12] The alternative hypothesis is that these seasonal coefficients change in 1972:1. The unrestricted model also allows the coefficients on income and the interest rate to change, so the hypothesis focuses exclusively on the seasonal coefficients. In the TSP program in Appendix B, the third and last olsq statement estimates the restricted model for hypothesis [7.12]. The results are in Table 7.8. The F-statistic is F(4, 159) = {( )/ }*(159/4) = 3.32 The critical F-value is F * (4, 120) = 2.45, so again the null hypothesis is rejected. It seems that there are no qualitative differences in the seasonal variations across the two sub-samples, but the quantitative differences are statistically significantly different. At the beginning of this section it was shown that the hypothesis that all the coefficients are constant is rejected. We then probed deeper to see if the instability could be traced to the slope coefficients on RPDI and R4Y or to changes in the seasonal coefficients. The result is that we rejected constancy in both the slope

20 Chapter 7 20 coefficients and the seasonal effects, so "overall instability" cannot be blamed on just the slopes alone or the seasonal intercepts alone. Table 7.8 Dependent variable: RSAV Current sample: 1951:1 to 1993:3 Number of observations: 171 Mean of dependent variable = Sum of squared residuals = R-squared = Estimated Standard Variable Coefficient Error t-statistic Q Q Q Q RPDI RPDI R4Y R4Y Appendix A The following TSP program calculates the regressions reported in Tables 7.3 and 7.4. The F- statistics for the two null hypotheses discussed in section 6 of this chapter are also computed by this TSP program. options memory=4; options crt; smpl ; read(file= '~/614/614.dat') price size lsize age poolag poolig bathp fire1 sagar dagar time; olsq price c sagar dagar size lsize age bathp poolag poolig time; set SSRu set h=@ncoef; w1=price-6000*sagar *dagar; olsq w1 c size lsize age bathp poolag poolig time; set SSRr=@ssr; set f=(ssrr-ssru)*(@nob-h)/(ssru*2);

21 Chapter 7 21 print f; w2=price-5000*sagar *dagar; olsq w2 c size lsize age bathp poolag poolig time; set SSRr=@ssr; set f=(ssrr-ssru)*(@nob-h)/(ssru*2); print f; ******** Appendix B This appendix contains the TSP program used to calculate the regressions reported in section 7 of this chapter (Structural change) options crt; in DAT374; smpl 1951:1 1993:3; trend t; rsav=sav*10**5/(cpi*pop); rpdi=pdi*10**5/(cpi*pop); msd rsav rpdi; dummy; d2=t>85; d1=t<=85; q11=q1*d1; q21=q2*d1; q31=q3*d1; q41=q4*d1; q12=q1*d2; q22=q2*d2; q32=q3*d2; q42=q4*d2; rpdi1=rpdi*d1; rpdi2=rpdi*d2; r4y1=r4y*d1; r4y2=r4y*d2; olsq rsav q11 q21 q31 q41 q12 q22 q32 q42 rpdi1 rpdi2 r4y1 r4y2; olsq rsav q11 q21 q31 q41 q12 q22 q32 q42 rpdi r4y; olsq rsav q1-q4 rpdi1 rpdi2 r4y1 r4y2; The following tsp program Explanatory notes: What are the roles of d1 & d2 in this program? d1 and d2 are dummy variables: d1 is unity between 1951:1 and 1972:1 but zero after that while d2 is zero up to 1972:1 an unity after that. The products d1*rpdi and d2*rpdi are two columns that stretch from 1951:1 to 1993:3. d1*rpdi equals rpdi from 1951:1 to 1972:1 but zero after that, d2*rpdi is zero up to 1972:1 but equals rpdi after that. By including d1*rpdi and d2*rpdi in a regression there will be two coefficients on income (one for the first sub-period and another for the second sub-period). If rpdi alone is included then the income coefficient is constant throughout the entire period.

22 Chapter 7 22 These ideas apply to the dummy variables and to the interest rate. How are d1 and d2 created? The tsp statement "trend t" creates a variable t that is 1 in the 1951:1 and increments by one every quarter. If you print t you get the values This is called a time trend because it just counts time. The first subsample is 1951:1 to 1972:1 (notice in the assignment you use a different date to split the sample) and t goes from 1 to 85. The second sample is 1972:2 to 1993:3 and t goes from 86 to 171. The dummy variable d1 is set to unity if t <= 85 and d1 = 0 otherwise. This is done in tsp by this command: d1 = t < = 85; This odd-looking statement is understood by tsp to mean: set d1 to unity if t<=85 is true and set d1=0 if t<=85 is false. Similarly, d2 = t > 85; sets d2=1 if t>85 and sets d2=0 if t <=85.

Chapter 3 Multiple Regression Complete Example

Chapter 3 Multiple Regression Complete Example Department of Quantitative Methods & Information Systems ECON 504 Chapter 3 Multiple Regression Complete Example Spring 2013 Dr. Mohammad Zainal Review Goals After completing this lecture, you should be

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables. Regression Analysis BUS 735: Business Decision Making and Research 1 Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn how to estimate

More information

Econometrics Homework 1

Econometrics Homework 1 Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

CHAPTER 5 FUNCTIONAL FORMS OF REGRESSION MODELS

CHAPTER 5 FUNCTIONAL FORMS OF REGRESSION MODELS CHAPTER 5 FUNCTIONAL FORMS OF REGRESSION MODELS QUESTIONS 5.1. (a) In a log-log model the dependent and all explanatory variables are in the logarithmic form. (b) In the log-lin model the dependent variable

More information

1 A Non-technical Introduction to Regression

1 A Non-technical Introduction to Regression 1 A Non-technical Introduction to Regression Chapters 1 and Chapter 2 of the textbook are reviews of material you should know from your previous study (e.g. in your second year course). They cover, in

More information

Chapter 3 Statistical Estimation of The Regression Function

Chapter 3 Statistical Estimation of The Regression Function Chapter 3 Statistical Estimation of The Regression Function 3.1 Statistical Estimation If the population can be observed, there is no statistical problem: all the features if the population are known.

More information

ECON 497: Lecture Notes 10 Page 1 of 1

ECON 497: Lecture Notes 10 Page 1 of 1 ECON 497: Lecture Notes 10 Page 1 of 1 Metropolitan State University ECON 497: Research and Forecasting Lecture Notes 10 Heteroskedasticity Studenmund Chapter 10 We'll start with a quote from Studenmund:

More information

Rockefeller College University at Albany

Rockefeller College University at Albany Rockefeller College University at Albany PAD 705 Handout: Suggested Review Problems from Pindyck & Rubinfeld Original prepared by Professor Suzanne Cooper John F. Kennedy School of Government, Harvard

More information

Regression Analysis. BUS 735: Business Decision Making and Research

Regression Analysis. BUS 735: Business Decision Making and Research Regression Analysis BUS 735: Business Decision Making and Research 1 Goals and Agenda Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn

More information

Chapter 11 Handout: Hypothesis Testing and the Wald Test

Chapter 11 Handout: Hypothesis Testing and the Wald Test Chapter 11 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic Manipulation o Wald Test Restricted Regression Reflects Unrestricted Regression

More information

Chapter 4: Regression Models

Chapter 4: Regression Models Sales volume of company 1 Textbook: pp. 129-164 Chapter 4: Regression Models Money spent on advertising 2 Learning Objectives After completing this chapter, students will be able to: Identify variables,

More information

Inference in Regression Model

Inference in Regression Model Inference in Regression Model Christopher Taber Department of Economics University of Wisconsin-Madison March 25, 2009 Outline 1 Final Step of Classical Linear Regression Model 2 Confidence Intervals 3

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 4 4- Basic Business Statistics th Edition Chapter 4 Introduction to Multiple Regression Basic Business Statistics, e 9 Prentice-Hall, Inc. Chap 4- Learning Objectives In this chapter, you learn:

More information

Inference with Simple Regression

Inference with Simple Regression 1 Introduction Inference with Simple Regression Alan B. Gelder 06E:071, The University of Iowa 1 Moving to infinite means: In this course we have seen one-mean problems, twomean problems, and problems

More information

Section 3: Simple Linear Regression

Section 3: Simple Linear Regression Section 3: Simple Linear Regression Carlos M. Carvalho The University of Texas at Austin McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction

More information

The Hong Kong University of Science & Technology ISOM551 Introductory Statistics for Business Assignment 4 Suggested Solution

The Hong Kong University of Science & Technology ISOM551 Introductory Statistics for Business Assignment 4 Suggested Solution 8 TUNG, Yik-Man The Hong Kong University of cience & Technology IOM55 Introductory tatistics for Business Assignment 4 uggested olution Note All values of statistics are obtained by Excel Qa Theoretically,

More information

Regression Analysis II

Regression Analysis II Regression Analysis II Measures of Goodness of fit Two measures of Goodness of fit Measure of the absolute fit of the sample points to the sample regression line Standard error of the estimate An index

More information

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H. ACE 564 Spring 2006 Lecture 8 Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information by Professor Scott H. Irwin Readings: Griffiths, Hill and Judge. "Collinear Economic Variables,

More information

Regression Models. Chapter 4. Introduction. Introduction. Introduction

Regression Models. Chapter 4. Introduction. Introduction. Introduction Chapter 4 Regression Models Quantitative Analysis for Management, Tenth Edition, by Render, Stair, and Hanna 008 Prentice-Hall, Inc. Introduction Regression analysis is a very valuable tool for a manager

More information

Finding Relationships Among Variables

Finding Relationships Among Variables Finding Relationships Among Variables BUS 230: Business and Economic Research and Communication 1 Goals Specific goals: Re-familiarize ourselves with basic statistics ideas: sampling distributions, hypothesis

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

Chapter 7 Student Lecture Notes 7-1

Chapter 7 Student Lecture Notes 7-1 Chapter 7 Student Lecture Notes 7- Chapter Goals QM353: Business Statistics Chapter 7 Multiple Regression Analysis and Model Building After completing this chapter, you should be able to: Explain model

More information

Use of Dummy (Indicator) Variables in Applied Econometrics

Use of Dummy (Indicator) Variables in Applied Econometrics Chapter 5 Use of Dummy (Indicator) Variables in Applied Econometrics Section 5.1 Introduction Use of Dummy (Indicator) Variables Model specifications in applied econometrics often necessitate the use of

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Statistical Inference. Part IV. Statistical Inference

Statistical Inference. Part IV. Statistical Inference Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define

More information

11.5 Regression Linear Relationships

11.5 Regression Linear Relationships Contents 11.5 Regression............................. 835 11.5.1 Linear Relationships................... 835 11.5.2 The Least Squares Regression Line........... 837 11.5.3 Using the Regression Line................

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test

Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test Amherst College Department of Economics Economics 360 Fall 2012 Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic

More information

LI EAR REGRESSIO A D CORRELATIO

LI EAR REGRESSIO A D CORRELATIO CHAPTER 6 LI EAR REGRESSIO A D CORRELATIO Page Contents 6.1 Introduction 10 6. Curve Fitting 10 6.3 Fitting a Simple Linear Regression Line 103 6.4 Linear Correlation Analysis 107 6.5 Spearman s Rank Correlation

More information

The Multiple Regression Model

The Multiple Regression Model Multiple Regression The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & or more independent variables (X i ) Multiple Regression Model with k Independent Variables:

More information

Inferences for Regression

Inferences for Regression Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In

More information

Correlation Analysis

Correlation Analysis Simple Regression Correlation Analysis Correlation analysis is used to measure strength of the association (linear relationship) between two variables Correlation is only concerned with strength of the

More information

Class time (Please Circle): 11:10am-12:25pm. or 12:45pm-2:00pm

Class time (Please Circle): 11:10am-12:25pm. or 12:45pm-2:00pm Name: UIN: Class time (Please Circle): 11:10am-12:25pm. or 12:45pm-2:00pm Instructions: 1. Please provide your name and UIN. 2. Circle the correct class time. 3. To get full credit on answers to this exam,

More information

Diploma Part 2. Quantitative Methods. Examiner s Suggested Answers

Diploma Part 2. Quantitative Methods. Examiner s Suggested Answers Diploma Part Quantitative Methods Examiner s Suggested Answers Question 1 (a) The standard normal distribution has a symmetrical and bell-shaped graph with a mean of zero and a standard deviation equal

More information

1 Correlation and Inference from Regression

1 Correlation and Inference from Regression 1 Correlation and Inference from Regression Reading: Kennedy (1998) A Guide to Econometrics, Chapters 4 and 6 Maddala, G.S. (1992) Introduction to Econometrics p. 170-177 Moore and McCabe, chapter 12 is

More information

Mathematics for Economics MA course

Mathematics for Economics MA course Mathematics for Economics MA course Simple Linear Regression Dr. Seetha Bandara Simple Regression Simple linear regression is a statistical method that allows us to summarize and study relationships between

More information

Chapter 14 Multiple Regression Analysis

Chapter 14 Multiple Regression Analysis Chapter 14 Multiple Regression Analysis 1. a. Multiple regression equation b. the Y-intercept c. $374,748 found by Y ˆ = 64,1 +.394(796,) + 9.6(694) 11,6(6.) (LO 1) 2. a. Multiple regression equation b.

More information

2 Prediction and Analysis of Variance

2 Prediction and Analysis of Variance 2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering

More information

Variance Decomposition and Goodness of Fit

Variance Decomposition and Goodness of Fit Variance Decomposition and Goodness of Fit 1. Example: Monthly Earnings and Years of Education In this tutorial, we will focus on an example that explores the relationship between total monthly earnings

More information

Monday, October 15 Handout: Multiple Regression Analysis Introduction

Monday, October 15 Handout: Multiple Regression Analysis Introduction Amherst College Department of Economics Economics 360 Fall 2012 Monday, October 15 Handout: Multiple Regression Analysis Introduction Review Simple and Multiple Regression Analysis o Distinction between

More information

Business Statistics. Lecture 10: Course Review

Business Statistics. Lecture 10: Course Review Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,

More information

Spring 1998 Final exam (35%)

Spring 1998 Final exam (35%) Econ 120C Ramu Ramanathan Spring 1998 Final exam (35%) Your name Your Id No. (NOT Soc. Sec. no.) Read the questions carefully and answer all points. For each part, confine your answers to the specific

More information

The multiple regression model; Indicator variables as regressors

The multiple regression model; Indicator variables as regressors The multiple regression model; Indicator variables as regressors Ragnar Nymoen University of Oslo 28 February 2013 1 / 21 This lecture (#12): Based on the econometric model specification from Lecture 9

More information

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK 1 ECONOMETRICS STUDY PACK MAY/JUNE 2016 Question 1 (a) (i) Describing economic reality (ii) Testing hypothesis about economic theory (iii) Forecasting future

More information

Chapter 14 Student Lecture Notes 14-1

Chapter 14 Student Lecture Notes 14-1 Chapter 14 Student Lecture Notes 14-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter 14 Multiple Regression Analysis and Model Building Chap 14-1 Chapter Goals After completing this

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

STAT 212 Business Statistics II 1

STAT 212 Business Statistics II 1 STAT 1 Business Statistics II 1 KING FAHD UNIVERSITY OF PETROLEUM & MINERALS DEPARTMENT OF MATHEMATICAL SCIENCES DHAHRAN, SAUDI ARABIA STAT 1: BUSINESS STATISTICS II Semester 091 Final Exam Thursday Feb

More information

Chapter 9. Dummy (Binary) Variables. 9.1 Introduction The multiple regression model (9.1.1) Assumption MR1 is

Chapter 9. Dummy (Binary) Variables. 9.1 Introduction The multiple regression model (9.1.1) Assumption MR1 is Chapter 9 Dummy (Binary) Variables 9.1 Introduction The multiple regression model y = β+β x +β x + +β x + e (9.1.1) t 1 2 t2 3 t3 K tk t Assumption MR1 is 1. yt =β 1+β 2xt2 + L+β KxtK + et, t = 1, K, T

More information

Types of economic data

Types of economic data Types of economic data Time series data Cross-sectional data Panel data 1 1-2 1-3 1-4 1-5 The distinction between qualitative and quantitative data The previous data sets can be used to illustrate an important

More information

Lab 11 - Heteroskedasticity

Lab 11 - Heteroskedasticity Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction

More information

Solutions: Monday, October 15

Solutions: Monday, October 15 Amherst College Department of Economics Economics 360 Fall 2012 1. Consider Nebraska petroleum consumption. Solutions: Monday, October 15 Petroleum Consumption Data for Nebraska: Annual time series data

More information

determine whether or not this relationship is.

determine whether or not this relationship is. Section 9-1 Correlation A correlation is a between two. The data can be represented by ordered pairs (x,y) where x is the (or ) variable and y is the (or ) variable. There are several types of correlations

More information

Chapter 13. Multiple Regression and Model Building

Chapter 13. Multiple Regression and Model Building Chapter 13 Multiple Regression and Model Building Multiple Regression Models The General Multiple Regression Model y x x x 0 1 1 2 2... k k y is the dependent variable x, x,..., x 1 2 k the model are the

More information

Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12)

Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12) Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12) Remember: Z.05 = 1.645, Z.01 = 2.33 We will only cover one-sided hypothesis testing (cases 12.3, 12.4.2, 12.5.2,

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 6: Multiple regression analysis: Further issues

Wooldridge, Introductory Econometrics, 4th ed. Chapter 6: Multiple regression analysis: Further issues Wooldridge, Introductory Econometrics, 4th ed. Chapter 6: Multiple regression analysis: Further issues What effects will the scale of the X and y variables have upon multiple regression? The coefficients

More information

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations. Exercises for the course of Econometrics Introduction 1. () A researcher is using data for a sample of 30 observations to investigate the relationship between some dependent variable y i and independent

More information

Chapter 9 Inferences from Two Samples

Chapter 9 Inferences from Two Samples Chapter 9 Inferences from Two Samples 9-1 Review and Preview 9-2 Two Proportions 9-3 Two Means: Independent Samples 9-4 Two Dependent Samples (Matched Pairs) 9-5 Two Variances or Standard Deviations Review

More information

Multiple Regression Analysis: The Problem of Inference

Multiple Regression Analysis: The Problem of Inference Multiple Regression Analysis: The Problem of Inference Jamie Monogan University of Georgia Intermediate Political Methodology Jamie Monogan (UGA) Multiple Regression Analysis: Inference POLS 7014 1 / 10

More information

ECON 497 Midterm Spring

ECON 497 Midterm Spring ECON 497 Midterm Spring 2009 1 ECON 497: Economic Research and Forecasting Name: Spring 2009 Bellas Midterm You have three hours and twenty minutes to complete this exam. Answer all questions and explain

More information

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit

More information

Multiple Regression: Inference

Multiple Regression: Inference Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled

More information

Statistics for Managers using Microsoft Excel 6 th Edition

Statistics for Managers using Microsoft Excel 6 th Edition Statistics for Managers using Microsoft Excel 6 th Edition Chapter 13 Simple Linear Regression 13-1 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value of

More information

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity LECTURE 10 Introduction to Econometrics Multicollinearity & Heteroskedasticity November 22, 2016 1 / 23 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists

More information

EXAMINATION QUESTIONS 63. P f. P d. = e n

EXAMINATION QUESTIONS 63. P f. P d. = e n EXAMINATION QUESTIONS 63 If e r is constant its growth rate is zero, so that this may be rearranged as e n = e n P f P f P d P d This is positive if the foreign inflation rate exceeds the domestic, indicating

More information

Practice exam questions

Practice exam questions Practice exam questions Nathaniel Higgins nhiggins@jhu.edu, nhiggins@ers.usda.gov 1. The following question is based on the model y = β 0 + β 1 x 1 + β 2 x 2 + β 3 x 3 + u. Discuss the following two hypotheses.

More information

Inference in Regression Analysis

Inference in Regression Analysis ECNS 561 Inference Inference in Regression Analysis Up to this point 1.) OLS is unbiased 2.) OLS is BLUE (best linear unbiased estimator i.e., the variance is smallest among linear unbiased estimators)

More information

STATISTICS 110/201 PRACTICE FINAL EXAM

STATISTICS 110/201 PRACTICE FINAL EXAM STATISTICS 110/201 PRACTICE FINAL EXAM Questions 1 to 5: There is a downloadable Stata package that produces sequential sums of squares for regression. In other words, the SS is built up as each variable

More information

Chapter 4. Regression Models. Learning Objectives

Chapter 4. Regression Models. Learning Objectives Chapter 4 Regression Models To accompany Quantitative Analysis for Management, Eleventh Edition, by Render, Stair, and Hanna Power Point slides created by Brian Peterson Learning Objectives After completing

More information

MGEC11H3Y L01 Introduction to Regression Analysis Term Test Friday July 5, PM Instructor: Victor Yu

MGEC11H3Y L01 Introduction to Regression Analysis Term Test Friday July 5, PM Instructor: Victor Yu Last Name (Print): Solution First Name (Print): Student Number: MGECHY L Introduction to Regression Analysis Term Test Friday July, PM Instructor: Victor Yu Aids allowed: Time allowed: Calculator and one

More information

Summary statistics. G.S. Questa, L. Trapani. MSc Induction - Summary statistics 1

Summary statistics. G.S. Questa, L. Trapani. MSc Induction - Summary statistics 1 Summary statistics 1. Visualize data 2. Mean, median, mode and percentiles, variance, standard deviation 3. Frequency distribution. Skewness 4. Covariance and correlation 5. Autocorrelation MSc Induction

More information

Testing and Model Selection

Testing and Model Selection Testing and Model Selection This is another digression on general statistics: see PE App C.8.4. The EViews output for least squares, probit and logit includes some statistics relevant to testing hypotheses

More information

Chapter 15 Panel Data Models. Pooling Time-Series and Cross-Section Data

Chapter 15 Panel Data Models. Pooling Time-Series and Cross-Section Data Chapter 5 Panel Data Models Pooling Time-Series and Cross-Section Data Sets of Regression Equations The topic can be introduced wh an example. A data set has 0 years of time series data (from 935 to 954)

More information

Chapter 10 Nonlinear Models

Chapter 10 Nonlinear Models Chapter 10 Nonlinear Models Nonlinear models can be classified into two categories. In the first category are models that are nonlinear in the variables, but still linear in terms of the unknown parameters.

More information

Instead of using all the sample observations for estimation, the suggested procedure is to divide the data set

Instead of using all the sample observations for estimation, the suggested procedure is to divide the data set Chow forecast test: Instead of using all the sample observations for estimation, the suggested procedure is to divide the data set of N sample observations into N 1 observations to be used for estimation

More information

Topic 14: Inference in Multiple Regression

Topic 14: Inference in Multiple Regression Topic 14: Inference in Multiple Regression Outline Review multiple linear regression Inference of regression coefficients Application to book example Inference of mean Application to book example Inference

More information

Variance Decomposition in Regression James M. Murray, Ph.D. University of Wisconsin - La Crosse Updated: October 04, 2017

Variance Decomposition in Regression James M. Murray, Ph.D. University of Wisconsin - La Crosse Updated: October 04, 2017 Variance Decomposition in Regression James M. Murray, Ph.D. University of Wisconsin - La Crosse Updated: October 04, 2017 PDF file location: http://www.murraylax.org/rtutorials/regression_anovatable.pdf

More information

Chapter 14 Simple Linear Regression (A)

Chapter 14 Simple Linear Regression (A) Chapter 14 Simple Linear Regression (A) 1. Characteristics Managerial decisions often are based on the relationship between two or more variables. can be used to develop an equation showing how the variables

More information

Business Statistics 41000: Homework # 5

Business Statistics 41000: Homework # 5 Business Statistics 41000: Homework # 5 Drew Creal Due date: Beginning of class in week # 10 Remarks: These questions cover Lectures #7, 8, and 9. Question # 1. Condence intervals and plug-in predictive

More information

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal Econ 3790: Business and Economics Statistics Instructor: Yogesh Uppal yuppal@ysu.edu Sampling Distribution of b 1 Expected value of b 1 : Variance of b 1 : E(b 1 ) = 1 Var(b 1 ) = σ 2 /SS x Estimate of

More information

Chapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics?

Chapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics? Econometrics and Quantitative Analysis Using Econometrics: A Practical Guide A.H. Studenmund 6th Edition. Addison Wesley Longman Chapter 1 An Overview of Regression Analysis Instructor: Dr. Samir Safi

More information

Midterm 2 - Solutions

Midterm 2 - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman Midterm 2 - Solutions You have until 10:20am to complete this exam. Please remember to put

More information

Unit 10: Simple Linear Regression and Correlation

Unit 10: Simple Linear Regression and Correlation Unit 10: Simple Linear Regression and Correlation Statistics 571: Statistical Methods Ramón V. León 6/28/2004 Unit 10 - Stat 571 - Ramón V. León 1 Introductory Remarks Regression analysis is a method for

More information

Chapter 10: Multiple Regression Analysis Introduction

Chapter 10: Multiple Regression Analysis Introduction Chapter 10: Multiple Regression Analysis Introduction Chapter 10 Outline Simple versus Multiple Regression Analysis Goal of Multiple Regression Analysis A One-Tailed Test: Downward Sloping Demand Theory

More information

Topic 1. Definitions

Topic 1. Definitions S Topic. Definitions. Scalar A scalar is a number. 2. Vector A vector is a column of numbers. 3. Linear combination A scalar times a vector plus a scalar times a vector, plus a scalar times a vector...

More information

1 Independent Practice: Hypothesis tests for one parameter:

1 Independent Practice: Hypothesis tests for one parameter: 1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)

More information

The simple linear regression model discussed in Chapter 13 was written as

The simple linear regression model discussed in Chapter 13 was written as 1519T_c14 03/27/2006 07:28 AM Page 614 Chapter Jose Luis Pelaez Inc/Blend Images/Getty Images, Inc./Getty Images, Inc. 14 Multiple Regression 14.1 Multiple Regression Analysis 14.2 Assumptions of the Multiple

More information

TESTING FOR CO-INTEGRATION

TESTING FOR CO-INTEGRATION Bo Sjö 2010-12-05 TESTING FOR CO-INTEGRATION To be used in combination with Sjö (2008) Testing for Unit Roots and Cointegration A Guide. Instructions: Use the Johansen method to test for Purchasing Power

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression Chapter 14 Student Lecture Notes 14-1 Department of Quantitative Methods & Information Systems Business Statistics Chapter 14 Multiple Regression QMIS 0 Dr. Mohammad Zainal Chapter Goals After completing

More information

Ch 13 & 14 - Regression Analysis

Ch 13 & 14 - Regression Analysis Ch 3 & 4 - Regression Analysis Simple Regression Model I. Multiple Choice:. A simple regression is a regression model that contains a. only one independent variable b. only one dependent variable c. more

More information

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006 Chapter 17 Simple Linear Regression and Correlation 17.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

Ordinary Least Squares Regression Explained: Vartanian

Ordinary Least Squares Regression Explained: Vartanian Ordinary Least Squares Regression Explained: Vartanian When to Use Ordinary Least Squares Regression Analysis A. Variable types. When you have an interval/ratio scale dependent variable.. When your independent

More information

Economics 618B: Time Series Analysis Department of Economics State University of New York at Binghamton

Economics 618B: Time Series Analysis Department of Economics State University of New York at Binghamton Problem Set #1 1. Generate n =500random numbers from both the uniform 1 (U [0, 1], uniformbetween zero and one) and exponential λ exp ( λx) (set λ =2and let x U [0, 1]) b a distributions. Plot the histograms

More information

Chapter 16. Simple Linear Regression and Correlation

Chapter 16. Simple Linear Regression and Correlation Chapter 16 Simple Linear Regression and Correlation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only

More information

Bayesian Analysis LEARNING OBJECTIVES. Calculating Revised Probabilities. Calculating Revised Probabilities. Calculating Revised Probabilities

Bayesian Analysis LEARNING OBJECTIVES. Calculating Revised Probabilities. Calculating Revised Probabilities. Calculating Revised Probabilities Valua%on and pricing (November 5, 2013) LEARNING OBJECTIVES Lecture 7 Decision making (part 3) Regression theory Olivier J. de Jong, LL.M., MM., MBA, CFD, CFFA, AA www.olivierdejong.com 1. List the steps

More information