Rockefeller College University at Albany

Similar documents
CHAPTER 6: SPECIFICATION VARIABLES

WISE International Masters

Rockefeller College University at Albany

Econometrics Summary Algebraic and Statistical Preliminaries

Economics 308: Econometrics Professor Moody

2. Linear regression with multiple regressors

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Lecture 4: Multivariate Regression, Part 2

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

FinQuiz Notes

Lecture 4: Multivariate Regression, Part 2

ECON 4230 Intermediate Econometric Theory Exam

LECTURE 11. Introduction to Econometrics. Autocorrelation

Final Exam - Solutions

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =

Statistical Inference with Regression Analysis

Multiple Regression Analysis

Making sense of Econometrics: Basics

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

Econometrics Review questions for exam


ECON 497 Midterm Spring

Handout 12. Endogeneity & Simultaneous Equation Models

Econometrics Part Three

Iris Wang.

2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0

Ec1123 Section 7 Instrumental Variables

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

ECONOMETRICS HONOR S EXAM REVIEW SESSION

Econometrics. 9) Heteroscedasticity and autocorrelation

The Simple Linear Regression Model

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

ECON Introductory Econometrics. Lecture 16: Instrumental variables

ECON 497: Lecture 4 Page 1 of 1

Lab 11 - Heteroskedasticity

PBAF 528 Week 8. B. Regression Residuals These properties have implications for the residuals of the regression.

Basic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

WISE International Masters

Final Exam - Solutions

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Multiple Linear Regression CIVL 7012/8012

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

Lecture 5: Omitted Variables, Dummy Variables and Multicollinearity

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Econometrics - 30C00200

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Binary Dependent Variables

ECON 497: Lecture Notes 10 Page 1 of 1

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 18th Class 7/2/10

Christopher Dougherty London School of Economics and Political Science

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Environmental Econometrics

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Lecture 4: Heteroskedasticity

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

Applied Quantitative Methods II


ECNS 561 Multiple Regression Analysis

Topic 7: Heteroskedasticity

Instrumental Variables and the Problem of Endogeneity

ACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.

Multiple Regression Analysis

New York University Department of Economics. Applied Statistics and Econometrics G Spring 2013

1 Motivation for Instrumental Variable (IV) Regression

Answer Key: Problem Set 6

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

1 Regression with Time Series Variables

An overview of applied econometrics

Applied Statistics and Econometrics

2 Prediction and Analysis of Variance

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold.

Lectures 5 & 6: Hypothesis Testing

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10)

Semester 2, 2015/2016

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Graduate Econometrics Lecture 4: Heteroskedasticity

Steps in Regression Analysis

Lecture 1: intro. to regresions

1 A Non-technical Introduction to Regression

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Chapter 7. Testing Linear Restrictions on Regression Coefficients

Gov 2000: 9. Regression with Two Independent Variables

Chapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics?

Heteroskedasticity. Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set

Measurement Error. Often a data set will contain imperfect measures of the data we would ideally like.

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

Reliability of inference (1 of 2 lectures)

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Econometrics Midterm Examination Answers

Mid-term exam Practice problems

Econ 510 B. Brown Spring 2014 Final Exam Answers

Panel Data Exercises Manuel Arellano. Using panel data, a researcher considers the estimation of the following system:

Handout 11: Measurement Error

Testing Linear Restrictions: cont.

Transcription:

Rockefeller College University at Albany PAD 705 Handout: Suggested Review Problems from Pindyck & Rubinfeld Original prepared by Professor Suzanne Cooper John F. Kennedy School of Government, Harvard University Here are a few recommended problems from Pindyck and Rubinfeld. For the mid-term, try the questions from Chapters 4 7 listed below; for the final, attempt all problems. There may be other problems in Pindyck & Rubinfeld that are relevant; the answers to all are found at the back of the textbook. However, these questions are most directly relevant to the course and Professor Cooper at the Kennedy School made her answer key available to me for the ones suggested below. Since I m borrowing this from Professor Cooper, it is possible that some explanations go beyond what I have taught in the course. Rest assured that the exam will only cover topics I have lectured upon at the level of depth I have presented. Chapter 4: 4.3 Chapter 5: 5.2, 5.3, 5.4(a), 5.5, 5.6 1 do not recommend running these regressions in preparation for the exams. The regression results are (t-statistics in parentheses): I. RENTPER = 35.92 + 20.02*SEX + 118.00*ROOMPER - 0.79*DIST (1.05) (1.21) (3.02) (-1.35) R 2 =.36 ESS = 44069 N = 32 II. RENTPER = 32.69 + 132.06*ROOMPER - 1.03*DIST (0.95) (3.51) (1.88) R 2 =.33 ESS = 46362 N = 32 III. RENTPER = 30.72 + 99.93*SEX + 130.51*ROOMPER - 1.15 *DIST (0.91) (0.83) (3.35) (-2.00) -121.94*(ROOMPER*SEX) + 4.56*(DIST*SEX) (-0.96) (2.22) R 2 =.47 ESS = 36707 N = 32 Revised: March 17, 2003

IV. RENTPER = 30.72 + 130.51*ROOMPER - 1.15*DIST (0.87) (3.20) (-0.60) (males only) R 2 =.38 ESS = 29357 N = 22 V. RENTPER = 130.65 + 8.57*ROOMPER + 3.41 *DIST (females only) (1.31) (0.08) (2.00) R 2 =.38 ESS = 7350.3 N = 10 Chapter 6: 6.1, 6.3 Chapter 7: 7.1, 7.7 Chapter 11: 11.2 a - d (Answers to parts e and f are included below, but they are more advanced and go beyond the level at which we will discuss the material in class.) 11.7 Chapter 10: 10.7a (Do not do the estimation; he results are as follows, with standard errors in parentheses.) Lin. Prob Probit Logit PUB 1-2.11 (.14).37 (.42).32 (.38) PUB3-4.22 (.16).70 (.46).62 (.42) PUB5.10 (.26).30 (.77).29 (.70) PRIV -.068 (.17) -.21 (.48) -.19 (.43) TYEAR -.0055 (.0055) -.016 (.016) -.014 (.015) SCHL.31 (.15) 1.58 (.78) 1.45 (.78) INCCON.38 (.41) 1.31 (.46) 1.20 (.43) PTCON -.41 (.18) -1.46 (.63) -1.32 (.60) 2

Suggested Answers for Practice Problems from Pindyck & Rubinfeld 4.3: Since consumption is exactly equal (in this simple model) to income minus savings, then if we run a regression of consumption on income and savings, we should get a perfect fit, with the coefficient on income equal to 1, and the coefficient on savings equal to -1 (and the intercept equal to zero). In other words, we would find β 2 = 1 and β 3 = -1, we would have C = 1*Y - 1*S. This regression would fit perfectly, i.e., the error sum of squares would be zero and the regression sum of squares would equal the total sum of squares (total variation of C about its mean). 5.2: a) Given the regression results, we can test the null hypothesis that β 3 = 0 We can use a t-test since we know that the least squares regression estimate divided by its standard error is distributed as t N-k. We are, in fact, given the t-statistics directly, and do not even have to calculate them in this case. The t-statistic for the null hypothesis β 3 = 0 is 3.02, which exceeds the 95% critical value (2.05 for a 2-tailed test or 1.70 for a one-tailed test, since there are 32 4 = 28 degrees of freedom), so we can reject the null hypothesis at the 95% level. b) Similarly, testing the null hypothesis that β 4 = 0, the t-statistic is -1.35, which does not exceed the one-tailed or two-tailed test critical values at the 95% level. Therefore we do not reject the null hypothesis. c) Using a t-test, we do not reject the null hypothesis at the 95% level since the t-statistic is 1.21, which does not exceed the 95% critical value (2.05 for a two-tailed test). Using an F-test, use the sum of squared residuals from equations I and II, where I is the unrestricted and II is the restricted regression, where the null hypothesis is imposed. The F-statistic is: ( ESS ) / R ESS q U (46362 44069) /1 = = 1.46 / N l 44069 / 32 4 ESS U Since 1.46 is less than the 95% critical value for the F 1,28 distribution (which is 4.17), we do not reject the null hypothesis that the coefficient on SEX is zero. Note that the F value is the square of the t value (i.e., 1.46 = 1.21 2 ). 5.3: a) In regression III there are six parameters estimated. So the answer to this part of the question is straightforward (i.e., a simple t-test), but since the degrees of freedom is a relatively small number here, we need to look up the critical value in the table. For 32-6 = 26 degrees of freedom, the 95% critical value is -2.06. Since the calculated t-statistic is -0.96, we do not reject the null hypothesis b5=0 at the 95% level. Similarly, for the null hypothesis on β 6 = 0, the t-statistic (2.22) exceeds the 95% critical value for a two-tailed test (2.06) so we reject the null hypothesis in favor of the alternative, b6 not equal 0 b) Using the formula for the F-test from 5.2c above, (44069 36707) 36707 / 32 6 / 2 = 2.61 since the restricted regression is I and the unrestricted regression is III. Since the critical value at the 95% level for 2,26 degrees of freedom is 3.39 (well, between 3.32 and 3.39), we do not reject the null hypothesis that the coefficients are jointly zero. 2 c) Given the values for R, computed the corrected value involves adjusting for degrees of freedom: 3

R = 1 (1 For model I, we have (1 - (1 -.36)(31/28)) =.29 For model II, we have (1 - (1 -.33)(31/29)) =.28 For model III, we have (1 - (1 -.47)(31/26)) =.37 2 R N 1 ) N k 5.4: a) Using two-tailed t-tests, we do not reject the null hypothesis that γ 1 = 0 (since the t-statistic is.08, which does not exceed the 95% critical value for 10 3 = 7 degrees of freedom). We also do not reject the null hypothesis that γ 2 = 0. 5.5) Since regression III includes a full set of dummy variables for SEX (i.e., allows for different intercept and different slopes for ROOMPER and DIST), we can get parameter estimates for females either from regression III or from regression V. Using regression III, the equation for females is obtained by setting SEX = 1 The intercept is then 30.72 + 99.93 = 130.65, which is the intercept in the equation estimated only for females (regression V). Similarly, the coefficient on ROOMPER for females from regression III is 130.51-121.94 = 8.57, which is the estimate from regression V, estimated on females only. Similarly for the coefficient on DIST, from regression III is -1.15 + 4.56 = 3.41 for females, which is what we find in regression V. Intuitively, with full intercept and interactive dummy variables we can get the intercept and slope values for either the group for which the dummy variable is zero, or for the group where it equals one. We will get exactly the same coefficient estimates if we estimate the equations separately for the two groups. 5.6: The coefficients in regression V are just the regression estimates for females alone. Regression IV gives us the estimates for males alone. Then regression III uses a full set of interactive dummy variables so we can see not only the estimates for males and females separately, but the incremental differences between them, which are seen as the coefficients on the interactive dummy variables in regression III, i.e., 2 γ 1 = β 3 + β 5, γ2 = β 4 + β 6 Intuitively, the difference between using full intercept and interactive dummy variables vs. estimating the equation separately for the two groups is not in the coefficient estimates (although recall that the standard errors can be affected). The coefficient on the interactive SEX*DIST variable gives the increment to the slope of RENTPER with respect to DIST for females. The coefficient on DIST gives the slope for males, and the sum of the coefficients on DIST and SEX*DIST gives the slope for females. So the coefficient on SEX*DIST is the increment to the slope for females vs. males. We might not expect that γ 2 < 0 since an apartment further from campus would likely cost less to rent. If, however, we have omitted variables that are positively correlated with distance (like how nice the apartment is), then the coefficient on distance might pick up the effect of quality. The omitted quality variable could bias the coefficient on distance, making it positive when in fact the true value is negative. 6. 1: When there is heteroskedasticity, OLS will tend to give a lot of weight to those observations with high variance. However, those observations actually contain less information about the true value of the regression parameter than observations with small variance (since those with small variance lie closer to the true regression line). Weighted least squares transforms the errors (and the whole regression equation) so that least squares on the weighted version does not give more weight to those observations with large variance. Rather, we get more information about the true regression parameter by giving more weight to those observations with smaller variance. In other words, if we use OLS, it will try very hard to fit the 4

points with large variance. This will yield a high variance for our slope estimate. We would prefer to give more weight to those observations with small variance to get a more precise (i.e., efficient) estimate of the regression parameter. This is what WLS does. 6.3: a) Least squares would still be unbiased and consistent, but the standard errors would be computed wrongly (unless we specify robust standard errors). Heteroskedasticity does not affect consistency or unbiasedness. However, even once we compute the standard errors properly, OLS will not be efficient for the reason given in 6.1 above (and more generally, since we violate an assumption of the classical linear regression model, the Gauss-Markov theorem no longer applies, and we do not have the BEST linear unbiased estimates). b) To get efficient estimates we would need to use weighted least squares. Since we are given the form of the variance (the variance for the large firms is 2 times the variance for the small firms), we can just weight each observation by its variance, i.e., the small firms observations (on the dependent and all independent variables) can be weighted by 1 (i.e., don t need to do anything, and their variance will just be the variance for the small firms), and the large firms observations can be weighted by 1/sqrt(2) (so their variance will now be 1/2 their old variance, which was twice the variance for the small firms, so now their variance is ½ * 2= 1 * the variance for the small firms). After doing this, the variance of the error for all observations will be the same. We don t know what that value is (it s just the variance for the observations on small firms), but as long as it is a constant we will get back to efficient estimates. c) We could use a Goldfeld-Quandt test (which is what hettest in Stata does). Intuitively, we would split the sample into small firms and large firms (no need to drop any observations since they fit nicely into two groups). We would fit least squares and get the ESS from each group. The ratio of the ESS values is an estimate of the ratio of the variances for the two groups and is distributed as F. We can compare this value to the critical value from the F distribution with numerator degrees of freedom equal to the number of observations on large firms minus the number of parameters estimated, and the denominator degrees of freedom equals the number of observations on small firms minus the number of parameters estimated. If the computed value exceeds the critical value, we reject the null hypothesis that the variances are the same. 7. 1: Measurement error in the dependent variable appears in the error term in the regression. However, it is not generally correlated with any of the independent variables. Therefore, it does not induce any correlation between independent variables and the error in the regression. When an independent variable is measured with error, this measurement error shows up in the error of the regression, and is in addition correlated with the observed value of the independent variable, inducing a correlation between the independent variable and the error in the regression. This leads to biased and inconsistent regression parameter estimates as least squares does not take into account this correlation. 7.7: We lose only efficiency by including an irrelevant variable. Since including such a variable does not lead to any correlation between error and independent variables (as would be the case for an omitted variable), we have unbiased and consistent estimates even with irrelevant variables. However, because an additional parameter needs to be estimated, we lose efficiency. We are violating the assumption of the classical linear regression model that we have correctly specified the regression model, and therefore we no longer have the BEST estimates, although they are still linear and unbiased. In addition, it should be the case that we obtain an unbiased estimate of the coefficient on the irrelevant variable, where its true value is zero. If the intercept belongs in the regression and we leave it out, however, we can get biased and inconsistent parameter estimates on other coefficients of interest. 5

11.2 a) To address the identification question, first categorize variables as either endogenous or predetermined, and then apply the order condition. In this particular problem (which is a classic case, and you have seen it before), Q D, Q S and P are endogenous (to see this just use what you know about supply and demand and that price is also determined by quantity supplied and demanded) and the lagged price level and Y are predetermined (again, using your knowledge of supply and demand, in this case). The supply equation then includes 2 endogenous variables (quantity supplied and price) and there are 2 excluded predetermined variables (i.e., lagged price and income are not included in that equation). Therefore, the supply equation is overidentified since #included endogenous -1 < #excluded predetermined. Using OLS will give biased and inconsistent parameter estimates since there is an endogenous variable on the right hand side which due to its endogeneity will be correlated with the error in the supply equation (i.e., price is determined in part by quantity supplied and quantity supplied is determined in part by the error in the supply equation, so price and the error will be correlated). b) Using the same analysis as in (a), the demand equation includes 2 endogenous variables (quantity demanded and price) and no excluded predetermined variables (since both predetermined variables are included in the demand equation). Therefore the demand equation is not identified. OLS will be biased and inconsistent for the same reason as in (a) above. c) To estimate the supply equation with IV, we would need an instrument. Both lagged price and current income are possible instruments. Since they are predetermined variables in the system of equations, they are correlated with price and can therefore be used as instruments for price. Of course, since the supply equation is overidentified, using IV involves choosing just one of the available instruments. We would choose either lagged price or current income and use that variable as an instrument for current price, using the IV estimation procedure. d) Using 2SLS permits using both available predetermined variables as to form an instrument for current price, rather than choosing just one variable as in (c) above. The procedure would be to regress current price on lagged price and current income. The fitted values from that regression would be a predetermined variable since it would be determined entirely by the two predetermined variables. Therefore, the fitted values from the first stage regression can be used in place of the price variable in the supply equation. When this instrument is used in place of current price, there is no longer an endogeneity problem since there is now a predetermined variable on the right hand side of the supply equation. e) ILS cannot be used to estimate the demand equation since it is not identified. Furthermore, OLS will be biased and inconsistent because of the presence of the price variable (which is endogenous) on the right hand side of the equation. f) If the error in the supply equation is autocorrelated (i.e., serially correlated), then this period s error is correlated with last period s error. Also, since this period s error is correlated with this period s price, and last period s error is correlated with last period s price, it follows that last period s price is correlated with this period s error. Therefore, last period s price is not a valid predetermined variable. 11.7 The endogenous variables are presumably Y 1, Y 2, Y 3 and the predetermined variables are presumably X 1, X 2 (this is a pretty stylized example where it s easy because the book labels endogenous variables Y and predetermined variables X ). Then the first equation includes 2 endogenous variables and there are no excluded predetermined variables, so the first equation is unidentified. The second equation includes 2 endogenous variables and there is I excluded predetermined variable, so #included endogenous - 1 = #excluded predetermined, so the second equation is exactly identified. The third equation includes 2 endogenous variables and there are 2 excluded predetermined variables, so the third equation is overidentified. 6

10.7a All three models yield different coefficients, but recall that the interpretation of the coefficients is also different in each case, so we should not be surprised that the magnitudes differ. In particular, in the linear probability model, each coefficient is interpreted as the marginal effect of that variable on Pr(Y = 1), holding other variables constant. In the probit model, each coefficient is the marginal effect on Z, where F(Z) = Pr(Y = 1), holding other variables constant, and similarly for the logit model, but where the function F(Z) is different since it is a different CDF. Note, however, that the signs of the coefficients are the same in each case, which makes sense since a negative coefficient in the linear probability model suggests that the effect of that variable on Pr(Y = 1) is negative, and in the probit or logit a negative coefficient means that the effect of that variable on Z is negative, and since F(Z) for both the probit and logit doesn t change the sign, a negative effect on Z leads to a negative effect on F(Z), and therefore on Pr(Y = 1). Also, coefficients that are large in magnitude relative to other coefficients in the linear probability model are also large relative to others in the probit and logit. Significance of particular coefficients is also roughly comparable across the linear probability, probit, and logit specifications. 7