Answer Key: Problem Set 6

Similar documents
Solutions to Problem Set 4 (Due November 13) Maximum number of points for Problem set 4 is: 66. Problem C 6.1

Answer Key: Problem Set 5

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is

CHAPTER 7. + ˆ δ. (1 nopc) + ˆ β1. =.157, so the new intercept is = The coefficient on nopc is.157.

Chapter 9: The Regression Model with Qualitative Information: Binary Variables (Dummies)

Problem 13.5 (10 points)

ECON 482 / WH Hong Binary or Dummy Variables 1. Qualitative Information

ECO375 Tutorial 7 Heteroscedasticity

At this point, if you ve done everything correctly, you should have data that looks something like:

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

Solutions to Problem Set 5 (Due November 22) Maximum number of points for Problem set 5 is: 220. Problem 7.3

Applied Statistics and Econometrics

ECO375 Tutorial 4 Wooldridge: Chapter 6 and 7

Unit 6 - Introduction to linear regression

Final Exam - Solutions

ECON3150/4150 Spring 2015

Multiple Regression Analysis

Multiple Regression Analysis: Heteroskedasticity

CHAPTER 4. > 0, where β

Inference in Regression Analysis

Multiple Regression Analysis

Can you tell the relationship between students SAT scores and their college grades?

Intro to Linear Regression

Intro to Linear Regression

ECNS 561 Topics in Multiple Regression Analysis

Simple Linear Regression: One Qualitative IV

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Intermediate Econometrics

Econometrics. 4) Statistical inference

Introduction to Econometrics. Heteroskedasticity

Making sense of Econometrics: Basics

Unit 6 - Simple linear regression

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

Econometrics Review questions for exam

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Economics 471: Econometrics Department of Economics, Finance and Legal Studies University of Alabama

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Answer Key. 9.1 Scatter Plots and Linear Correlation. Chapter 9 Regression and Correlation. CK-12 Advanced Probability and Statistics Concepts 1

Problem C7.10. points = exper.072 exper guard forward (1.18) (.33) (.024) (1.00) (1.00)

Homework Set 3, ECO 311, Spring 2014

Lab 11 - Heteroskedasticity

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company


Business Statistics. Lecture 9: Simple Regression

Marketing Research Session 10 Hypothesis Testing with Simple Random samples (Chapter 12)

Multiple Linear Regression

Course Econometrics I

Lecture 18: Simple Linear Regression

ECON 497 Midterm Spring

Econometrics - 30C00200

ECON 497: Lecture Notes 10 Page 1 of 1

Reliability of inference (1 of 2 lectures)

OSU Economics 444: Elementary Econometrics. Ch.10 Heteroskedasticity

Lecture 11 Multiple Linear Regression

2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0

15.8 MULTIPLE REGRESSION WITH MANY EXPLANATORY VARIABLES

General Linear Model (Chapter 4)

The Simple Linear Regression Model

Chapter 10. Correlation and Regression. McGraw-Hill, Bluman, 7th ed., Chapter 10 1

Applied Statistics and Econometrics

Correlation and Linear Regression

Chapter 13. Multiple Regression and Model Building

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

Lecture 4: Heteroskedasticity

Heteroskedasticity. Part VII. Heteroskedasticity

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Lecture 8: Heteroskedasticity. Causes Consequences Detection Fixes

Econometrics Multiple Regression Analysis: Heteroskedasticity

Chapter 8 Heteroskedasticity

In Class Review Exercises Vartanian: SW 540

Lecture 10 Multiple Linear Regression

Rockefeller College University at Albany

PBAF 528 Week 8. B. Regression Residuals These properties have implications for the residuals of the regression.

Econometrics -- Final Exam (Sample)

ST430 Exam 1 with Answers

Economics Introduction to Econometrics - Fall 2007 Final Exam - Answers

Lecture 3: Inference in SLR

Multiple Regression: Chapter 13. July 24, 2015

Regression Models - Introduction

Sociology 593 Exam 2 Answer Key March 28, 2002

ECON 4230 Intermediate Econometric Theory Exam

Business Statistics. Lecture 10: Course Review

Lecture 14. Analysis of Variance * Correlation and Regression. The McGraw-Hill Companies, Inc., 2000

Lecture 14. Outline. Outline. Analysis of Variance * Correlation and Regression Analysis of Variance (ANOVA)

(a) Briefly discuss the advantage of using panel data in this situation rather than pure crosssections

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

y response variable x 1, x 2,, x k -- a set of explanatory variables

LECTURE 11. Introduction to Econometrics. Autocorrelation

Statistics and Quantitative Analysis U4320

Solutions to Problem Set 6 (Due December 8) Maximum number of points for Problem set 8 is: 220. Problem 10.6

Introductory Econometrics

Information Sources. Class webpage (also linked to my.ucdavis page for the class):

ECON3150/4150 Spring 2016

WISE International Masters

Homoskedasticity. Var (u X) = σ 2. (23)

mrw.dat is used in Section 14.2 to illustrate heteroskedasticity-robust tests of linear restrictions.

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables

Transcription:

: Problem Set 6 1. Consider a linear model to explain monthly beer consumption: beer = + inc + price + educ + female + u 0 1 3 4 E ( u inc, price, educ, female ) = 0 ( u inc price educ female) σ inc var,,, =. Write the transformed equation that ahs a homoskedastic error term. With Var(u inc,price,educ,female) = σ inc, h(x) = inc, where h(x) is the heteroskedasticity function defined in equation (8.1). Therefore, h( x ) = inc, and so the transformed equation is obtained by dividing the original equation by inc: beer = 0(1/ inc) + 1+ ( price / inc) + 3( educ / inc) + 4( female / inc) + ( u / inc). inc Notice that 1, which is the slope on inc in the original model, is now a constant in the transformed equation. This is simply a consequence of the form of the heteroskedasticity and the functional forms of the explanatory variables in the original equation.. Using the data in GPA3.DTA, the following equation was estimated for the fall and second semester students: trmgpa =.1 + 0.900 + 0.193 cumgpa + 0.0014 tothrs (0.55) (0.175) (0.064) (0.001) [0.55] [0.166] [0.074] [0.001] + 0.0018sat 0.0039hsperc + 0.351female 0.157season (0.000) (0.0018) (0.085) (0.098) [0.000] [0.0019] [0.079] [0.080] n = 69, R = 0.465 Here, trmgpa is term GPA, is a weighted average of overall GPA in courses taken, cumgpa is GPA prior to the current semester, tothrs is total credit hours prior to the semester, sat is SAT score, hsperc is graduating percentile in high school class, female is a gender dummy, and season is a dummy variable equal to unity if the 1

student's sport is in season during the fall. The usual and heteroskedasticity-robust standard errors are reported in parentheses and brackets, respectively. i. Do the variables, cumgpa, and tothrs have the expected estimated effects? Which of these variables are statistically significant at the 5% level? Does it matter which standard errors are used? These coefficients have the anticipated signs. If a student takes courses where grades are, on average, higher as reflected by higher then his/her grades will be higher. The better the student has been in the past as measured by cumgpa the better the student does (on average) in the current semester. Finally, tothrs is a measure of experience, and its coefficient indicates an increasing return to experience. The t statistic for is very large, over five using the usual standard error (which is the largest of the two). Using the robust standard error for cumgpa, its t statistic is about.61, which is also significant at the 5% level. The t statistic for tothrs is only about 1.17 using either standard error, so it is not significant at the 5% level. ii. Why does the hypothesis H = 1 make sense? Test this hypothesis against 0 : the two-sided alternative at the 5% level, using both standard errors. Describe your conclusion. This is easiest to see without other explanatory variables in the model. If were the only explanatory variable, H 0 : = 1 means that, without any information about the student, the best predictor of term GPA is the average GPA in the students courses; this holds essentially by definition. (The intercept would be zero in this case.) With additional explanatory variables it is not necessarily true that = 1 because could be correlated with characteristics of the student. (For example, perhaps the courses students take are influenced by ability as measured by test scores and past college performance.) But it is still interesting to test this hypothesis. The t statistic using the usual standard error is t = (.900 1)/.175.57; using the heteroskedasticity-robust standard error gives t.60. In either case we fail to reject H 0 : = 1 at any reasonable significance level, certainly including 5%.

iii. Test whether there is an in-season effect on term GPA, using both standard errors,. Does the significance level at which the null can be rejected depend on the standard error used? The in-season effect is given by the coefficient on season, which implies that, other things equal, an athlete s GPA is about.16 points lower when his/her sport is competing. The t statistic using the usual standard error is about 1.60, while that using the robust standard error is about 1.96. Against a two-sided alternative, the t statistic using the robust standard error is just significant at the 5% level (the standard normal critical value is 1.96), while using the usual standard error, the t statistic is not quite significant at the 10% level (cv 1.65). So the standard error used makes a difference in this case. This example is somewhat unusual, as the robust standard error is more often the larger of the two. Computer Exercise 3. Use the data set GPA1.DTA for this exercise i. Use OLS to estimate a model relating colgpa to hsgpa, ACT, skipped, and PC. Obtain the OLS residuals The equation estimated by OLS is colgpa = 1.36 +.41 hsgpa +.013 ACT.071 skipped +.14 PC (.33) (.09) (.010) (.06) (.057) n = 141, R =.59, R =.38 ii. Compute the special case of the White test for heteroskedasticity. In the regression of u ˆi on colgpa i, colgpa i, obtain the fitted values, say h ˆi. The F statistic obtained for the White test is about 3.58. With and 138 df, this gives p-value.031. So, at the 5% level, we conclude there is evidence of heteroskedasticity in the errors of the colgpa equation. (As an aside, note that the t statistics for each of the terms is very small, and we could have simply dropped the 3

quadratic term without losing anything of value.) iii. Verify that the fitted values from part ii are all strictly positive. Then, obtain the weighted least squares estimates using weights 1/ h ˆ i. Compare the weighted least squares estimates for the effect of skipping lectures and the effect of PC ownership with the corresponding OLS estimates. What about their statistical significance? In fact, the smallest fitted value from the regression in part (ii) is about.07, while the largest is about.165. Using these fitted values as the squares regression gives the following: hˆi in a weighted least colgpa = 1.40 +.40 hsgpa +.013 ACT.076 skipped +.16 PC (.30) (.083) (.010) (.0) (.056) n = 141, R =.306, R =.86 There is very little difference in the estimated coefficient on PC, and the OLS t statistic and WLS t statistic are also very close. Note that we have used the usual OLS standard error, even though it would be more appropriate to use the heteroskedasticity-robust form (since we have evidence of heteroskedasticity). The R-squared in the weighted least squares estimation is larger than that from the OLS regression in part (i), but, remember, these are not comparable. iv. With the WLS estimation from part iii, obtain hetroskedasticity-robust standard errors. In other words, allow for the fact that the variance function estimated in part ii might be misspecified. Do the standard error change much from part iii? [Hint: In Stata, use option 'robust' at the end of the 'regress' command to get the heteroskedasticity-robust standard errors.] With robust standard errors that is, with standard errors that are robust to misspecifying the function h(x) the equation is colgpa = 1.40 +.40 hsgpa +.013 ACT.076 skipped +.16 PC (.31) (.086) (.010) (.01) (.059) n = 141, R =.306, R =.86 4

The robust standard errors do not differ by much from those in part (iii); in most cases, they are slightly higher, but all explanatory variables that were statistically significant before are still statistically significant. But the confidence interval for PC is a bit wider. 5