Regression #8: Loose Ends

Similar documents
Lecture 4: Multivariate Regression, Part 2

Problem Set #5-Key Sonoma State University Dr. Cuellar Economics 317- Introduction to Econometrics

Lecture 4: Multivariate Regression, Part 2

Statistical Inference with Regression Analysis

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Problem Set 10: Panel Data

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11

ECON 5350 Class Notes Functional Form and Structural Change

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is

Practice exam questions

Lab 07 Introduction to Econometrics

Applied Statistics and Econometrics

Introductory Econometrics. Lecture 13: Hypothesis testing in the multiple regression model, Part 1

Graduate Econometrics Lecture 4: Heteroskedasticity

Applied Statistics and Econometrics

Lab 6 - Simple Regression

ECON Introductory Econometrics. Lecture 16: Instrumental variables

THE MULTIVARIATE LINEAR REGRESSION MODEL

Lecture 3: Multiple Regression

Computer Exercise 3 Answers Hypothesis Testing

Economics 326 Methods of Empirical Research in Economics. Lecture 14: Hypothesis testing in the multiple regression model, Part 2

Applied Statistics and Econometrics

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests

Lab 10 - Binary Variables

Problem Set #3-Key. wage Coef. Std. Err. t P> t [95% Conf. Interval]

Applied Statistics and Econometrics

Nonlinear Regression Functions

2.1. Consider the following production function, known in the literature as the transcendental production function (TPF).

Multiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor)

1. The shoe size of five randomly selected men in the class is 7, 7.5, 6, 6.5 the shoe size of 4 randomly selected women is 6, 5.

Multivariate Regression: Part I

Lecture 7: OLS with qualitative information

Handout 12. Endogeneity & Simultaneous Equation Models

Multiple Regression: Inference

Making sense of Econometrics: Basics

Simultaneous Equations with Error Components. Mike Bronner Marko Ledic Anja Breitwieser

ECON3150/4150 Spring 2015

Model Specification and Data Problems. Part VIII

The Simple Regression Model. Part II. The Simple Regression Model

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors

5.2. a. Unobserved factors that tend to make an individual healthier also tend

Econometrics Homework 1

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58

ECON3150/4150 Spring 2016

Please discuss each of the 3 problems on a separate sheet of paper, not just on a separate page!

7 Introduction to Time Series

The regression model with one fixed regressor cont d

University of California at Berkeley Fall Introductory Applied Econometrics Final examination. Scores add up to 125 points

7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15

ECON Introductory Econometrics. Lecture 17: Experiments

Econometrics Midterm Examination Answers

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10)

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

Essential of Simple regression

ECO220Y Simple Regression: Testing the Slope

4 Instrumental Variables Single endogenous variable One continuous instrument. 2

ECON3150/4150 Spring 2016

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators

Econometrics I Lecture 7: Dummy Variables

Applied Quantitative Methods II

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables

2. (3.5) (iii) Simply drop one of the independent variables, say leisure: GP A = β 0 + β 1 study + β 2 sleep + β 3 work + u.

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

1 Independent Practice: Hypothesis tests for one parameter:

Lecture 5: Omitted Variables, Dummy Variables and Multicollinearity

Lecture 8: Instrumental Variables Estimation

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Question 1 carries a weight of 25%; Question 2 carries 20%; Question 3 carries 20%; Question 4 carries 35%.

ECON Introductory Econometrics. Lecture 13: Internal and external validity

Interpreting coefficients for transformed variables

4 Instrumental Variables Single endogenous variable One continuous instrument. 2

Analysing data: regression and correlation S6 and S7

Empirical Application of Simple Regression (Chapter 2)

Data Analysis 1 LINEAR REGRESSION. Chapter 03

Course Econometrics I

Econometrics. 9) Heteroscedasticity and autocorrelation

Chapter 6: Linear Regression With Multiple Regressors

Binary Dependent Variables

Statistical Inference. Part IV. Statistical Inference

Lecture 3: Multiple Regression. Prof. Sharyn O Halloran Sustainable Development U9611 Econometrics II

1: a b c d e 2: a b c d e 3: a b c d e 4: a b c d e 5: a b c d e. 6: a b c d e 7: a b c d e 8: a b c d e 9: a b c d e 10: a b c d e

sociology 362 regression

1: a b c d e 2: a b c d e 3: a b c d e 4: a b c d e 5: a b c d e. 6: a b c d e 7: a b c d e 8: a b c d e 9: a b c d e 10: a b c d e

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

Econometrics Review questions for exam

Linear Regression. Junhui Qian. October 27, 2014

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Ecmt 675: Econometrics I

Basic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler

Problem 13.5 (10 points)

ECON Introductory Econometrics. Lecture 11: Binary dependent variables

CHAPTER 6: SPECIFICATION VARIABLES

The linear model. Our models so far are linear. Change in Y due to change in X? See plots for: o age vs. ahe o carats vs.

4. Nonlinear regression functions

Transcription:

Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30

In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch on nonetheless. These include: 1 Multicollinearity 2 Coefficient interpretation with log transformations. 3 Dummy / Indicator Variables 4 Nonlinearities. Justin L. Tobias (Purdue) Regression #8 2 / 30

Multicollinearity So, what is multicollinearity? So, why is this important? 1 2 3 Justin L. Tobias (Purdue) Regression #8 3 / 30

Multicollinearity In a very real sense, the importance of multicollinearity, and its perception as a problem in econometrics, is overblown. For example, OLS estimators are still unbiased, consistent and efficient in the presence of high (but not perfect) collinearity. Goldberger (1991) likens the problem of multicollinearity to micronumerosity - the problem of having a small sample size. A series of interesting (and entertaining) quotes on this issue are taken from his book: Justin L. Tobias (Purdue) Regression #8 4 / 30

Multicollinearity The extreme case, exact micronumerosity arises when n = 0, in which case the sample estimate of µ is not unique... The extreme case is easy enough to recognize. Near micronumerosity is more subtle, and yet very serious. It arises when the rank condition n > 0 is barely satisfied... He continues by noting the similarity of consequences with multicollinearity... The consequences of micronumerosity are serious. Precision of estimation is reduced... Investigators will sometimes be led to acccept the hypothesis µ = 0... even though the true situation may be not that µ = 0 but that the sample data have not enablesd us to pick µ up. The estimate of µ will [also] be very sensitive to the sample data... Justin L. Tobias (Purdue) Regression #8 5 / 30

Multicollinearity Finally, he suggests some tests for micronumerosity... Tests for the presence of micronumerosity require the judicious use of various fingers. Some researchers prefer a single finger, others use their toes, still others let their thumbs rule. A generally reliable guide may be obtained by counting the number of observations. Most of the time in econometric analysis, when n is close to zero, it is also far from infinity. While these are entertaining, they illustrate that: 1 The problems associated with small sample sizes are like those associated with multicollinearity. 2 Multicollinearity does not violate any of our fundamental assumptions; it is simply a feature of the regression model itself. Large standard errors are not wrong or misleading as the coefficient estimates should vary a lot from sample to sample. Justin L. Tobias (Purdue) Regression #8 6 / 30

While multicollinearity may be bad, in the sense that individual t-statistics are small, leading the applied researcher to update his/her beliefs about the publishability of the work and subsequently wanting to throw himself/herself out the window, it can also aid in inference, (e.g., prediction), as the following example suggests: Suppose: and, to fix ideas, set σ 2 = 1. In addition, suppose that the explanatory variables have been scaled so that: In this case, since the off-diagonal, i x 1ix 2i is the sample correlation, denoted as ρ. Justin L. Tobias (Purdue) Regression #8 7 / 30

Multicollinearity: Example It follows that Thus, for the purposes of getting a small variance for ˆβ 1 and ˆβ 2, we would want to set ρ = 0. However, consider the parameter With very little work, we can show: For this parameter, it is clear that ρ > 0 leads to increased precision! Justin L. Tobias (Purdue) Regression #8 8 / 30

Multicollinearity: Example Note that this logic translates to the exercise of prediction. To this end, suppose we wish to predict y when x 1 and x 2 equal the same value, say c. (Note this is not completely unreasonable given our initial scaling of the data). Then, at which point the preference for a small variance associated with becomes clear. This argument also extends to general problems of prediction. Justin L. Tobias (Purdue) Regression #8 9 / 30

Consider two different regressions: y i = β 1 + β 2 Educ i + u i y i = θ 1 + θ 2 Educ i + θ 3 Educ 2 i + θ 4 Educ 3 i + θ 5 Educ 4 i + ɛ i. regress lwage educat Source SS df MS Number of obs = 1260 -------------+------------------------------ F( 1, 1258) = 95.89 Model 31.5149966 1 31.5149966 Prob > F = 0.0000 Residual 413.464976 1258.328668502 R-squared = 0.0708 -------------+------------------------------ Adj R-squared = 0.0701 Total 444.979972 1259.353439215 Root MSE =.5733 ------------------------------------------------------------------------------ lwage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- educat.0602839.0061563 9.79 0.000.0482061.0723616 _cons.9014239.0790132 11.41 0.000.7464117 1.056436 ------------------------------------------------------------------------------. regress lwage educat educat2 educat3 educat4 Source SS df MS Number of obs = 1260 -------------+------------------------------ F( 4, 1255) = 26.63 Model 34.8081566 4 8.70203915 Prob > F = 0.0000 Residual 410.171816 1255.326830132 R-squared = 0.0782 -------------+------------------------------ Adj R-squared = 0.0753 Total 444.979972 1259.353439215 Root MSE =.57169 ------------------------------------------------------------------------------ lwage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- educat -.1293852.9497547-0.14 0.892-1.992667 1.733897 educat2.0233402.1400233 0.17 0.868 -.2513654.2980458 educat3 -.0015552.0087018-0.18 0.858 -.0186269.0155166 educat4.000042.0001941 0.22 0.829 -.0003388.0004229 _cons 1.602221 2.250264 0.71 0.477-2.812474 6.016916 ------------------------------------------------------------------------------ Justin L. Tobias (Purdue) Regression #8 10 / 30

In the simple regression model, education looks clearly significant. In the second model, however, education does not appear related to wages, as none of the coefficients are statistically significant. This, however, is an artifact of multicollinearity. The relevant question to ask is if all of the education variables are jointly equal to zero. In fact, we calculate a χ 2 4 statistic equal to 106.5 for the joint null hypothesis that θ 2 = θ 3 = θ 4 = θ 5 = 0. Justin L. Tobias (Purdue) Regression #8 11 / 30

Multicollinearity We close this discussion with a general derivation that cleanly reveals the problem of multicollinearity. Consider a regression equation that has been transformed into deviation from means: where x 1 is a scalar and z i a vector. We then have: Using the partitioned inverse theorem to select off the (1,1) element of this matrix, we obtain: Justin L. Tobias (Purdue) Regression #8 12 / 30

Multicollinearity Continuing, where R 2 1 is the R-squared value from a regression of x 1 on all the Z s. (To see this last point, recall our earlier derivation of R 2 in the lecture notes). Thus, high values of R1 2 lead to high variances (the multicolinearity problem). Conversely, lots of variation in x 1 mitigates the variance. Justin L. Tobias (Purdue) Regression #8 13 / 30

Interpretations with Common Transformations Models with logarithmic transformations on the dependent and independent variables are ubiquitous in applied work and thus it is useful to pause and explain how to interpret coefficients in such cases. As a benchmark, consider a model without any transformations: y i = β 1 + β 2 x 2i + + β k x ki + ɛ i It is clear in such a situation that β j represents a marginal effect - the expected change in y corresponding to a unit change in x j (holding all else constant): thus producing its interpretation. Justin L. Tobias (Purdue) Regression #8 14 / 30

Interpretations with Common Transformations Now, consider a model in which both the dependent and all the explanatory variables have logarithmic transformations: To see the interpretation here, it is useful to take the differential of both sides of this equation, noting that d[f (x)] = f (x)dx. Thus, or yielding Justin L. Tobias (Purdue) Regression #8 15 / 30

Interpretations with Common Transformations The left hand side of the last equation is a partial elasticity - the percentage change in y corresponding to a percentage change in x j (holding all else constant). Thus, in log-log models, the coefficients represent (partial) elasticities. Example : logcoffeedemanded =.77.253logPrice. This would indicate that the demand for coffee is price-inelastic. Justin L. Tobias (Purdue) Regression #8 16 / 30

Perhaps equally common is the case where the dependent variable has a log transformation, but the independent variables do not: Performing a similar operation, we obtain: which rearranges to: or 100β j represents the percentage change in y corresponding to a unit change in x. Example : LogWage = 2.2 +.12Education. That is, an added year of schooling increases your wages by 12 percent, on average. Justin L. Tobias (Purdue) Regression #8 17 / 30

Interpretations with Common Transformations Also note that other parameters of interest can be obtained via simple manipulations of this formula. For example, in the log-levels model, we can re-arrange things to obtain: Evaluated at y = 10, for example, this would imply that an added year of schooling increases your (hourly) wage by about $1.20 on average. Finally, similar (and obvious) manipulations can be performed to provide the correct interpretation when the explanatory variables are measured in logs while the dependent variable is measured in levels. Justin L. Tobias (Purdue) Regression #8 18 / 30

Dummy Variables and Interactions Dummy variables (or indicator variables) represent a useful way to represent qualitative information in a quantitive way. For example, one can control for variation across race, gender or region of residence through the creation of dummy variables. Often data sets will code such information in a way that is not directly useful to the econometrician. For example, gender may be listed as F or M while region of residence may be coded as, say, 1-4, denoting the East, West, North and South, respectively. Justin L. Tobias (Purdue) Regression #8 19 / 30

Dummy Variables and Interactions When dummy variables are used, one must take care to interpret the parameters correctly. To see this, consider two models: where y i = β 1 D 1i + β 2 D 2i + β 3 D 3i + u i y i = α 1 + α 2 D2i + α 3 D3i + v i Model 1 Model 2 D 1 = I (Ed < 12) D 2 = I (Ed = 12) D 2 = I (Ed 12) D 3 = I (Ed > 12) D 3 = I (Ed > 12) Justin L. Tobias (Purdue) Regression #8 20 / 30

Dummy Variables and Interactions The conditional expectations reveal the interpretation of the coefficients in each model: E(y Dropout, Model1) = β 1 E(y Dropout, Model2) = α 1 E(y HSGrad, Model1) = β 2 E(y HSGrad, Model2) = α 1 + α 2 Justin L. Tobias (Purdue) Regression #8 21 / 30

Dummy Variables and Interactions E(y MorethanHS, Model1) = β 3 E(y MorethanHS, Model2) = α 1 + α 2 + α 3 Thus, in Model 2, the α s are interpreted as the gains (or losses) from moving to the higher education group while the β s are the average wages for the given group. Note that the interpretation of the coefficients changes across models, even though D 3 and D 3, for example are the same variable. Justin L. Tobias (Purdue) Regression #8 22 / 30

Dummies and Interactions: Example 1 The STATA output clearly shows that the α s are estimated as differences between the β s. Wages are significantly larger when moving to the higher education group.. regress wage D1 D2 D3, noconst ------------------------------------------------------------------------------ wage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+--------------------------------------------------------------- D1 774.25 41.29344 18.75 0.000 693.2111 855.2889 D2 862.6718 19.54007 44.15 0.000 824.3241 901.0194 D3 1076.024 18.18003 59.19 0.000 1040.346 1111.703 ------------------------------------------------------------------------------. regress wage tilded2 tilded3 ------------------------------------------------------------------------------ wage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+--------------------------------------------------------------- tilded2 88.42176 45.68329 1.94 0.053-1.232276 178.0758 tilded3 213.3525 26.68947 7.99 0.000 160.9741 265.7309 _cons 774.25 41.29344 18.75 0.000 693.2111 855.2889 ------------------------------------------------------------------------------ Justin L. Tobias (Purdue) Regression #8 23 / 30

Interaction What is an interaction? This is sometimes done to add flexibility to a regression model. However, most of the time, the interaction is added to enable the researcher to test some hypothesis of interest. Consider, for example the regression model: LogWage i = β 1 + β 2 Fem i + β 3 Educ i + β 4 Fem i Educ i + β 5 Exper i + u i. What potentially interesting hypotheses would this enable us to test? Justin L. Tobias (Purdue) Regression #8 24 / 30

What would you conclude based on these results? ------------------------------------------------------------------------------ lwage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- female -.2958839.1787929-1.65 0.099 -.6471275.0553597 educ.0927793.0089777 10.33 0.000.0751423.1104163 femaleed -.0038152.013975-0.27 0.785 -.0312694.0236391 exper.0094302.0014518 6.50 0.000.0065781.0122823 _cons.4614994.1267468 3.64 0.000.2125017.7104971 ------------------------------------------------------------------------------ Justin L. Tobias (Purdue) Regression #8 25 / 30

Nonlinearities Nonlinearities are typically handled in a regression framework by including powers of the explanatory variables and including these as separate regressors. In a sense, this might be thought of as a special interaction. Sometimes these are included to make the regression model more flexible, but if this is the case, researchers often tend to prefer nonparametric methods. However, economic theory (and common sense) often suggests the inclusion of such variables to allow, for example, quadratic profiles of certain covariates. Consider, for example, the model below: LogWage i = β 1 + β 2 Fem i + β 3 Educ i + β 4 Exper i + β 5 Exper 2 i + u i, with output presented on the following page: Justin L. Tobias (Purdue) Regression #8 26 / 30

When are the returns to experience at a maximum? ------------------------------------------------------------------------------ lwage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- female -.3371868.0363214-9.28 0.000 -.4085411 -.2658324 educ.0841361.0069568 12.09 0.000.0704692.0978029 exper.03891.0048235 8.07 0.000.029434.0483859 exper2 -.000686.0001074-6.39 0.000 -.000897 -.0004751 _cons.390483.1022096 3.82 0.000.1896894.5912767 ------------------------------------------------------------------------------ Justin L. Tobias (Purdue) Regression #8 27 / 30

What if you wanted to test if return to experience profiles are different for men and women? To this end, you might want to estimate a model like: LogWage i = β 1 + β 2 Educ i + β 3 Exper i + β 4 Experi 2 + β 5 F i + β 6 F i Exper i + β 7 F i Experi 2 + u i, where F represents the Female Dummy. What would be some hypothesis tests of interest? Justin L. Tobias (Purdue) Regression #8 28 / 30

When are wages highest for men? For Women? Do we see the same experience profiles for both genders? ------------------------------------------------------------------------------ lwage Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- educ.0857402.0068921 12.44 0.000.0722004.0992801 exper.0543082.0065378 8.31 0.000.0414643.067152 exper2 -.000929.0001447-6.42 0.000 -.0012134 -.0006447 female -.035051.0800504-0.44 0.662 -.1923137.1222117 fexper -.0320967.0094877-3.38 0.001 -.0507357 -.0134578 fexper2.0005158.0002093 2.46 0.014.0001046.000927 _cons.2186471.1078223 2.03 0.043.0068253.4304688 ------------------------------------------------------------------------------ Justin L. Tobias (Purdue) Regression #8 29 / 30

Maximum Experience, Males: 29.2, Maximum Experience, Females: 26.9 2.1 2 Log Wage Profile: Men Log Wage Profile: Women 1.9 1.8 Log Wage 1.7 1.6 1.5 1.4 1.3 0 5 10 15 20 25 30 35 40 Experience Justin L. Tobias (Purdue) Regression #8 30 / 30