σ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity

Size: px
Start display at page:

Download "σ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity"

Transcription

1 Comparison of SLR and MLR analysis: What s New? Roadmap Multicollinearity and standard errors F Tests of linear restrictions F stats, adjusted R-squared, RMSE and t stats Playing with Bodyfat: F tests in action Comparison of SLR and MLR Analysis: What s new? Those (S/M)LR Assumptions If these hold: SLR SLR.: Linear Model Y = β + β X + U 0 SLR: Random sampling SLR3: Sampling variation in the independent variable SLR.4: Zero Conditional Mean EU ( X= ) = 0 for any LUE: OLS provides unbiased parameter estimates MLR MLR.: Linear Model Y = β + β X + U 0 j j MLR: Random sampling MLR3: No perfect collinearity MLR.4: Zero Conditional Mean EU (,,, k ) = 0 for any, LUE: OLS provides unbiased parameter estimates PRF (Population Regression Function) EY ( X ) = = β0 + β EY (,, k) β0 β jj = + Add one more assumption SLR.5: Homoskedastic errors (conditional on ) MLR.5: Homoskedastic errors (conditional on the j s) If all 5 conditions hold: s = ( i ) σ Var( B ' ) σ Var( B, w, z... etc) = ( i ) R VIF (Variance Inflation Factor) VIFj = R j

2 So Var( B, w, z... etc) = σ ( ) i VIF ˆ SSR n σ = = MSE ˆ SSR n ( k + ) σ = = MSE MSE unbiased estimator of Gauss-Markov Thm. SLR.-SLR.5 BLUE σ MSE unbiased estimator of MLR.-MLR.5 BLUE σ. Inference SLR MLR Add one last Assumption: Distribution of Y i SLR.6: Ui N(0, σ ) and indept. of X MLR.6: Ui N(0, σ ) and indept. of Xj s Degrees of Freedom n- n-k- Distribution of Estimator: t statistic B β tn where se( B ) MSE se( B ) = ( i ) B j β j se( B ) j se( B ) t where n k MSE = ( i ) i VIF Confidence Intervals (C is confidence level) [ B c se( B )] ±, P[ tn < c) = C Bj ± c se( Bj), P[ t < c) = C n k Null Hypothesis: H0 : β = 0 H : 0 0 β j = t stat (under H0) B tstat = t se( B ) n B j tstat = t se( B ) j n k p value (for given data set and OLS estimates) p P tn tstat ˆ β = [ > ) p = P[ tn k > tstat ˆ β ) j Hypothesis Test (for given ) : Reject H0 if t ˆ > or if p α β c < Reject H0 if t ˆ > c or if p < α β j

3 Roadmap. There are remarkably few substantive differences between SLR and MLR models with respect to estimation and inference. Or put differently: Much of what we learned about SLR models carries over to MLR models.. Some new issues that arise because we now have the possibility of more than one eplanatory variable in the model: a. Small differences in formulas for adjusted R-squared, MSE, estimators, confidence intervals and hypothesis tests (VIFj and n-k-) b. F statistics and F tests - testing linear restrictions on unknown parameters (plural) c. And, umm, that s about it. 3. The MLR.-MLR.6 conditions are basically the same as the SLR conditions, though SLR.3 (variation in the RHS variable) is replaced by MLR.3 (no perfect collinearity amongst the RHS variables). 4. In several cases, we replace the (n-) term in various SLR formulas with (n-k-) for MLR formulas, where k is the number of eplanatory variables in the model. a. The t-distributions used in Hypothesis Testing and Confidence Intervals now have n-k- degrees of freedom. And the F-statistic does as well, sort of. b. The formula for adjusted R-squared now has a n-k- term. c. In the MSE formula, we now divide by n-k- to generate an unbiased estimator. And yes, n- = n-k- for SLR models. Adjusted R-squared will now be of more value as we compare goodness of fit across models with differing numbers of eplanatory variables.

4 Multicollinearity and Standard Errors. Multicollinearity: Recall that multicollinearity captures the etent to which the values of one (RHS) variable can be predicted by a linear function of other RHS variables a. Typically measured as R j (R squared j): The R of the collinearity regression in which one of the RHS variables is regressed on the others RHS variables. So, for eample, R is the R of the following MLR model: X = γ0 + γx + γ3x γnxn + U. Variance Inflation Factors: VIFj =, and so named because given MLR.-5 and R j conditional on the s, σ σ σ Var( B ) = = VIF = VIF. ( ) ( ) ( n ) S i i R i i 3. As we saw earlier in the semester, while the presence of multicollinearity will lead to higher standard errors, the real problem with it is that it can lead to wacky estimated coefficients because of the ceteris paribus condition (in some cases it doesn t make sense to hold everything else fied while changing just one eplanatory variable). So if you have strange estimated coefficients, see if multicollinearity is driving them. a. Note that since VIFj =, R j =. Recall that the second relationship is R j VIFj useful in calculating R using the output from Stata s vif command. j

5 F Tests of Linear Restrictions 4. To do an F test of a set of q linear restrictions on the estimated coefficient. a. Start with the Null hypothesis that it s A-OK to impose some linear restrictions on the OLS model. i. Eamples of linear restriction: a) β = β, b) β+ β = 0, c) β = 0and β = 0, and d) β = and β =. Linear restrictions are linear functions of the unknown parameters. ii. We ll be interested in counting the number of restrictions. To do that, just count the equals signs. So in the eamples, a) and b) have one restriction, and c) and d) have two. b. Estimate the model with and without the restrictions. c. Since we ve imposed a restriction (or restrictions) on the estimated coefficients, the SSRs will almost always increase: SSRR SSRUR. (The subscripts refer to the Restricted and Unrestricted models.) d. But by how much? a lot? or maybe not so much? e. If SSRs increase by a lot (whatever that is) then the restrictions severely impacted the performance of the model, and so we reject the Null Hypothesis (which was that the restrictions were A-OK). f. But if not so much, then maybe those restrictions weren t so bad after all, and we might fail to reject. Which is to say that it really was A-OK to impose those restrictions. 5. F statistic: We use the F statistic (and the F distribution) to do the F test. The F statistic is defined by: ( ) / SSRR SSRUR q F, where q is the number of restrictions (eg the # of = s), and SSRUR / ( n k ) n k is the number of degrees of freedom in the unrestricted (UR) model. By construction F Elasticity - Another interpretation of the F statistic: We can rewrite the equation for F as ( ) / % SSRR SSRUR SSRUR SSR F = so the F statistics tells you the %change in q / ( n k ) % dofs SSRs for a given %change in degrees of freedom (you might call this bang per buck).

6 7. Under MLR.-MLR.6, the following F statistic will have a Fqn (, k ) distribution. Here are some F distributions for different parameter values: F(,0) Distribution F(3,0) Distribution F(0,0) Distribution F(5,0) Distribution 8. We do the F test following the classical hypothesis testing protocol: a. The Null Hypothesis is that the restrictions are A-OK. b. Focus on Type I errors and the probability of a False Rejection. c. Pick a small significance level, α, say α =.05, which will be the maimum acceptable probability of a False Rejection. d. For the given Fstat, compute the p-value as the probability in the tail to the right of the Fstat: p = P( F( q, n k ) > Fstat). e. The p value is the probability that you d observe an F statistic at least as large as Fstat when the Null Hypothesis was in fact true.

7 f. Reject the Null Hypothesis if the p < α and fail to reject otherwise. i. So reject the Null Hypothesis if the F statistic is large and the associated probability level is small (less than the significance level α ). In other words: the restrictions are rejected because when they were imposed, the world got a whole lot worse. ii. If the Fstat (elasticity) is small, so that relatively speaking, SSRs did not increase by much, then the associated probability level is high, and the restrictions are not rejected and are deemed to be OK. 9. And so we reject the Null Hypothesis if the probability value associated with this test statistic is below the desired significance level or equivalently, if the F statistic is too large. 0. Eample: Testing hgt parameter is zero in the following SLR model. reg Brozek hgt Source SS df MS Number of obs = F(, 50) =.00 Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE = Brozek Coef. Std. Err. t P> t [95% Conf. Interval] hgt _cons test hgt=0 ( ) hgt = 0 F(, 50) =.00 Prob > F = Notice that q =, n-k- = 50, the Fstat is.00 and the p value (Prob > F) is We cannot reject at the 5% level the Null Hypothesis that the true parameter value for hgt is 0. Notice that the p value for the F test is the same as the p-value for the t stat n the regression results and that the Fstat is the square of the reported tstat. These are not coincidences (see below).

8 . F stats defined using R-squared: You can also derive the F statistic using R-squared. SSRR SSRUR a. Define RR = and RUR =, and so SST SST SST ( R R R UR ) / q ( RUR RR ) / q F =. SST R UR / ( n k ) R UR / ( n k ) b. Relation to t-stats: If you apply the F test to the case of testing just one parameter value (say testing β = 0 ), the F test will be the same as the t test, as above. So there is no inconsistency here. (See the eample above.) c. Fstat will be the same as tstat, and the p-values for the two statistics will be the same since prob( F(, n k ) < ) = prob( < t < ), for > 0.. Babies and bathwater: Be careful about throwing out the baby with the bath water you don t want to eclude a significant eplanatory variable just because it happens to be associated with a set of variables that are jointly insignificant. 3. Eample: Here s an eample using a sample from a sovereign debt dataset. The F test does not reject at the 0% level the Null Hypothesis that the inflation and deficit parameters are zero even though deficit_gdp is almost statistically significant at the 5% level. But inflation is so statistically insignificant, it pulls down deficit_gdp in the process. n k. reg NSRate corrupt gdp inflation deficit_gdp debt_gdp eurozone if _n < 30 Source SS df MS Number of obs = F(6, ) =.99 Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE =.0674 NSRate Coef. Std. Err. t P> t [95% Conf. Interval] corrupt gdp inflation deficit_gdp debt_gdp eurozone _cons

9 . test inflation deficit_gdp ( ) inflation = 0 ( ) deficit_gdp = 0 F(, ) =. Prob > F = The reported F stat in OLS output: Standard regression packages report the F statistic associated with test that all of the (non-intercept) parameter values are zero. This is sometimes called the overall significance of the regression. a. If you cannot reject that hypothesis, you have a really crummy model. b. Using the R formula and the fact that the restricted model regressing Y on a constant only will have R R = 0. And since RUR = R (the reported R-squared for the regression), R / k we have F = the reported F statistic. R / ( n k ) c. Note that this is confirmed in the previous regression results:. di (0.7799/6)/(( )/).9948 F stats, adjusted R-squared, RMSE and tstats 5. Recall that adjusted R is an attempt to adjust the coefficient of determination for the fact that when you add RHS variables to a model, R cannot decline since SSR cannot increase. Small decreases in SSRs will not generate a higher adj R ; larger decreases will and what is small or large will also depend on how many changes were made. 6. Adj R is defined: SSR ( n ) R = SST ( n k ) SSR / ( n k ) = SST / ( n ) a. Thus for given S YY adding or subtracting RHS variables from a MLR model, eactly opposite directions. ( n ) b. And since > 0, R ( n k ) MSE =. S, adjusted R increases if and only if MSE decreases. So if you are YY R and MSE will move in R, with the difference inversely related to k. 7. Consider two models: UR unrestricted and R- restricted. You can think of the process of adding RHS variables to a model as being equivalent to moving from a restricted model (in which the coefficients for all of the added variables are zero) to an unrestricted model (in which they are no longer constrained to be zero).

10 adjr from adding RHS var(s) to the model R Model: fewer RHS vars UR Model: more RHS vars F statistic from dropping RHS var(s) (restricting coeffs to be zero) 8. It turns out that the change in adj R in dropping the restrictions (adding the variables) is closely related to the F-statistic associated with imposing the restrictions. The change in adjusted R in going from one model to the other is defined by: SSRR SSRUR RUR RR = ( MSER MSEUR ) =. SYY SYY ( n k ) + q n k q R ( n k ) + q UR After much algebra, this is: R R = [ F ] UR R a. The sign of this epression will depend on whether the F statistic is greater or less than : adjusted R increases if and only if the F statistic associated with the restrictions eceeds. b. Recall that for a single restriction, the F statistic is the square of the t stat, and so adjusted R increases if and only if the magnitude of the t stat of the added RHS variable eceeds. 9. Adding and dropping RHS variables: This last fact is useful when considering dropping RHS variables from a model. a. One RHS variable: If a RHS variable has a t stat > then dropping that variable from the model will cause adjusted R-squared to decline. And if t stat <, then adjusted R- squared will increase when that variable is dropped from the model. 3 Or put differently: Adding RHS variables with t stat > will lead to higher adjusted R-squared. b. Multiple RHS variables: : If the Fstat associated with dropping multiple RHS variables is > then dropping those variables from the model will cause adjusted R-squared to decline. And if Fstat <, then adjusted R-squared will increase when those variables are. 3 And yes, if t stat = 0, then there will be no change to adjusted R-squared when the variable is dropped from the model.

11 dropped from the model. Or put differently: Adding RHS variables with Fstat > will lead to higher adjusted R-squared. c. Relation to statistical significance: It could be that additional RHS variables that lead to a higher adjusted R-squared are not statistically significant adding a RHS variable with t stat > will be associated with higher adjusted R-squared; but statistical significance will generally want t stat > or so. So if the added variable is statistically significant, then adjusted R-squared increased when that variable was added to the model but it could be that the new variable is not statistically significant, even though adjusted R-squared increased when it was added to the model.

12 An Eample using the bodyfat dataset. bcuse bodyfat () () (3) (4) Brozek Brozek Brozek Brozek wgt -0.5*** -0.9*** -0.54*** -0.3*** (-5.) (-3.68) (-4.9) (-3.56) abd 0.937*** 0.9*** 0.940*** 0.94*** (7.9) (5.37) (6.7) (5.0) hip (-.) (-.4) (-.) (-.4) thigh 0.77* 0.55* 0.77* 0.55* (.43) (.) (.4) (.0) hgt (-.3) (-.) ankle (0.4) (0.3) _cons -4.80*** -3.33** -4.73*** -33.4** (-6.45) (-3.05) (-5.65) (-.93) N R-sq adj. R-sq rmse t statistics in parentheses * p<0.05, ** p<0.0, *** p<0.00 Note: Add hgt -- () to (): hgt tstat >, R and adj R increase, RMSE decreases Add ankle -- () to (3): ankle tstat <, R increases, adj R decreases, and RMSE increases Add hgt -- (3) to (4): hgt tstat >, R and adj R increase, RMSE decreases Add ankle -- () to (4): ankle tstat <, R increases, adj R decreases, and RMSE increases Take Aways: adj R and RMSE (MSE) move in opposite directions, which depend on whether the t-stat of the added variable is smaller or larger than in magnitude

13 Playing with Bodyfat: F tests in action wgt _ kg About that BMI metric: BMI = hgt _ m If the BMI is a useful metric, then it should do a good job eplaining body fat levels: β β wgt _ kg Brozek = αbmi = α or hgt _ m ln( Brozek) = ln( α) + βln( BMI ) = ln( α) + βln( wgt _ kg) βln( hgt _ m) So if we regress Brozek on lnhgt and lnwgt, we should find the lnhgt coeff will be twice the magnitude of the lnwgt coeff and opposite in sign. So let s try that.. bcuse bodyfat. gen lnbrozek=ln(brozek) ( missing value generated). gen lnhgt=ln(hgt_m). gen lnwgt=ln(wgt_kg). reg lnbrozek lnhgt lnwgt Source SS df MS Number of obs = F(, 48) = Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE =.3779 lnbrozek Coef. Std. Err. t P> t [95% Conf. Interval] lnhgt lnwgt _cons Oops the lnhgt coefficient is actually smaller in magnitude than the lnwgt coeff so much for the BMI or maybe not!

14 . bcuse bodyfat Looking for data issues I Compare reported Brozek with calculation (Brozek). list Case Brozek Brozek if abs( Brozek- Brozek)> Case Brozek Brozek Looking for data issues II compare reported heights and weights. scatter hgt wgt.. list if hgt < Case Brozek Siri density age wgt hgt BMI fatfre~t neck chest abd hip thigh knee ankle biceps forearm wrist wgt_kg hgt_m BMI Brozek Siri Yes, that s right. 9.5 inches tall and weighs 05 pounds What happens when we drop those curious observations?. drop if Case == 33 Case == 48 Case == 76 Case == 96 Case == 8 Case == 4 (6 observations deleted). reg lnbrozek lnhgt lnwgt Source SS df MS Number of obs = F(, 43) = Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE =.3576 lnbrozek Coef. Std. Err. t P> t [95% Conf. Interval] lnhgt lnwgt _cons Bingo! - ratio! we can use the F test to test for the BMI restriction.

15 Assume that the restriction in true (the Null hypothesis is that the magnitude ratio is -).. test lnhgt = - *lnwgt ( ) lnhgt + *lnwgt = 0 F(, 43) = 0.00 Prob > F = We fail to reject the restriction, since the Prob > F is so large we typically reject when that probability is less than.,.05,.0 etc. This probability is similar to the p-value it s the probability that we will falsely reject the Null hypothesis. So let s impose the restriction and see what we get:. gen lnbmi=ln(bmi). reg lnbrozek lnbmi Source SS df MS Number of obs = F(, 44) = 00.9 Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE = lnbrozek Coef. Std. Err. t P> t [95% Conf. Interval] lnbmi _cons Reported R-squared didn t change, but SSRs increased (barely) from to (Recall that SSRs must increase (or more precisely, cannot decrease) if you impose a restriction.) The F-test basically looks at how much the SSRs increase when you impose a restriction (or restrictions): If SSRs increase by a small amount, then little eplanatory power is lost when the restriction is imposed, and the F test fails to reject the Null Hypothesis that the restriction is OK. If the SSRs increase by a lot, then the restriction is leading to a much poorer model (as measured by R squared) and the restriction is rejected. More F tests later. Stay tuned!

16 and the Chow Test 0. The Chow Test is a test of whether you can combine two datasets (for eample, estimating one model for both males and females rather than estimating separate models for the two populations). a. The easy way to do the Chow test is to define a dummy variable (say, for males) and to then interact that dummy with all of the other RHS variables in the model. b. Then do an F test to see if you can reject the Null hypothesis that all of the true parameter values for these newly interacted variables are zero. c. If you cannot reject that joint hypothesis, then it s OK to combine the datasets (which in the eample, would essentially amount to requiring identical coefficients for males and females). d. If you reject the Null hypothesis, then you should run separate regression for the two populations.. We combine datasets all the time, often without thinking about it. For instance, in the Pythagorean Theorem regressions, without giving it any thought, we included all MLB games played 960 0, for both National and American League teams. By estimating a single model on the full dataset, we essentially imposed a restriction that the coefficients for the NL teams are the same as the coefficients for the AL teams. With more data you generally have more precise coefficient estimates, but you don t want to combine NL and AL data if the two leagues follow fundamentally different models.. Here are the combined and separate results. The eyeball test says that combining the datasets is fine. We ll do the formal Chow Test when we cover dummy variables (spoiler alert: The Chow test will not reject the combining of the AL and NL datasets) (MLB) (AL) (NL) G_W G_W G_W RA_RS 0.935*** 0.945*** 0.94*** (95.54) (76.5) (60.74) _cons.069***.058***.080*** (00.9) (78.46) (64.98) N t statistics in parentheses * p<0.05, ** p<0.0, *** p<0.00

F Tests and F statistics

F Tests and F statistics F Tests and F statistics Testing Linear estrictions F Stats and F Tests F Distributions F stats (w/ ) F Stats and tstat s eported F Stat's in OLS Output Example I: Bodyfat Babies and Bathwater F Stats,

More information

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Analytics What s New? Not Much!

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Analytics What s New? Not Much! Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Analytics What s New? Not Much! OLS: Comparison of SLR and MLR Analysis Interpreting Coefficients I (SRF): Marginal effects ceteris paribus

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Assessment I What s New? & Goodness-of-Fit

Ordinary Least Squares (OLS): Multiple Linear Regression (MLR) Assessment I What s New? & Goodness-of-Fit Ordinary Least Squares (OLS): Multiple Linear egression (ML) Assessment I What s New? & Goodness-of-Fit Introduction OLS: A Quick Comparison of SL and ML Assessment Not much that's new! ML Goodness-of-Fit:

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

Statistical Inference with Regression Analysis

Statistical Inference with Regression Analysis Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing

More information

Multiple Regression: Inference

Multiple Regression: Inference Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled

More information

Lecture 5: Hypothesis testing with the classical linear model

Lecture 5: Hypothesis testing with the classical linear model Lecture 5: Hypothesis testing with the classical linear model Assumption MLR6: Normality MLR6 is not one of the Gauss-Markov assumptions. It s not necessary to assume the error is normally distributed

More information

1 Multiple Regression

1 Multiple Regression 1 Multiple Regression In this section, we extend the linear model to the case of several quantitative explanatory variables. There are many issues involved in this problem and this section serves only

More information

1 Independent Practice: Hypothesis tests for one parameter:

1 Independent Practice: Hypothesis tests for one parameter: 1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)

More information

Lab 10 - Binary Variables

Lab 10 - Binary Variables Lab 10 - Binary Variables Spring 2017 Contents 1 Introduction 1 2 SLR on a Dummy 2 3 MLR with binary independent variables 3 3.1 MLR with a Dummy: different intercepts, same slope................. 4 3.2

More information

Economics 326 Methods of Empirical Research in Economics. Lecture 14: Hypothesis testing in the multiple regression model, Part 2

Economics 326 Methods of Empirical Research in Economics. Lecture 14: Hypothesis testing in the multiple regression model, Part 2 Economics 326 Methods of Empirical Research in Economics Lecture 14: Hypothesis testing in the multiple regression model, Part 2 Vadim Marmer University of British Columbia May 5, 2010 Multiple restrictions

More information

THE MULTIVARIATE LINEAR REGRESSION MODEL

THE MULTIVARIATE LINEAR REGRESSION MODEL THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus

More information

Lecture 7: OLS with qualitative information

Lecture 7: OLS with qualitative information Lecture 7: OLS with qualitative information Dummy variables Dummy variable: an indicator that says whether a particular observation is in a category or not Like a light switch: on or off Most useful values:

More information

Econometrics Midterm Examination Answers

Econometrics Midterm Examination Answers Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i

More information

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47 ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with

More information

Problem Set 1 ANSWERS

Problem Set 1 ANSWERS Economics 20 Prof. Patricia M. Anderson Problem Set 1 ANSWERS Part I. Multiple Choice Problems 1. If X and Z are two random variables, then E[X-Z] is d. E[X] E[Z] This is just a simple application of one

More information

Essential of Simple regression

Essential of Simple regression Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship

More information

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11 Econ 495 - Econometric Review 1 Contents 1 Linear Regression Analysis 4 1.1 The Mincer Wage Equation................. 4 1.2 Data............................. 6 1.3 Econometric Model.....................

More information

Inference in Regression Analysis

Inference in Regression Analysis ECNS 561 Inference Inference in Regression Analysis Up to this point 1.) OLS is unbiased 2.) OLS is BLUE (best linear unbiased estimator i.e., the variance is smallest among linear unbiased estimators)

More information

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests ECON4150 - Introductory Econometrics Lecture 7: OLS with Multiple Regressors Hypotheses tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 7 Lecture outline 2 Hypothesis test for single

More information

Lab 11 - Heteroskedasticity

Lab 11 - Heteroskedasticity Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

1 Warm-Up: 2 Adjusted R 2. Introductory Applied Econometrics EEP/IAS 118 Spring Sylvan Herskowitz Section #

1 Warm-Up: 2 Adjusted R 2. Introductory Applied Econometrics EEP/IAS 118 Spring Sylvan Herskowitz Section # Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Sylvan Herskowitz Section #10 4-1-15 1 Warm-Up: Remember that exam you took before break? We had a question that said this A researcher wants to

More information

2.1. Consider the following production function, known in the literature as the transcendental production function (TPF).

2.1. Consider the following production function, known in the literature as the transcendental production function (TPF). CHAPTER Functional Forms of Regression Models.1. Consider the following production function, known in the literature as the transcendental production function (TPF). Q i B 1 L B i K i B 3 e B L B K 4 i

More information

Nonlinear Regression Functions

Nonlinear Regression Functions Nonlinear Regression Functions (SW Chapter 8) Outline 1. Nonlinear regression functions general comments 2. Nonlinear functions of one variable 3. Nonlinear functions of two variables: interactions 4.

More information

Lab 07 Introduction to Econometrics

Lab 07 Introduction to Econometrics Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand

More information

Empirical Application of Simple Regression (Chapter 2)

Empirical Application of Simple Regression (Chapter 2) Empirical Application of Simple Regression (Chapter 2) 1. The data file is House Data, which can be downloaded from my webpage. 2. Use stata menu File Import Excel Spreadsheet to read the data. Don t forget

More information

At this point, if you ve done everything correctly, you should have data that looks something like:

At this point, if you ve done everything correctly, you should have data that looks something like: This homework is due on July 19 th. Economics 375: Introduction to Econometrics Homework #4 1. One tool to aid in understanding econometrics is the Monte Carlo experiment. A Monte Carlo experiment allows

More information

Lecture 3: Multivariate Regression

Lecture 3: Multivariate Regression Lecture 3: Multivariate Regression Rates, cont. Two weeks ago, we modeled state homicide rates as being dependent on one variable: poverty. In reality, we know that state homicide rates depend on numerous

More information

(a) Briefly discuss the advantage of using panel data in this situation rather than pure crosssections

(a) Briefly discuss the advantage of using panel data in this situation rather than pure crosssections Answer Key Fixed Effect and First Difference Models 1. See discussion in class.. David Neumark and William Wascher published a study in 199 of the effect of minimum wages on teenage employment using a

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

General Linear Model (Chapter 4)

General Linear Model (Chapter 4) General Linear Model (Chapter 4) Outcome variable is considered continuous Simple linear regression Scatterplots OLS is BLUE under basic assumptions MSE estimates residual variance testing regression coefficients

More information

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one

More information

Regression #8: Loose Ends

Regression #8: Loose Ends Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch

More information

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C = Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution

More information

ECO220Y Simple Regression: Testing the Slope

ECO220Y Simple Regression: Testing the Slope ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x

More information

1. The shoe size of five randomly selected men in the class is 7, 7.5, 6, 6.5 the shoe size of 4 randomly selected women is 6, 5.

1. The shoe size of five randomly selected men in the class is 7, 7.5, 6, 6.5 the shoe size of 4 randomly selected women is 6, 5. Economics 3 Introduction to Econometrics Winter 2004 Professor Dobkin Name Final Exam (Sample) You must answer all the questions. The exam is closed book and closed notes you may use calculators. You must

More information

Linear Regression with Multiple Regressors

Linear Regression with Multiple Regressors Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution

More information

1 The basics of panel data

1 The basics of panel data Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Related materials: Steven Buck Notes to accompany fixed effects material 4-16-14 ˆ Wooldridge 5e, Ch. 1.3: The Structure of Economic Data ˆ Wooldridge

More information

Inference in Regression Model

Inference in Regression Model Inference in Regression Model Christopher Taber Department of Economics University of Wisconsin-Madison March 25, 2009 Outline 1 Final Step of Classical Linear Regression Model 2 Confidence Intervals 3

More information

Introduction to Econometrics. Multiple Regression (2016/2017)

Introduction to Econometrics. Multiple Regression (2016/2017) Introduction to Econometrics STAT-S-301 Multiple Regression (016/017) Lecturer: Yves Dominicy Teaching Assistant: Elise Petit 1 OLS estimate of the TS/STR relation: OLS estimate of the Test Score/STR relation:

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple

More information

Multicollinearity Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 13, 2015

Multicollinearity Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 13, 2015 Multicollinearity Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 13, 2015 Stata Example (See appendices for full example).. use http://www.nd.edu/~rwilliam/stats2/statafiles/multicoll.dta,

More information

Lab 6 - Simple Regression

Lab 6 - Simple Regression Lab 6 - Simple Regression Spring 2017 Contents 1 Thinking About Regression 2 2 Regression Output 3 3 Fitted Values 5 4 Residuals 6 5 Functional Forms 8 Updated from Stata tutorials provided by Prof. Cichello

More information

Homework 1 Solutions

Homework 1 Solutions Homework 1 Solutions January 18, 2012 Contents 1 Normal Probability Calculations 2 2 Stereo System (SLR) 2 3 Match Histograms 3 4 Match Scatter Plots 4 5 Housing (SLR) 4 6 Shock Absorber (SLR) 5 7 Participation

More information

Problem Set 10: Panel Data

Problem Set 10: Panel Data Problem Set 10: Panel Data 1. Read in the data set, e11panel1.dta from the course website. This contains data on a sample or 1252 men and women who were asked about their hourly wage in two years, 2005

More information

STAT 350: Summer Semester Midterm 1: Solutions

STAT 350: Summer Semester Midterm 1: Solutions Name: Student Number: STAT 350: Summer Semester 2008 Midterm 1: Solutions 9 June 2008 Instructor: Richard Lockhart Instructions: This is an open book test. You may use notes, text, other books and a calculator.

More information

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time

Autocorrelation. Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time Autocorrelation Given the model Y t = b 0 + b 1 X t + u t Think of autocorrelation as signifying a systematic relationship between the residuals measured at different points in time This could be caused

More information

Problem 4.1. Problem 4.3

Problem 4.1. Problem 4.3 BOSTON COLLEGE Department of Economics EC 228 01 Econometric Methods Fall 2008, Prof. Baum, Ms. Phillips (tutor), Mr. Dmitriev (grader) Problem Set 3 Due at classtime, Thursday 14 Oct 2008 Problem 4.1

More information

Lecture 3: Inference in SLR

Lecture 3: Inference in SLR Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals

More information

Question 1 [17 points]: (ch 11)

Question 1 [17 points]: (ch 11) Question 1 [17 points]: (ch 11) A study analyzed the probability that Major League Baseball (MLB) players "survive" for another season, or, in other words, play one more season. They studied a model of

More information

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators 1 2 Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE Hüseyin Taştan 1 1 Yıldız Technical University Department of Economics These presentation notes are based on Introductory

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1 Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when

More information

Handout 12. Endogeneity & Simultaneous Equation Models

Handout 12. Endogeneity & Simultaneous Equation Models Handout 12. Endogeneity & Simultaneous Equation Models In which you learn about another potential source of endogeneity caused by the simultaneous determination of economic variables, and learn how to

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

STATISTICS 110/201 PRACTICE FINAL EXAM

STATISTICS 110/201 PRACTICE FINAL EXAM STATISTICS 110/201 PRACTICE FINAL EXAM Questions 1 to 5: There is a downloadable Stata package that produces sequential sums of squares for regression. In other words, the SS is built up as each variable

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Chapter 6: Linear Regression With Multiple Regressors

Chapter 6: Linear Regression With Multiple Regressors Chapter 6: Linear Regression With Multiple Regressors 1-1 Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution

More information

Chapter 27 Summary Inferences for Regression

Chapter 27 Summary Inferences for Regression Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 13 Nonlinearities Saul Lach October 2018 Saul Lach () Applied Statistics and Econometrics October 2018 1 / 91 Outline of Lecture 13 1 Nonlinear regression functions

More information

Week 3: Simple Linear Regression

Week 3: Simple Linear Regression Week 3: Simple Linear Regression Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ALL RIGHTS RESERVED 1 Outline

More information

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58

Inference. ME104: Linear Regression Analysis Kenneth Benoit. August 15, August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58 Inference ME104: Linear Regression Analysis Kenneth Benoit August 15, 2012 August 15, 2012 Lecture 3 Multiple linear regression 1 1 / 58 Stata output resvisited. reg votes1st spend_total incumb minister

More information

Computer Exercise 3 Answers Hypothesis Testing

Computer Exercise 3 Answers Hypothesis Testing Computer Exercise 3 Answers Hypothesis Testing. reg lnhpay xper yearsed tenure ---------+------------------------------ F( 3, 6221) = 512.58 Model 457.732594 3 152.577531 Residual 1851.79026 6221.297667619

More information

Hypothesis Tests and Confidence Intervals in Multiple Regression

Hypothesis Tests and Confidence Intervals in Multiple Regression Hypothesis Tests and Confidence Intervals in Multiple Regression (SW Chapter 7) Outline 1. Hypothesis tests and confidence intervals for one coefficient. Joint hypothesis tests on multiple coefficients

More information

Introduction to Econometrics. Multiple Regression

Introduction to Econometrics. Multiple Regression Introduction to Econometrics The statistical analysis of economic (and related) data STATS301 Multiple Regression Titulaire: Christopher Bruffaerts Assistant: Lorenzo Ricci 1 OLS estimate of the TS/STR

More information

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X. Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.

More information

Lecture 11: Simple Linear Regression

Lecture 11: Simple Linear Regression Lecture 11: Simple Linear Regression Readings: Sections 3.1-3.3, 11.1-11.3 Apr 17, 2009 In linear regression, we examine the association between two quantitative variables. Number of beers that you drink

More information

University of California at Berkeley Fall Introductory Applied Econometrics Final examination. Scores add up to 125 points

University of California at Berkeley Fall Introductory Applied Econometrics Final examination. Scores add up to 125 points EEP 118 / IAS 118 Elisabeth Sadoulet and Kelly Jones University of California at Berkeley Fall 2008 Introductory Applied Econometrics Final examination Scores add up to 125 points Your name: SID: 1 1.

More information

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors: Wooldridge, Introductory Econometrics, d ed. Chapter 3: Multiple regression analysis: Estimation In multiple regression analysis, we extend the simple (two-variable) regression model to consider the possibility

More information

ECON3150/4150 Spring 2016

ECON3150/4150 Spring 2016 ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

Introductory Econometrics. Lecture 13: Hypothesis testing in the multiple regression model, Part 1

Introductory Econometrics. Lecture 13: Hypothesis testing in the multiple regression model, Part 1 Introductory Econometrics Lecture 13: Hypothesis testing in the multiple regression model, Part 1 Jun Ma School of Economics Renmin University of China October 19, 2016 The model I We consider the classical

More information

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem - First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III

More information

Interpreting coefficients for transformed variables

Interpreting coefficients for transformed variables Interpreting coefficients for transformed variables! Recall that when both independent and dependent variables are untransformed, an estimated coefficient represents the change in the dependent variable

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 7 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 68 Outline of Lecture 7 1 Empirical example: Italian labor force

More information

Econometrics Homework 1

Econometrics Homework 1 Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z

More information

Sociology 593 Exam 1 Answer Key February 17, 1995

Sociology 593 Exam 1 Answer Key February 17, 1995 Sociology 593 Exam 1 Answer Key February 17, 1995 I. True-False. (5 points) Indicate whether the following statements are true or false. If false, briefly explain why. 1. A researcher regressed Y on. When

More information

Course Econometrics I

Course Econometrics I Course Econometrics I 3. Multiple Regression Analysis: Binary Variables Martin Halla Johannes Kepler University of Linz Department of Economics Last update: April 29, 2014 Martin Halla CS Econometrics

More information

Chapter 3 Multiple Regression Complete Example

Chapter 3 Multiple Regression Complete Example Department of Quantitative Methods & Information Systems ECON 504 Chapter 3 Multiple Regression Complete Example Spring 2013 Dr. Mohammad Zainal Review Goals After completing this lecture, you should be

More information

ECO375 Tutorial 4 Introduction to Statistical Inference

ECO375 Tutorial 4 Introduction to Statistical Inference ECO375 Tutorial 4 Introduction to Statistical Inference Matt Tudball University of Toronto Mississauga October 19, 2017 Matt Tudball (University of Toronto) ECO375H5 October 19, 2017 1 / 26 Statistical

More information

Specification Error: Omitted and Extraneous Variables

Specification Error: Omitted and Extraneous Variables Specification Error: Omitted and Extraneous Variables Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised February 5, 05 Omitted variable bias. Suppose that the correct

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Using the same data as before, here is part of the output we get in Stata when we do a logistic regression of Grade on Gpa, Tuce and Psi.

Using the same data as before, here is part of the output we get in Stata when we do a logistic regression of Grade on Gpa, Tuce and Psi. Logistic Regression, Part III: Hypothesis Testing, Comparisons to OLS Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised January 14, 2018 This handout steals heavily

More information

Lecture 8: Heteroskedasticity. Causes Consequences Detection Fixes

Lecture 8: Heteroskedasticity. Causes Consequences Detection Fixes Lecture 8: Heteroskedasticity Causes Consequences Detection Fixes Assumption MLR5: Homoskedasticity 2 var( u x, x,..., x ) 1 2 In the multivariate case, this means that the variance of the error term does

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

Basic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler

Basic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler Basic econometrics Tutorial 3 Dipl.Kfm. Introduction Some of you were asking about material to revise/prepare econometrics fundamentals. First of all, be aware that I will not be too technical, only as

More information

Group Comparisons: Differences in Composition Versus Differences in Models and Effects

Group Comparisons: Differences in Composition Versus Differences in Models and Effects Group Comparisons: Differences in Composition Versus Differences in Models and Effects Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised February 15, 2015 Overview.

More information

Econometrics Review questions for exam

Econometrics Review questions for exam Econometrics Review questions for exam Nathaniel Higgins nhiggins@jhu.edu, 1. Suppose you have a model: y = β 0 x 1 + u You propose the model above and then estimate the model using OLS to obtain: ŷ =

More information

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018 Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate

More information

sociology sociology Scatterplots Quantitative Research Methods: Introduction to correlation and regression Age vs Income

sociology sociology Scatterplots Quantitative Research Methods: Introduction to correlation and regression Age vs Income Scatterplots Quantitative Research Methods: Introduction to correlation and regression Scatterplots can be considered as interval/ratio analogue of cross-tabs: arbitrarily many values mapped out in -dimensions

More information