Multiple Regression Analysis

Size: px
Start display at page:

Download "Multiple Regression Analysis"

Transcription

1 Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x β k x k + u 2. Inference 0

2 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators 2. That OLS is BLUE, given the Gauss-Markov assumptions! This still leaves open the shape of the distribution of the OLS estimators! To do classical hypothesis testing, we need to make an assumption about this distribution Assumption 6 (Normality): u is independent of x 1, x 2,, x k and u is normally distributed with zero mean and variance σ 2 : u ~ Normal(0, σ 2 ) 1

3 CLM Assumptions (cont)! This is a very strong assumption, it implies: E(u x)=e(u)=0, and Var(u x)=var(u)=σ 2! Assumptions 1-6 are called the Classical Linear Model Assumptions, CLM Assumptions! Under CLM, OLS is not only BLUE, but is the minimum variance unbiased estimator - no longer limited to the class of linear estimators 2

4 CLM Assumptions (cont)! We can summarize the population assumptions of CLM as follows! y x ~ Normal(β 0 + β 1 x β k x k, σ 2 )! While for now we just assume normality, clear that sometimes not the case! Example: Suppose we re interested in modeling house prices (our y variable) as a function of x variables: House prices are never below zero (a normally distributed variable takes on negative values); same problem with wages Could take log of house prices to convert it to something close to normal; analysis of wages as y variable often involves taking logs Or, if our sample is large enough, non-normality ceases to be a problem, thanks to the Central Limit Theorem. 3

5 The homoskedastic normal distribution with a single explanatory variable f(y x) y.. E(y x) = β 0 + β 1 x Normal distribution x 1 x 2 4

6 Normal Sampling Distributions! Under the CLM assumptions, conditional on the sample values of the independent variables, (x), ˆ! j ~ N "! j,var( ˆ! j ) $ # %! Why? Because each estimated β is a linear combination of the errors, u, in the sample. Recall: ˆ " j = " j + n $ i=1 n $ (x ij # x j )u i (x ij # x j ) 2, and u i ~ N(0,% 2 ) i=1! Any linear combination of iid normal random variables is normal (recall Appendix C) 5

7 Normal Sampling Distributions thus, ( ˆ! j "! ) j sd ( ˆ! ) j ~ Normal( 0,1)! To standardize (as before), we just subtract the mean and divide by the standard deviation! The fact that the estimated β s are distributed normally has further implications we will use: 1) Any linear combination of the estimated β s will be normally distributed. 2) Any subset of the estimated β s will be jointly normally distributed 6

8 The t Test! Testing hypotheses about a single population parameter! Remember, the β j are unknown! However, we can hypothesize about the value of β j and use statistical inference to test our hypothesis! To perform a hypothesis test we first note that under CLM: ( ˆ! j "! j ) se( ˆ! j ) ~ t n" k "1 7

9 The t Test (cont)! Knowing the sampling distribution for the standardized estimator allows us to carry out hypothesis tests! Start with a null hypothesis! For example, H 0 :! j = 0! The null here is making the claim that variable x j has no effect on y. 8

10 The t test (cont)! To perform our test (given this particular stating of the null hypothesis) we first form the t statistic for our estimate of β j for ˆ! j : t ˆ! j " ˆ! j # 0 se( ˆ! ) = j ˆ! j se( ˆ! ) j! Notice that we subtract the value of β j under the null (in this case, 0), from our estimate of β j. Intuitively, if our estimate is close to the null, our t- statistic will be close to zero, and we will be unable to reject the null that β j =0. 9

11 Things to notice about t-stat! The denominator is always positive. So for a zero null, the t-statistic will always take on the sign of your estimate of β j. For positive estimated β j it will be positive, for negative estimated β j it will be negative.! We can interpret the t-statistic as giving us an estimate of how many standard deviations from the null (0, in this case) our estimate of β j lies. Say it equals 3. This suggests that it lies three standard deviations to the right of 0. If it equals -3, this suggests it lies 3 sd s to the left of 0. 10

12 Intuition for t Test! Because it is a continuous random variable, the estimated β j will never precisely equal 0 even if the underlying population parameter β j equals 0! We want to know how far estimated β j is from 0 Essentially, if our estimated β j lies near the null, we think there s a good chance that estimate would be generated if the null is true (so, we will fail to reject the null); if our estimated β j lies far away from the null, we think it s unlikely that estimate would be generated under the null: thus we reject the null.! The question is, How far is far enough from the null to warrant rejection?! We need a rejection rule! 11

13 t-test: One-Sided Alternatives! Besides our null, H 0, we need an alternative hypothesis, H 1, and a significance level! H 1 may be one-sided, or two-sided! Examples of different one-sided alternatives include H 1 :! j < 0 or H 1 :! j > 0! A 2-sided alternative is H 1 :! j " 0! We also need to choose a significance level of our test (recall that the significance level is the probability of Type I error: that is the probability of rejecting the null when the null is true) 12

14 One-Sided Alternatives (cont)! Having picked a significance level, α, we look up the (1 α) th percentile in a t distribution with n k 1 df and call this c, the critical value! With a 5% significance level and a 1-sided positive alternative, c is the 95th percentile. If H 0 is true, 95% of the time our estimate of t will lie below c, so 95% of the time we will correctly fail to reject the null.! If our estimate of t is greater than c, then either 1) H 0 is true, and an unlikely event (Type I error) has occurred; or 2) H 0 is false.! Since the odds of (1) are very small (assuming we choose a conventional significance level like 5%), we will choose the second interpretation (but never forget that (1) is a possibility!) 13

15 Rejection Rule! Thus, our rejection rule for H 1 : β j >0 is as follows:! We reject the null hypothesis if the t-statistic is greater than the critical value! We fail to reject the null hypothesis if the t statistic is less than the critical value 14

16 One-Sided Alternative y i = β 0 + β 1 x i1 + + β k x ik + u i H 0 : β j = 0 H 1 : β j > 0 Fail to reject (1 α) reject α 0 c 15

17 One-sided vs Two-sided! Because the t distribution is symmetric, testing H 1 : β j < 0 is straightforward.! The critical value is just the negative of before! We can reject the null if the t statistic < c, while if t > c then we fail to reject the null.! For a two-sided test, we set the critical value based on α/2 and reject H 1 : β j 0 if the absolute value of the t statistic > c In the 5% case, c would be the 97.5 percentile of the t distribution with n-k-1 df 16

18 Two-Sided Alternative y i = β 0 + β 1 X i1 + + β k X ik + u i H 0 : β j = 0 H 1 : β j 0 fail to reject reject α/2 (1 α) reject α/2 -c 0 c 17

19 Summary for H 0 : β j = 0! Unless otherwise stated, the alternative is assumed to be two-sided; in fact EViews will automatically generate t-statistics for you assuming your null value is 0 and your test is two sided.! If we reject the null, we typically say that our estimated β j is statistically significant at the 5% level or statistically different from zero (the null) at the 5% level! If we fail to reject the null, we typically say our estimate of β j is statistically insignificant or statistically indistinguishable from zero (the null). 18

20 Testing other hypotheses! A more general form of the t statistic recognizes that we may want to test something like H 0 : β j = a j! a j is a constant, e.g. test H 0 : β j =1! In this case, the appropriate t statistic is t! ˆ" j # a j se( ˆ" j ), where a j = 0 for the standard test! The t-statistic tells us how many standard deviations from the null our estimate lies.! As before, this can be used with a one-sided or two-sided alternative. 19

21 Confidence Intervals! Another way to use classical statistical testing is to construct a confidence interval using the same critical value as was used for a two-sided test! A (1 - α) % confidence interval is defined as ˆ! j ± c se( ˆ! # j ), where c is the 1- " $ % 2 in a t n) k )1 distribution! e.g. α=5%; 95% confidence interval & ' ( percentile This implies that the true value of β j will be contained in the interval in 95% of random samples. 20

22 Confidence Intervals (cont)! The confidence intervals are easily calculated since we have ˆ " j and se ˆ (" ) j! Once we have the confidence interval it is easy to do a two-tailed test! If testing H 0 : β j = a j against H 1 : β j a j, then Reject H 0 if a j does not lie in the confidence interval for your sample. Do not reject H 0 if a j does lie in the confidence interval for your sample. 21

23 Computing p-values for t tests! An alternative to the classical approach is to ask, what is the smallest significance level at which the null would be rejected?! So, compute the t statistic, and then look up what percentile it is in the appropriate t distribution--this is called the p-value Interpretation:! The p-value gives the probability of observing a t-statistic as large as we did, if the null is in fact true.! If p=0.01, this tells us we would obtain a t-statistic as big in magnitude (or bigger) in only 1% of samples, if the null is true.! Small p-values are evidence against the null hypothesis 22

24 EViews and p-values, t tests, etc.! Most computer packages will compute the p-value for you, assuming a two-sided test! If you really want a one-sided alternative, just divide the two-sided p-value by 2! EViews provides the t statistic, and p-value for H 0 : β j = 0 for you, in columns labeled t-statistic, Prob., respectively. It also gives you the coefficient estimate and standard error, which can be used to calculate confidence intervals. 23

25 Testing Hypotheses about a Linear Combination of Parameters! Suppose instead of testing whether β 1 is equal to a constant, you want to test whether it is equal to another parameter, e.g. H 0 : β 1 = β 2! This is a hypothesis test concerning 2 parameters (haven t seen this before)! Can use same basic procedure as before to form the t statistic if we write H 0 : β 1 -β 2 =0, t = ˆ! 1 " ˆ! 2 se( ˆ! 1 " ˆ! 2 ) 24

26 Testing Linear Combo (cont)! Problem/difficulty is calculating the standard error se( " ˆ 1 # " ˆ 2 ). Recall (from review) that Var " ˆ 1 # ˆ ( ) Since " se # ˆ! j ( ) = " se( ˆ! ) 1 se ˆ! 1 & ˆ! 2 ( " ) 2 = Var ˆ " 1 ( ) $ is an unbiased estimate of Var ˆ! j %2 { # ( ) + Var( " ˆ ) 2 # 2Cov( " ˆ 1," ˆ ) 2 ( ) $ + " se ˆ!2 %2 # $ & 2s12 %2 }1/2 where s 12 is an estimate of Cov( ˆ! 1, ˆ! ) 2 25

27 Testing a Linear Combo (cont)! So, to use formula, need s 12, which standard EViews output does not give! Many packages will have an option to get it, or will just perform the test for you! In lab you will learn how to use EViews to test whether ˆ! 1 = ˆ! 2.! More generally, you can always restate the problem to get the test you want 26

28 Example! Suppose you are interested in the effect of campaign expenditures by Party A and Party B on votes received by Party A s candidate! Model is votea = β 0 + β 1 log(expenda) + β 2 log(expendb) + β 3 prtystra + u! You might think that expenda and expendb should have an equal yet opposite effect on votes received by Party A s candidate i.e. H 0 :β 1 =-β 2 We could define a new parameter, θ 1, where θ 1 = β 1 +β 2 27

29 Example (cont):! If we could generate an estimate of se(θ 1 ) we could test H 0 : θ 1 = β 1 + β 2 = 0! We can do this by rewriting the model! recall: β 1 = θ 1 β 2, so substitute in and rearrange votea = β 0 + θ 1 log(expenda) + β 2 [log(expendb)-log (expenda)] + β 3 prtystra + u! This is the same model as before, but now we get a direct estimate of θ 1 and the standard error of θ 1 which makes for easier hypothesis testing. 28

30 Example (cont):! The t-statistic can be written: ˆ! t = 1 se( ˆ! ) 1! In fact, any linear combination of parameters could be tested in a similar manner! Other examples of hypotheses about a single linear combination of parameters:! β 1 = 1 + β 2! β 1 = 5β 2! β 1 = -1/2β 2 ; etc 29

31 Multiple Linear Restrictions! Everything we ve done so far has involved testing a single linear restriction, (e.g. β 1 = 0 or β 1 = β 2 )! However, we may want to jointly test multiple hypotheses about our parameters! A typical example is testing exclusion restrictions we want to know if a group of parameters are all equal to zero Example: Suppose that you are a college administrator and you want to know what determines which students will do well 30

32 Exclusion Restrictions! You might estimate the following model (US applicable): ColGPA=β 0 + β 1 highscgpa+ β 2 VerbSAT+β 3 MathSAT+u! Once we control for high-school GPA do the SAT scores have any effect on (or at least predictive power for) college GPA?! Here, the null hypothesis is H 0 : β 2 =0, β 3 =0 (where each of these is referred to as an exclusion restriction ; the alternative hypothesis is that H 0 is not true.! If we can t reject the null, it suggests we should drop VerbSAT and MathSAT from the model! While it s tempting to do separate t-tests on β 2 and β 3, this is not appropriate, because the t-test is done assuming no restrictions are simultaneously being placed on other parameter values (i.e. the t-test is designed for single exclusion restrictions) 31

33 Testing Exclusion Restrictions! We want to know if the parameters are jointly significant at a given level! For example it is possible that neither β 2 nor β 3 is individually significant at, say, the 5% level but that they are jointly significant at that level! To test the restrictions we need to estimate 1. The Restricted Model--without x 2 and x 3 2. The Unrestricted Model--with x 1, x 2 and x 3! We can use the SSR s from each of these regressions to form a test statistic 32

34 Exclusion Restrictions (cont)! Clearly, the SSR will be greater in the restricted model (because with fewer RHS variables, we explain less of the variation in y and so get larger residuals, on average).! Intuitively, we want to know if the increase in SSR is big enough to warrant inclusion of x 1 and x 2! The F-statistic in the general case with q exclusion restrictions is: F! (SSR r " SSR ur ) / q SSR ur / (n " k " 1) Where r refers to the restricted model, and ur refers to the unrestricted model. 33

35 The F statistic! The F statistic is always positive, since the SSR from the restricted model can t be less than the SSR from the unrestricted model! Essentially the F statistic is measuring the relative increase in SSR when moving from the unrestricted to the restricted model! q = number of restrictions, or the numerator degrees of freedom df r df ur! n k 1 = denominator degrees of freedom = df ur 34

36 The F statistic (cont)! To decide if the increase in SSR when we move to a restricted model is big enough to reject the exclusions, we need to know about the sampling distribution of our F statistic under H 0! It turns our that F~F q,n-k-1, where q is the numerator degrees of freedom and n-k-1 is the denominator degrees of freedom.! We can use the back of the text (Table G.3) or EViews to tabulate the F-distribution 35

37 The F statistic (cont) f(f) fail to reject Reject H 0 at α significance level if F > c (1 α) 0 c α reject F 36

38 Summary of the F statistic! If H 0 is rejected we say, for example, that x 2 and x 3 are jointly significant! If H 0 is not rejected the variables are jointly insignificant! The F-test is particularly useful when a group of the x variables are highly correlated e.g. expenditures on hospitals, doctors, nurses, etc. We don t know which are more important Multicollinearity won t allow for individual t-tests (due to high standard errors of individual coefficient estimates) 37

39 The R 2 form of the F statistic! Because the SSR s may be large and unwieldy, an alternative form of the formula is useful! We use the fact that SSR = SST(1 R 2 ) for any regression, so can substitute in for SSR r and SSR ur F! ( R 2 2 ur " R ) r q ( ) n " k " " R ur ( )! Again, r = restricted and ur = unrestricted! Notice in the numerator that the unrestricted R- squared comes first. 38

40 Overall Significance! A special case of exclusion restrictions is to test H 0 : β 1 = β 2 = = β k = 0! Since the R 2 from a model with only an intercept will be zero, the F statistic is simply R 2 / k F = (1! R 2 ) / (n! k! 1)! If we can t reject the null, this suggests that data are consistent with the RHS variables explaining none of the variation in y.! Even if the R-squared is very small, we can use this statistic to test whether the x s are explaining any of y. 39

41 General Linear Restrictions! The basic form of the F statistic will work for any set of linear restrictions (not just setting a couple betas equal to zero)! First estimate the unrestricted model and then estimate the restricted model! In each case, make note of the SSR! Imposing the restrictions can be tricky will likely have to redefine variables again 40

42 Example:! Use same voting model as before! Model is votea = β 0 + β 1 log(expenda) + β 2 log(expendb) + β 3 prtystra + u! now suppose null is H 0 : β 1 = 1, β 3 = 0! Substituting in the restrictions: votea = β 0 + log (expenda) + β 2 log(expendb) + u, so use votealog(expenda)= β 0 + β 2 log(expendb) + u as the restricted model.! Note: Can t use the R-squared form of the F- statistic here because SSTs will be different. 41

43 F-statistic Summary! Just as with t statistics, p-values can be calculated by looking up the percentile in the appropriate F distribution! In lab, you ll learn how to do this in EViews.! If only one exclusion is being tested, then F = t 2, and the p-values will be the same 42

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Inference in Regression Analysis

Inference in Regression Analysis ECNS 561 Inference Inference in Regression Analysis Up to this point 1.) OLS is unbiased 2.) OLS is BLUE (best linear unbiased estimator i.e., the variance is smallest among linear unbiased estimators)

More information

CHAPTER 4. > 0, where β

CHAPTER 4. > 0, where β CHAPTER 4 SOLUTIONS TO PROBLEMS 4. (i) and (iii) generally cause the t statistics not to have a t distribution under H. Homoskedasticity is one of the CLM assumptions. An important omitted variable violates

More information

Inference in Regression Model

Inference in Regression Model Inference in Regression Model Christopher Taber Department of Economics University of Wisconsin-Madison March 25, 2009 Outline 1 Final Step of Classical Linear Regression Model 2 Confidence Intervals 3

More information

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1 Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Multiple Regression: Inference

Multiple Regression: Inference Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled

More information

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define

More information

Lectures 5 & 6: Hypothesis Testing

Lectures 5 & 6: Hypothesis Testing Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across

More information

Problem 4.1. Problem 4.3

Problem 4.1. Problem 4.3 BOSTON COLLEGE Department of Economics EC 228 01 Econometric Methods Fall 2008, Prof. Baum, Ms. Phillips (tutor), Mr. Dmitriev (grader) Problem Set 3 Due at classtime, Thursday 14 Oct 2008 Problem 4.1

More information

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators 1 2 Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE Hüseyin Taştan 1 1 Yıldız Technical University Department of Economics These presentation notes are based on Introductory

More information

Multiple Regression Analysis: Heteroskedasticity

Multiple Regression Analysis: Heteroskedasticity Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing

More information

Statistical Inference. Part IV. Statistical Inference

Statistical Inference. Part IV. Statistical Inference Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided

More information

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor 1. The regression equation 2. Estimating the equation 3. Assumptions required for

More information

THE MULTIVARIATE LINEAR REGRESSION MODEL

THE MULTIVARIATE LINEAR REGRESSION MODEL THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Multiple Regression Analysis. Part III. Multiple Regression Analysis Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant

More information

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random

More information

Inferences for Regression

Inferences for Regression Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

CHAPTER 6: SPECIFICATION VARIABLES

CHAPTER 6: SPECIFICATION VARIABLES Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero

More information

F Tests and F statistics

F Tests and F statistics F Tests and F statistics Testing Linear estrictions F Stats and F Tests F Distributions F stats (w/ ) F Stats and tstat s eported F Stat's in OLS Output Example I: Bodyfat Babies and Bathwater F Stats,

More information

ECO220Y Simple Regression: Testing the Slope

ECO220Y Simple Regression: Testing the Slope ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x

More information

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression

More information

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit

More information

ECNS 561 Multiple Regression Analysis

ECNS 561 Multiple Regression Analysis ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking

More information

LECTURE 5. Introduction to Econometrics. Hypothesis testing

LECTURE 5. Introduction to Econometrics. Hypothesis testing LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

The general linear regression with k explanatory variables is just an extension of the simple regression as follows 3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because

More information

Lecture 5: Hypothesis testing with the classical linear model

Lecture 5: Hypothesis testing with the classical linear model Lecture 5: Hypothesis testing with the classical linear model Assumption MLR6: Normality MLR6 is not one of the Gauss-Markov assumptions. It s not necessary to assume the error is normally distributed

More information

Introduction to Econometrics. Multiple Regression (2016/2017)

Introduction to Econometrics. Multiple Regression (2016/2017) Introduction to Econometrics STAT-S-301 Multiple Regression (016/017) Lecturer: Yves Dominicy Teaching Assistant: Elise Petit 1 OLS estimate of the TS/STR relation: OLS estimate of the Test Score/STR relation:

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

Chapter 3 Multiple Regression Complete Example

Chapter 3 Multiple Regression Complete Example Department of Quantitative Methods & Information Systems ECON 504 Chapter 3 Multiple Regression Complete Example Spring 2013 Dr. Mohammad Zainal Review Goals After completing this lecture, you should be

More information

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C = Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =

More information

The Simple Regression Model. Simple Regression Model 1

The Simple Regression Model. Simple Regression Model 1 The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising

More information

Chapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression

Chapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression Chapter 7 Hypothesis Tests and Confidence Intervals in Multiple Regression Outline 1. Hypothesis tests and confidence intervals for a single coefficie. Joint hypothesis tests on multiple coefficients 3.

More information

Statistical Inference with Regression Analysis

Statistical Inference with Regression Analysis Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing

More information

STAT Chapter 8: Hypothesis Tests

STAT Chapter 8: Hypothesis Tests STAT 515 -- Chapter 8: Hypothesis Tests CIs are possibly the most useful forms of inference because they give a range of reasonable values for a parameter. But sometimes we want to know whether one particular

More information

STA121: Applied Regression Analysis

STA121: Applied Regression Analysis STA121: Applied Regression Analysis Linear Regression Analysis - Chapters 3 and 4 in Dielman Artin Department of Statistical Science September 15, 2009 Outline 1 Simple Linear Regression Analysis 2 Using

More information

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Answer Key: Problem Set 6

Answer Key: Problem Set 6 : Problem Set 6 1. Consider a linear model to explain monthly beer consumption: beer = + inc + price + educ + female + u 0 1 3 4 E ( u inc, price, educ, female ) = 0 ( u inc price educ female) σ inc var,,,

More information

2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0

2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0 Introduction to Econometrics Midterm April 26, 2011 Name Student ID MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. (5,000 credit for each correct

More information

Lecture 3: Inference in SLR

Lecture 3: Inference in SLR Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals

More information

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests: One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get

More information

Business Statistics. Lecture 10: Course Review

Business Statistics. Lecture 10: Course Review Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,

More information

Intro to Linear Regression

Intro to Linear Regression Intro to Linear Regression Introduction to Regression Regression is a statistical procedure for modeling the relationship among variables to predict the value of a dependent variable from one or more predictor

More information

Chapter 27 Summary Inferences for Regression

Chapter 27 Summary Inferences for Regression Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test

More information

Lecture 4: Testing Stuff

Lecture 4: Testing Stuff Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted

More information

Lecture 10 Multiple Linear Regression

Lecture 10 Multiple Linear Regression Lecture 10 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 10-1 Topic Overview Multiple Linear Regression Model 10-2 Data for Multiple Regression Y i is the response variable

More information

Intro to Linear Regression

Intro to Linear Regression Intro to Linear Regression Introduction to Regression Regression is a statistical procedure for modeling the relationship among variables to predict the value of a dependent variable from one or more predictor

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique

More information

Lab 11 - Heteroskedasticity

Lab 11 - Heteroskedasticity Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction

More information

Distribution-Free Procedures (Devore Chapter Fifteen)

Distribution-Free Procedures (Devore Chapter Fifteen) Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

Multiple Linear Regression

Multiple Linear Regression Chapter 3 Multiple Linear Regression 3.1 Introduction Multiple linear regression is in some ways a relatively straightforward extension of simple linear regression that allows for more than one independent

More information

Chapter 8 Handout: Interval Estimates and Hypothesis Testing

Chapter 8 Handout: Interval Estimates and Hypothesis Testing Chapter 8 Handout: Interval Estimates and Hypothesis esting Preview Clint s Assignment: aking Stock General Properties of the Ordinary Least Squares (OLS) Estimation Procedure Estimate Reliability: Interval

More information

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 CIVL - 7904/8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 Chi-square Test How to determine the interval from a continuous distribution I = Range 1 + 3.322(logN) I-> Range of the class interval

More information

Using SPSS for One Way Analysis of Variance

Using SPSS for One Way Analysis of Variance Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

appstats27.notebook April 06, 2017

appstats27.notebook April 06, 2017 Chapter 27 Objective Students will conduct inference on regression and analyze data to write a conclusion. Inferences for Regression An Example: Body Fat and Waist Size pg 634 Our chapter example revolves

More information

Chapter 11 Handout: Hypothesis Testing and the Wald Test

Chapter 11 Handout: Hypothesis Testing and the Wald Test Chapter 11 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic Manipulation o Wald Test Restricted Regression Reflects Unrestricted Regression

More information

Recall that a measure of fit is the sum of squared residuals: where. The F-test statistic may be written as:

Recall that a measure of fit is the sum of squared residuals: where. The F-test statistic may be written as: 1 Joint hypotheses The null and alternative hypotheses can usually be interpreted as a restricted model ( ) and an model ( ). In our example: Note that if the model fits significantly better than the restricted

More information

Introduction to Econometrics. Multiple Regression

Introduction to Econometrics. Multiple Regression Introduction to Econometrics The statistical analysis of economic (and related) data STATS301 Multiple Regression Titulaire: Christopher Bruffaerts Assistant: Lorenzo Ricci 1 OLS estimate of the TS/STR

More information

Multiple Regression Analysis

Multiple Regression Analysis Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple

More information

Chapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics?

Chapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics? Econometrics and Quantitative Analysis Using Econometrics: A Practical Guide A.H. Studenmund 6th Edition. Addison Wesley Longman Chapter 1 An Overview of Regression Analysis Instructor: Dr. Samir Safi

More information

Economics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Fall 2016 Instructor: Martin Farnham

Economics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Fall 2016 Instructor: Martin Farnham Economics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Fall 2016 Instructor: Martin Farnham Last name (family name): First name (given name):

More information

Lecture 2 Multiple Regression and Tests

Lecture 2 Multiple Regression and Tests Lecture 2 and s Dr.ssa Rossella Iraci Capuccinello 2017-18 Simple Regression Model The random variable of interest, y, depends on a single factor, x 1i, and this is an exogenous variable. The true but

More information

Econometrics Review questions for exam

Econometrics Review questions for exam Econometrics Review questions for exam Nathaniel Higgins nhiggins@jhu.edu, 1. Suppose you have a model: y = β 0 x 1 + u You propose the model above and then estimate the model using OLS to obtain: ŷ =

More information

1 Independent Practice: Hypothesis tests for one parameter:

1 Independent Practice: Hypothesis tests for one parameter: 1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)

More information

Hint: The following equation converts Celsius to Fahrenheit: F = C where C = degrees Celsius F = degrees Fahrenheit

Hint: The following equation converts Celsius to Fahrenheit: F = C where C = degrees Celsius F = degrees Fahrenheit Amherst College Department of Economics Economics 360 Fall 2014 Exam 1: Solutions 1. (10 points) The following table in reports the summary statistics for high and low temperatures in Key West, FL from

More information

Business Statistics. Lecture 9: Simple Regression

Business Statistics. Lecture 9: Simple Regression Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals

More information

ECO375 Tutorial 4 Introduction to Statistical Inference

ECO375 Tutorial 4 Introduction to Statistical Inference ECO375 Tutorial 4 Introduction to Statistical Inference Matt Tudball University of Toronto Mississauga October 19, 2017 Matt Tudball (University of Toronto) ECO375H5 October 19, 2017 1 / 26 Statistical

More information

Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test

Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test Amherst College Department of Economics Economics 360 Fall 2012 Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic

More information

Solutions to Problem Set 4 (Due November 13) Maximum number of points for Problem set 4 is: 66. Problem C 6.1

Solutions to Problem Set 4 (Due November 13) Maximum number of points for Problem set 4 is: 66. Problem C 6.1 Solutions to Problem Set 4 (Due November 13) EC 228 01, Fall 2013 Prof. Baum, Mr. Lim Maximum number of points for Problem set 4 is: 66 Problem C 6.1 (i) (3 pts.) If the presence of the incinerator depresses

More information

CHAPTER 9: HYPOTHESIS TESTING

CHAPTER 9: HYPOTHESIS TESTING CHAPTER 9: HYPOTHESIS TESTING THE SECOND LAST EXAMPLE CLEARLY ILLUSTRATES THAT THERE IS ONE IMPORTANT ISSUE WE NEED TO EXPLORE: IS THERE (IN OUR TWO SAMPLES) SUFFICIENT STATISTICAL EVIDENCE TO CONCLUDE

More information

STA Module 10 Comparing Two Proportions

STA Module 10 Comparing Two Proportions STA 2023 Module 10 Comparing Two Proportions Learning Objectives Upon completing this module, you should be able to: 1. Perform large-sample inferences (hypothesis test and confidence intervals) to compare

More information

σ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity

σ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity Comparison of SLR and MLR analysis: What s New? Roadmap Multicollinearity and standard errors F Tests of linear restrictions F stats, adjusted R-squared, RMSE and t stats Playing with Bodyfat: F tests

More information

Lecture 3: Multiple Regression

Lecture 3: Multiple Regression Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

Correlation and regression

Correlation and regression NST 1B Experimental Psychology Statistics practical 1 Correlation and regression Rudolf Cardinal & Mike Aitken 11 / 12 November 2003 Department of Experimental Psychology University of Cambridge Handouts:

More information

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2

LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2 LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2 Data Analysis: The mean egg masses (g) of the two different types of eggs may be exactly the same, in which case you may be tempted to accept

More information

Simple Linear Regression: One Quantitative IV

Simple Linear Regression: One Quantitative IV Simple Linear Regression: One Quantitative IV Linear regression is frequently used to explain variation observed in a dependent variable (DV) with theoretically linked independent variables (IV). For example,

More information

POL 681 Lecture Notes: Statistical Interactions

POL 681 Lecture Notes: Statistical Interactions POL 681 Lecture Notes: Statistical Interactions 1 Preliminaries To this point, the linear models we have considered have all been interpreted in terms of additive relationships. That is, the relationship

More information

Chapter 8 Heteroskedasticity

Chapter 8 Heteroskedasticity Chapter 8 Walter R. Paczkowski Rutgers University Page 1 Chapter Contents 8.1 The Nature of 8. Detecting 8.3 -Consistent Standard Errors 8.4 Generalized Least Squares: Known Form of Variance 8.5 Generalized

More information

LECTURE 15: SIMPLE LINEAR REGRESSION I

LECTURE 15: SIMPLE LINEAR REGRESSION I David Youngberg BSAD 20 Montgomery College LECTURE 5: SIMPLE LINEAR REGRESSION I I. From Correlation to Regression a. Recall last class when we discussed two basic types of correlation (positive and negative).

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Final Exam - Solutions

Final Exam - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis March 19, 2010 Instructor: John Parman Final Exam - Solutions You have until 5:30pm to complete this exam. Please remember to put your

More information

Measuring the fit of the model - SSR

Measuring the fit of the model - SSR Measuring the fit of the model - SSR Once we ve determined our estimated regression line, we d like to know how well the model fits. How far/close are the observations to the fitted line? One way to do

More information

Hypothesis Tests and Confidence Intervals in Multiple Regression

Hypothesis Tests and Confidence Intervals in Multiple Regression Hypothesis Tests and Confidence Intervals in Multiple Regression (SW Chapter 7) Outline 1. Hypothesis tests and confidence intervals for one coefficient. Joint hypothesis tests on multiple coefficients

More information

Chapter 1 Review of Equations and Inequalities

Chapter 1 Review of Equations and Inequalities Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve

More information

Steps in Regression Analysis

Steps in Regression Analysis MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4) 2-1 Steps in Regression Analysis 1. Review the literature and develop the theoretical model 2. Specify the model:

More information

Multiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

Multiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Multiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Introduction When you perform statistical inference, you are primarily doing one of two things: Estimating the boundaries

More information

Problem Set 4 ANSWERS

Problem Set 4 ANSWERS Economics 20 Problem Set 4 ANSWERS Prof. Patricia M. Anderson 1. Suppose that our variable for consumption is measured with error, so cons = consumption + e 0, where e 0 is uncorrelated with inc, educ

More information