Multiple Regression Analysis
|
|
- Percival Blankenship
- 6 years ago
- Views:
Transcription
1 Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x β k x k + u 2. Inference 0
2 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators 2. That OLS is BLUE, given the Gauss-Markov assumptions! This still leaves open the shape of the distribution of the OLS estimators! To do classical hypothesis testing, we need to make an assumption about this distribution Assumption 6 (Normality): u is independent of x 1, x 2,, x k and u is normally distributed with zero mean and variance σ 2 : u ~ Normal(0, σ 2 ) 1
3 CLM Assumptions (cont)! This is a very strong assumption, it implies: E(u x)=e(u)=0, and Var(u x)=var(u)=σ 2! Assumptions 1-6 are called the Classical Linear Model Assumptions, CLM Assumptions! Under CLM, OLS is not only BLUE, but is the minimum variance unbiased estimator - no longer limited to the class of linear estimators 2
4 CLM Assumptions (cont)! We can summarize the population assumptions of CLM as follows! y x ~ Normal(β 0 + β 1 x β k x k, σ 2 )! While for now we just assume normality, clear that sometimes not the case! Example: Suppose we re interested in modeling house prices (our y variable) as a function of x variables: House prices are never below zero (a normally distributed variable takes on negative values); same problem with wages Could take log of house prices to convert it to something close to normal; analysis of wages as y variable often involves taking logs Or, if our sample is large enough, non-normality ceases to be a problem, thanks to the Central Limit Theorem. 3
5 The homoskedastic normal distribution with a single explanatory variable f(y x) y.. E(y x) = β 0 + β 1 x Normal distribution x 1 x 2 4
6 Normal Sampling Distributions! Under the CLM assumptions, conditional on the sample values of the independent variables, (x), ˆ! j ~ N "! j,var( ˆ! j ) $ # %! Why? Because each estimated β is a linear combination of the errors, u, in the sample. Recall: ˆ " j = " j + n $ i=1 n $ (x ij # x j )u i (x ij # x j ) 2, and u i ~ N(0,% 2 ) i=1! Any linear combination of iid normal random variables is normal (recall Appendix C) 5
7 Normal Sampling Distributions thus, ( ˆ! j "! ) j sd ( ˆ! ) j ~ Normal( 0,1)! To standardize (as before), we just subtract the mean and divide by the standard deviation! The fact that the estimated β s are distributed normally has further implications we will use: 1) Any linear combination of the estimated β s will be normally distributed. 2) Any subset of the estimated β s will be jointly normally distributed 6
8 The t Test! Testing hypotheses about a single population parameter! Remember, the β j are unknown! However, we can hypothesize about the value of β j and use statistical inference to test our hypothesis! To perform a hypothesis test we first note that under CLM: ( ˆ! j "! j ) se( ˆ! j ) ~ t n" k "1 7
9 The t Test (cont)! Knowing the sampling distribution for the standardized estimator allows us to carry out hypothesis tests! Start with a null hypothesis! For example, H 0 :! j = 0! The null here is making the claim that variable x j has no effect on y. 8
10 The t test (cont)! To perform our test (given this particular stating of the null hypothesis) we first form the t statistic for our estimate of β j for ˆ! j : t ˆ! j " ˆ! j # 0 se( ˆ! ) = j ˆ! j se( ˆ! ) j! Notice that we subtract the value of β j under the null (in this case, 0), from our estimate of β j. Intuitively, if our estimate is close to the null, our t- statistic will be close to zero, and we will be unable to reject the null that β j =0. 9
11 Things to notice about t-stat! The denominator is always positive. So for a zero null, the t-statistic will always take on the sign of your estimate of β j. For positive estimated β j it will be positive, for negative estimated β j it will be negative.! We can interpret the t-statistic as giving us an estimate of how many standard deviations from the null (0, in this case) our estimate of β j lies. Say it equals 3. This suggests that it lies three standard deviations to the right of 0. If it equals -3, this suggests it lies 3 sd s to the left of 0. 10
12 Intuition for t Test! Because it is a continuous random variable, the estimated β j will never precisely equal 0 even if the underlying population parameter β j equals 0! We want to know how far estimated β j is from 0 Essentially, if our estimated β j lies near the null, we think there s a good chance that estimate would be generated if the null is true (so, we will fail to reject the null); if our estimated β j lies far away from the null, we think it s unlikely that estimate would be generated under the null: thus we reject the null.! The question is, How far is far enough from the null to warrant rejection?! We need a rejection rule! 11
13 t-test: One-Sided Alternatives! Besides our null, H 0, we need an alternative hypothesis, H 1, and a significance level! H 1 may be one-sided, or two-sided! Examples of different one-sided alternatives include H 1 :! j < 0 or H 1 :! j > 0! A 2-sided alternative is H 1 :! j " 0! We also need to choose a significance level of our test (recall that the significance level is the probability of Type I error: that is the probability of rejecting the null when the null is true) 12
14 One-Sided Alternatives (cont)! Having picked a significance level, α, we look up the (1 α) th percentile in a t distribution with n k 1 df and call this c, the critical value! With a 5% significance level and a 1-sided positive alternative, c is the 95th percentile. If H 0 is true, 95% of the time our estimate of t will lie below c, so 95% of the time we will correctly fail to reject the null.! If our estimate of t is greater than c, then either 1) H 0 is true, and an unlikely event (Type I error) has occurred; or 2) H 0 is false.! Since the odds of (1) are very small (assuming we choose a conventional significance level like 5%), we will choose the second interpretation (but never forget that (1) is a possibility!) 13
15 Rejection Rule! Thus, our rejection rule for H 1 : β j >0 is as follows:! We reject the null hypothesis if the t-statistic is greater than the critical value! We fail to reject the null hypothesis if the t statistic is less than the critical value 14
16 One-Sided Alternative y i = β 0 + β 1 x i1 + + β k x ik + u i H 0 : β j = 0 H 1 : β j > 0 Fail to reject (1 α) reject α 0 c 15
17 One-sided vs Two-sided! Because the t distribution is symmetric, testing H 1 : β j < 0 is straightforward.! The critical value is just the negative of before! We can reject the null if the t statistic < c, while if t > c then we fail to reject the null.! For a two-sided test, we set the critical value based on α/2 and reject H 1 : β j 0 if the absolute value of the t statistic > c In the 5% case, c would be the 97.5 percentile of the t distribution with n-k-1 df 16
18 Two-Sided Alternative y i = β 0 + β 1 X i1 + + β k X ik + u i H 0 : β j = 0 H 1 : β j 0 fail to reject reject α/2 (1 α) reject α/2 -c 0 c 17
19 Summary for H 0 : β j = 0! Unless otherwise stated, the alternative is assumed to be two-sided; in fact EViews will automatically generate t-statistics for you assuming your null value is 0 and your test is two sided.! If we reject the null, we typically say that our estimated β j is statistically significant at the 5% level or statistically different from zero (the null) at the 5% level! If we fail to reject the null, we typically say our estimate of β j is statistically insignificant or statistically indistinguishable from zero (the null). 18
20 Testing other hypotheses! A more general form of the t statistic recognizes that we may want to test something like H 0 : β j = a j! a j is a constant, e.g. test H 0 : β j =1! In this case, the appropriate t statistic is t! ˆ" j # a j se( ˆ" j ), where a j = 0 for the standard test! The t-statistic tells us how many standard deviations from the null our estimate lies.! As before, this can be used with a one-sided or two-sided alternative. 19
21 Confidence Intervals! Another way to use classical statistical testing is to construct a confidence interval using the same critical value as was used for a two-sided test! A (1 - α) % confidence interval is defined as ˆ! j ± c se( ˆ! # j ), where c is the 1- " $ % 2 in a t n) k )1 distribution! e.g. α=5%; 95% confidence interval & ' ( percentile This implies that the true value of β j will be contained in the interval in 95% of random samples. 20
22 Confidence Intervals (cont)! The confidence intervals are easily calculated since we have ˆ " j and se ˆ (" ) j! Once we have the confidence interval it is easy to do a two-tailed test! If testing H 0 : β j = a j against H 1 : β j a j, then Reject H 0 if a j does not lie in the confidence interval for your sample. Do not reject H 0 if a j does lie in the confidence interval for your sample. 21
23 Computing p-values for t tests! An alternative to the classical approach is to ask, what is the smallest significance level at which the null would be rejected?! So, compute the t statistic, and then look up what percentile it is in the appropriate t distribution--this is called the p-value Interpretation:! The p-value gives the probability of observing a t-statistic as large as we did, if the null is in fact true.! If p=0.01, this tells us we would obtain a t-statistic as big in magnitude (or bigger) in only 1% of samples, if the null is true.! Small p-values are evidence against the null hypothesis 22
24 EViews and p-values, t tests, etc.! Most computer packages will compute the p-value for you, assuming a two-sided test! If you really want a one-sided alternative, just divide the two-sided p-value by 2! EViews provides the t statistic, and p-value for H 0 : β j = 0 for you, in columns labeled t-statistic, Prob., respectively. It also gives you the coefficient estimate and standard error, which can be used to calculate confidence intervals. 23
25 Testing Hypotheses about a Linear Combination of Parameters! Suppose instead of testing whether β 1 is equal to a constant, you want to test whether it is equal to another parameter, e.g. H 0 : β 1 = β 2! This is a hypothesis test concerning 2 parameters (haven t seen this before)! Can use same basic procedure as before to form the t statistic if we write H 0 : β 1 -β 2 =0, t = ˆ! 1 " ˆ! 2 se( ˆ! 1 " ˆ! 2 ) 24
26 Testing Linear Combo (cont)! Problem/difficulty is calculating the standard error se( " ˆ 1 # " ˆ 2 ). Recall (from review) that Var " ˆ 1 # ˆ ( ) Since " se # ˆ! j ( ) = " se( ˆ! ) 1 se ˆ! 1 & ˆ! 2 ( " ) 2 = Var ˆ " 1 ( ) $ is an unbiased estimate of Var ˆ! j %2 { # ( ) + Var( " ˆ ) 2 # 2Cov( " ˆ 1," ˆ ) 2 ( ) $ + " se ˆ!2 %2 # $ & 2s12 %2 }1/2 where s 12 is an estimate of Cov( ˆ! 1, ˆ! ) 2 25
27 Testing a Linear Combo (cont)! So, to use formula, need s 12, which standard EViews output does not give! Many packages will have an option to get it, or will just perform the test for you! In lab you will learn how to use EViews to test whether ˆ! 1 = ˆ! 2.! More generally, you can always restate the problem to get the test you want 26
28 Example! Suppose you are interested in the effect of campaign expenditures by Party A and Party B on votes received by Party A s candidate! Model is votea = β 0 + β 1 log(expenda) + β 2 log(expendb) + β 3 prtystra + u! You might think that expenda and expendb should have an equal yet opposite effect on votes received by Party A s candidate i.e. H 0 :β 1 =-β 2 We could define a new parameter, θ 1, where θ 1 = β 1 +β 2 27
29 Example (cont):! If we could generate an estimate of se(θ 1 ) we could test H 0 : θ 1 = β 1 + β 2 = 0! We can do this by rewriting the model! recall: β 1 = θ 1 β 2, so substitute in and rearrange votea = β 0 + θ 1 log(expenda) + β 2 [log(expendb)-log (expenda)] + β 3 prtystra + u! This is the same model as before, but now we get a direct estimate of θ 1 and the standard error of θ 1 which makes for easier hypothesis testing. 28
30 Example (cont):! The t-statistic can be written: ˆ! t = 1 se( ˆ! ) 1! In fact, any linear combination of parameters could be tested in a similar manner! Other examples of hypotheses about a single linear combination of parameters:! β 1 = 1 + β 2! β 1 = 5β 2! β 1 = -1/2β 2 ; etc 29
31 Multiple Linear Restrictions! Everything we ve done so far has involved testing a single linear restriction, (e.g. β 1 = 0 or β 1 = β 2 )! However, we may want to jointly test multiple hypotheses about our parameters! A typical example is testing exclusion restrictions we want to know if a group of parameters are all equal to zero Example: Suppose that you are a college administrator and you want to know what determines which students will do well 30
32 Exclusion Restrictions! You might estimate the following model (US applicable): ColGPA=β 0 + β 1 highscgpa+ β 2 VerbSAT+β 3 MathSAT+u! Once we control for high-school GPA do the SAT scores have any effect on (or at least predictive power for) college GPA?! Here, the null hypothesis is H 0 : β 2 =0, β 3 =0 (where each of these is referred to as an exclusion restriction ; the alternative hypothesis is that H 0 is not true.! If we can t reject the null, it suggests we should drop VerbSAT and MathSAT from the model! While it s tempting to do separate t-tests on β 2 and β 3, this is not appropriate, because the t-test is done assuming no restrictions are simultaneously being placed on other parameter values (i.e. the t-test is designed for single exclusion restrictions) 31
33 Testing Exclusion Restrictions! We want to know if the parameters are jointly significant at a given level! For example it is possible that neither β 2 nor β 3 is individually significant at, say, the 5% level but that they are jointly significant at that level! To test the restrictions we need to estimate 1. The Restricted Model--without x 2 and x 3 2. The Unrestricted Model--with x 1, x 2 and x 3! We can use the SSR s from each of these regressions to form a test statistic 32
34 Exclusion Restrictions (cont)! Clearly, the SSR will be greater in the restricted model (because with fewer RHS variables, we explain less of the variation in y and so get larger residuals, on average).! Intuitively, we want to know if the increase in SSR is big enough to warrant inclusion of x 1 and x 2! The F-statistic in the general case with q exclusion restrictions is: F! (SSR r " SSR ur ) / q SSR ur / (n " k " 1) Where r refers to the restricted model, and ur refers to the unrestricted model. 33
35 The F statistic! The F statistic is always positive, since the SSR from the restricted model can t be less than the SSR from the unrestricted model! Essentially the F statistic is measuring the relative increase in SSR when moving from the unrestricted to the restricted model! q = number of restrictions, or the numerator degrees of freedom df r df ur! n k 1 = denominator degrees of freedom = df ur 34
36 The F statistic (cont)! To decide if the increase in SSR when we move to a restricted model is big enough to reject the exclusions, we need to know about the sampling distribution of our F statistic under H 0! It turns our that F~F q,n-k-1, where q is the numerator degrees of freedom and n-k-1 is the denominator degrees of freedom.! We can use the back of the text (Table G.3) or EViews to tabulate the F-distribution 35
37 The F statistic (cont) f(f) fail to reject Reject H 0 at α significance level if F > c (1 α) 0 c α reject F 36
38 Summary of the F statistic! If H 0 is rejected we say, for example, that x 2 and x 3 are jointly significant! If H 0 is not rejected the variables are jointly insignificant! The F-test is particularly useful when a group of the x variables are highly correlated e.g. expenditures on hospitals, doctors, nurses, etc. We don t know which are more important Multicollinearity won t allow for individual t-tests (due to high standard errors of individual coefficient estimates) 37
39 The R 2 form of the F statistic! Because the SSR s may be large and unwieldy, an alternative form of the formula is useful! We use the fact that SSR = SST(1 R 2 ) for any regression, so can substitute in for SSR r and SSR ur F! ( R 2 2 ur " R ) r q ( ) n " k " " R ur ( )! Again, r = restricted and ur = unrestricted! Notice in the numerator that the unrestricted R- squared comes first. 38
40 Overall Significance! A special case of exclusion restrictions is to test H 0 : β 1 = β 2 = = β k = 0! Since the R 2 from a model with only an intercept will be zero, the F statistic is simply R 2 / k F = (1! R 2 ) / (n! k! 1)! If we can t reject the null, this suggests that data are consistent with the RHS variables explaining none of the variation in y.! Even if the R-squared is very small, we can use this statistic to test whether the x s are explaining any of y. 39
41 General Linear Restrictions! The basic form of the F statistic will work for any set of linear restrictions (not just setting a couple betas equal to zero)! First estimate the unrestricted model and then estimate the restricted model! In each case, make note of the SSR! Imposing the restrictions can be tricky will likely have to redefine variables again 40
42 Example:! Use same voting model as before! Model is votea = β 0 + β 1 log(expenda) + β 2 log(expendb) + β 3 prtystra + u! now suppose null is H 0 : β 1 = 1, β 3 = 0! Substituting in the restrictions: votea = β 0 + log (expenda) + β 2 log(expendb) + u, so use votealog(expenda)= β 0 + β 2 log(expendb) + u as the restricted model.! Note: Can t use the R-squared form of the F- statistic here because SSTs will be different. 41
43 F-statistic Summary! Just as with t statistics, p-values can be calculated by looking up the percentile in the appropriate F distribution! In lab, you ll learn how to do this in EViews.! If only one exclusion is being tested, then F = t 2, and the p-values will be the same 42
Multiple Regression Analysis
Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,
More informationInference in Regression Analysis
ECNS 561 Inference Inference in Regression Analysis Up to this point 1.) OLS is unbiased 2.) OLS is BLUE (best linear unbiased estimator i.e., the variance is smallest among linear unbiased estimators)
More informationCHAPTER 4. > 0, where β
CHAPTER 4 SOLUTIONS TO PROBLEMS 4. (i) and (iii) generally cause the t statistics not to have a t distribution under H. Homoskedasticity is one of the CLM assumptions. An important omitted variable violates
More informationInference in Regression Model
Inference in Regression Model Christopher Taber Department of Economics University of Wisconsin-Madison March 25, 2009 Outline 1 Final Step of Classical Linear Regression Model 2 Confidence Intervals 3
More informationcoefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1
Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationMultiple Regression: Inference
Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled
More informationLECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit
LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define
More informationLectures 5 & 6: Hypothesis Testing
Lectures 5 & 6: Hypothesis Testing in which you learn to apply the concept of statistical significance to OLS estimates, learn the concept of t values, how to use them in regression work and come across
More informationProblem 4.1. Problem 4.3
BOSTON COLLEGE Department of Economics EC 228 01 Econometric Methods Fall 2008, Prof. Baum, Ms. Phillips (tutor), Mr. Dmitriev (grader) Problem Set 3 Due at classtime, Thursday 14 Oct 2008 Problem 4.1
More informationMultiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators
1 2 Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE Hüseyin Taştan 1 1 Yıldız Technical University Department of Economics These presentation notes are based on Introductory
More informationMultiple Regression Analysis: Heteroskedasticity
Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing
More informationStatistical Inference. Part IV. Statistical Inference
Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided
More informationLinear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons
Linear Regression with 1 Regressor Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor 1. The regression equation 2. Estimating the equation 3. Assumptions required for
More informationTHE MULTIVARIATE LINEAR REGRESSION MODEL
THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus
More informationEconometrics Multiple Regression Analysis: Heteroskedasticity
Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More information1 Motivation for Instrumental Variable (IV) Regression
ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data
More informationMultiple Linear Regression
Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random
More informationInferences for Regression
Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In
More informationMotivation for multiple regression
Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope
More informationBias Variance Trade-off
Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]
More informationCHAPTER 6: SPECIFICATION VARIABLES
Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero
More informationF Tests and F statistics
F Tests and F statistics Testing Linear estrictions F Stats and F Tests F Distributions F stats (w/ ) F Stats and tstat s eported F Stat's in OLS Output Example I: Bodyfat Babies and Bathwater F Stats,
More informationECO220Y Simple Regression: Testing the Slope
ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x
More informationRegression with a Single Regressor: Hypothesis Tests and Confidence Intervals
Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals (SW Chapter 5) Outline. The standard error of ˆ. Hypothesis tests concerning β 3. Confidence intervals for β 4. Regression
More informationThe t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies
The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit
More informationECNS 561 Multiple Regression Analysis
ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking
More informationLECTURE 5. Introduction to Econometrics. Hypothesis testing
LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will
More informationMultiple Linear Regression CIVL 7012/8012
Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for
More informationThe general linear regression with k explanatory variables is just an extension of the simple regression as follows
3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because
More informationLecture 5: Hypothesis testing with the classical linear model
Lecture 5: Hypothesis testing with the classical linear model Assumption MLR6: Normality MLR6 is not one of the Gauss-Markov assumptions. It s not necessary to assume the error is normally distributed
More informationIntroduction to Econometrics. Multiple Regression (2016/2017)
Introduction to Econometrics STAT-S-301 Multiple Regression (016/017) Lecturer: Yves Dominicy Teaching Assistant: Elise Petit 1 OLS estimate of the TS/STR relation: OLS estimate of the Test Score/STR relation:
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationChapter 3 Multiple Regression Complete Example
Department of Quantitative Methods & Information Systems ECON 504 Chapter 3 Multiple Regression Complete Example Spring 2013 Dr. Mohammad Zainal Review Goals After completing this lecture, you should be
More informationMultiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =
Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =
More informationThe Simple Regression Model. Simple Regression Model 1
The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising
More informationChapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression
Chapter 7 Hypothesis Tests and Confidence Intervals in Multiple Regression Outline 1. Hypothesis tests and confidence intervals for a single coefficie. Joint hypothesis tests on multiple coefficients 3.
More informationStatistical Inference with Regression Analysis
Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing
More informationSTAT Chapter 8: Hypothesis Tests
STAT 515 -- Chapter 8: Hypothesis Tests CIs are possibly the most useful forms of inference because they give a range of reasonable values for a parameter. But sometimes we want to know whether one particular
More informationSTA121: Applied Regression Analysis
STA121: Applied Regression Analysis Linear Regression Analysis - Chapters 3 and 4 in Dielman Artin Department of Statistical Science September 15, 2009 Outline 1 Simple Linear Regression Analysis 2 Using
More informationHypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima
Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationAnswer Key: Problem Set 6
: Problem Set 6 1. Consider a linear model to explain monthly beer consumption: beer = + inc + price + educ + female + u 0 1 3 4 E ( u inc, price, educ, female ) = 0 ( u inc price educ female) σ inc var,,,
More information2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0
Introduction to Econometrics Midterm April 26, 2011 Name Student ID MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. (5,000 credit for each correct
More informationLecture 3: Inference in SLR
Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals
More informationOne sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:
One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get
More informationBusiness Statistics. Lecture 10: Course Review
Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,
More informationIntro to Linear Regression
Intro to Linear Regression Introduction to Regression Regression is a statistical procedure for modeling the relationship among variables to predict the value of a dependent variable from one or more predictor
More informationChapter 27 Summary Inferences for Regression
Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test
More informationLecture 4: Testing Stuff
Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted
More informationLecture 10 Multiple Linear Regression
Lecture 10 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 10-1 Topic Overview Multiple Linear Regression Model 10-2 Data for Multiple Regression Y i is the response variable
More informationIntro to Linear Regression
Intro to Linear Regression Introduction to Regression Regression is a statistical procedure for modeling the relationship among variables to predict the value of a dependent variable from one or more predictor
More informationReview of Statistics 101
Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationIntroduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs
Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique
More informationLab 11 - Heteroskedasticity
Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction
More informationDistribution-Free Procedures (Devore Chapter Fifteen)
Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationMultiple Linear Regression
Chapter 3 Multiple Linear Regression 3.1 Introduction Multiple linear regression is in some ways a relatively straightforward extension of simple linear regression that allows for more than one independent
More informationChapter 8 Handout: Interval Estimates and Hypothesis Testing
Chapter 8 Handout: Interval Estimates and Hypothesis esting Preview Clint s Assignment: aking Stock General Properties of the Ordinary Least Squares (OLS) Estimation Procedure Estimate Reliability: Interval
More informationCIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8
CIVL - 7904/8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 Chi-square Test How to determine the interval from a continuous distribution I = Range 1 + 3.322(logN) I-> Range of the class interval
More informationUsing SPSS for One Way Analysis of Variance
Using SPSS for One Way Analysis of Variance This tutorial will show you how to use SPSS version 12 to perform a one-way, between- subjects analysis of variance and related post-hoc tests. This tutorial
More information2. Linear regression with multiple regressors
2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions
More informationECON3150/4150 Spring 2015
ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2
More informationappstats27.notebook April 06, 2017
Chapter 27 Objective Students will conduct inference on regression and analyze data to write a conclusion. Inferences for Regression An Example: Body Fat and Waist Size pg 634 Our chapter example revolves
More informationChapter 11 Handout: Hypothesis Testing and the Wald Test
Chapter 11 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic Manipulation o Wald Test Restricted Regression Reflects Unrestricted Regression
More informationRecall that a measure of fit is the sum of squared residuals: where. The F-test statistic may be written as:
1 Joint hypotheses The null and alternative hypotheses can usually be interpreted as a restricted model ( ) and an model ( ). In our example: Note that if the model fits significantly better than the restricted
More informationIntroduction to Econometrics. Multiple Regression
Introduction to Econometrics The statistical analysis of economic (and related) data STATS301 Multiple Regression Titulaire: Christopher Bruffaerts Assistant: Lorenzo Ricci 1 OLS estimate of the TS/STR
More informationMultiple Regression Analysis
Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple
More informationChapter 1. An Overview of Regression Analysis. Econometrics and Quantitative Analysis. What is Econometrics? (cont.) What is Econometrics?
Econometrics and Quantitative Analysis Using Econometrics: A Practical Guide A.H. Studenmund 6th Edition. Addison Wesley Longman Chapter 1 An Overview of Regression Analysis Instructor: Dr. Samir Safi
More informationEconomics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Fall 2016 Instructor: Martin Farnham
Economics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Fall 2016 Instructor: Martin Farnham Last name (family name): First name (given name):
More informationLecture 2 Multiple Regression and Tests
Lecture 2 and s Dr.ssa Rossella Iraci Capuccinello 2017-18 Simple Regression Model The random variable of interest, y, depends on a single factor, x 1i, and this is an exogenous variable. The true but
More informationEconometrics Review questions for exam
Econometrics Review questions for exam Nathaniel Higgins nhiggins@jhu.edu, 1. Suppose you have a model: y = β 0 x 1 + u You propose the model above and then estimate the model using OLS to obtain: ŷ =
More information1 Independent Practice: Hypothesis tests for one parameter:
1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)
More informationHint: The following equation converts Celsius to Fahrenheit: F = C where C = degrees Celsius F = degrees Fahrenheit
Amherst College Department of Economics Economics 360 Fall 2014 Exam 1: Solutions 1. (10 points) The following table in reports the summary statistics for high and low temperatures in Key West, FL from
More informationBusiness Statistics. Lecture 9: Simple Regression
Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals
More informationECO375 Tutorial 4 Introduction to Statistical Inference
ECO375 Tutorial 4 Introduction to Statistical Inference Matt Tudball University of Toronto Mississauga October 19, 2017 Matt Tudball (University of Toronto) ECO375H5 October 19, 2017 1 / 26 Statistical
More informationWednesday, October 17 Handout: Hypothesis Testing and the Wald Test
Amherst College Department of Economics Economics 360 Fall 2012 Wednesday, October 17 Handout: Hypothesis Testing and the Wald Test Preview No Money Illusion Theory: Calculating True] o Clever Algebraic
More informationSolutions to Problem Set 4 (Due November 13) Maximum number of points for Problem set 4 is: 66. Problem C 6.1
Solutions to Problem Set 4 (Due November 13) EC 228 01, Fall 2013 Prof. Baum, Mr. Lim Maximum number of points for Problem set 4 is: 66 Problem C 6.1 (i) (3 pts.) If the presence of the incinerator depresses
More informationCHAPTER 9: HYPOTHESIS TESTING
CHAPTER 9: HYPOTHESIS TESTING THE SECOND LAST EXAMPLE CLEARLY ILLUSTRATES THAT THERE IS ONE IMPORTANT ISSUE WE NEED TO EXPLORE: IS THERE (IN OUR TWO SAMPLES) SUFFICIENT STATISTICAL EVIDENCE TO CONCLUDE
More informationSTA Module 10 Comparing Two Proportions
STA 2023 Module 10 Comparing Two Proportions Learning Objectives Upon completing this module, you should be able to: 1. Perform large-sample inferences (hypothesis test and confidence intervals) to compare
More informationσ σ MLR Models: Estimation and Inference v.3 SLR.1: Linear Model MLR.1: Linear Model Those (S/M)LR Assumptions MLR3: No perfect collinearity
Comparison of SLR and MLR analysis: What s New? Roadmap Multicollinearity and standard errors F Tests of linear restrictions F stats, adjusted R-squared, RMSE and t stats Playing with Bodyfat: F tests
More informationLecture 3: Multiple Regression
Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u
More informationHypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =
Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,
More informationCorrelation and regression
NST 1B Experimental Psychology Statistics practical 1 Correlation and regression Rudolf Cardinal & Mike Aitken 11 / 12 November 2003 Department of Experimental Psychology University of Cambridge Handouts:
More informationLAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2
LAB 2. HYPOTHESIS TESTING IN THE BIOLOGICAL SCIENCES- Part 2 Data Analysis: The mean egg masses (g) of the two different types of eggs may be exactly the same, in which case you may be tempted to accept
More informationSimple Linear Regression: One Quantitative IV
Simple Linear Regression: One Quantitative IV Linear regression is frequently used to explain variation observed in a dependent variable (DV) with theoretically linked independent variables (IV). For example,
More informationPOL 681 Lecture Notes: Statistical Interactions
POL 681 Lecture Notes: Statistical Interactions 1 Preliminaries To this point, the linear models we have considered have all been interpreted in terms of additive relationships. That is, the relationship
More informationChapter 8 Heteroskedasticity
Chapter 8 Walter R. Paczkowski Rutgers University Page 1 Chapter Contents 8.1 The Nature of 8. Detecting 8.3 -Consistent Standard Errors 8.4 Generalized Least Squares: Known Form of Variance 8.5 Generalized
More informationLECTURE 15: SIMPLE LINEAR REGRESSION I
David Youngberg BSAD 20 Montgomery College LECTURE 5: SIMPLE LINEAR REGRESSION I I. From Correlation to Regression a. Recall last class when we discussed two basic types of correlation (positive and negative).
More informationPractice Problems Section Problems
Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,
More informationFinal Exam - Solutions
Ecn 102 - Analysis of Economic Data University of California - Davis March 19, 2010 Instructor: John Parman Final Exam - Solutions You have until 5:30pm to complete this exam. Please remember to put your
More informationMeasuring the fit of the model - SSR
Measuring the fit of the model - SSR Once we ve determined our estimated regression line, we d like to know how well the model fits. How far/close are the observations to the fitted line? One way to do
More informationHypothesis Tests and Confidence Intervals in Multiple Regression
Hypothesis Tests and Confidence Intervals in Multiple Regression (SW Chapter 7) Outline 1. Hypothesis tests and confidence intervals for one coefficient. Joint hypothesis tests on multiple coefficients
More informationChapter 1 Review of Equations and Inequalities
Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve
More informationSteps in Regression Analysis
MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4) 2-1 Steps in Regression Analysis 1. Review the literature and develop the theoretical model 2. Specify the model:
More informationMultiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD
Multiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD Introduction When you perform statistical inference, you are primarily doing one of two things: Estimating the boundaries
More informationProblem Set 4 ANSWERS
Economics 20 Problem Set 4 ANSWERS Prof. Patricia M. Anderson 1. Suppose that our variable for consumption is measured with error, so cons = consumption + e 0, where e 0 is uncorrelated with inc, educ
More information