ECON 497: Lecture 4 Page 1 of 1
|
|
- Tracey Singleton
- 5 years ago
- Views:
Transcription
1 ECON 497: Lecture 4 Page 1 of 1 Metropolitan State University ECON 497: Research and Forecasting Lecture Notes 4 The Classical Model: Assumptions and Violations Studenmund Chapter 4 Ordinary least squares (OLS) is a very good way of estimating a linear relationship between a dependent variable and some independent or explanatory variables. In some sense, it is the best method of doing this. For it to be "best", however, there are seven assumptions that must be satisfied. These assumptions are somewhat technical, but here is an attempt to explain what each means and how it affects regression results. I. The regression model is linear in the coefficients, is correctly specified and has an additive error term. This assumption has three parts. Let's look at them each in turn. First, the model must be linear in the coefficients. This means that the process that is actually occurring in the real world is described by a relationship of the form Yi = β 0 + β 1 X 1i + β 2 X 2i +β 3 X 3i ε i or that the actual relationship can be rewritten in this form by, for example, taking logs. This part is best considered in conjunction with the second. The second is that the model is correctly specified. Combining these first two assumptions, we see that we need to know the actual process through which the dependent value is determined and that relationship has to be linear or some derivation of linear. While this may be possible for simple physical processes, it is virtually impossible that the decision making process of any rational person operating in a complex environment can be correctly modeled by a linear equation. What must be believed instead is that the model is sufficiently close to the actual process that the difference isn't important. In a practical sense, you can defend a specification by saying that it is a standard specification used in looking at the situation being examined. If you find other studies, papers or reports that have used a particular model, that may be an acceptable reason for using it. The opposite also holds. On the other hand, if you use a model which no one else has used, the results may argue for rejection of other models in favor of yours. The third part of this is that the error term is additive rather than being multiplicative or entering in some other way. This third part of the assumption is probably no more suspicious that the first two parts (that you have the correct model and that it is linear) but by looking at the residuals (the errors that you get after you estimate your model, also know as the e i s) it can be
2 ECON 497: Lecture 4 Page 2 of 2 determined whether or not this may be true. If the model is not correct, we may have problems such as those we've seen in class where the correct model is, say, curved but the estimated model is a straight line. II. The error term has a zero population mean. This means that the expected value of ε i is zero (E[ε i ]=0). There is nothing earthshaking about this. When you do your regression, your residuals will, basically by construction, have a mean value of zero. This is a practical matter and is only loosely related to the theoretical presentation above. Basically, the estimated model will have an error term with a mean value of zero, so if the theoretical model doesn t have an error term with an expected value of zero, the two versions will be inconsistent. III. All explanatory variables are uncorrelated with the error term (no endogeneity) This means that the error term is not likely to be larger or smaller, positive or negative when any of the explanatory variables are larger or smaller. If this were not true, then you might know, for example, that the error term is larger when one of the explanatory variables is larger and smaller when the explanatory variable is smaller. If this were true, the model could be improved based on the value of that explanatory variable. Whether or not this condition is satisfied can be investigated in at least two ways. A. The residuals (the differences between the actual value of the dependent variable and the predicted values) can be graphed against the various explanatory variables. B. Correlation coefficients between the residuals and the various explanatory variables can be calculated. There should be no discernable patterns in the graphs and only very small correlation coefficients. An Example of Endogeneity According to Prof. Lundberg, "Endogeneity is a problem when one of the right hand side variables is correlated with the error term because it is being determined as part of the whole behavioral system that this regression equation is part of. So, if we're trying to explain hours of TV watching, putting the number of sets in the household on the righthand side is a no-no. Both hours and sets will be driven by taste for TV watching, and the coefficient on number of sets will be meaningless behaviorally (though probably big and significant). So, endogeneity is a specification problem, and needs to be dealt with by estimating reduced-form models with only exogenous variables on the right, IV, or some simultaneous-equations method." Let's pursue this a bit. Imagine that there is some variable, Y i, which you are interested in. There are a number of explanatory variables that you want to include in the regression, X 1i, X 2i, X 3i, X 4i and X 5i. So, the equation you estimate is:
3 ECON 497: Lecture 4 Page 3 of 3 Y i = β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i + β 4 X 4i + β 5 X 5i + ε i However, if X 1i is edogenously determined, we will get a violation of the assumptions of the classical model, meaning that the happy results associated with OLS may not hold. For example, let's say that: X 1i = α0 + α 1 X 2i + α 2 X 3i + φ i where φ i is the error term. If this is how Y i is actually determined, then the correct model we should be estimating is: Y i = β 0 + β 1 (α 0 + α 1 X 2i + α 2 X 3i + φ i ) + β 2 X 2i + β 3 X 3i + β 4 X 4i + β 5 X 5i + ε i Rewriting this a bit gives us: Y i = β 0 + β 1 α 0 + (α 1 + β 2 )X 2i + (α 2 + β 3 )X 3i + β 4 X 4i + β 5 X 5i + ε i + φ i To summarize this equation, there is a constant term (β 0 + β 1 α 0 ) and a error term (ε i + φ i ) and coefficients attached to each of the remaining explanatory variables. If X 1i is included in this regression, it will be correlated with the error term because X 1i is a linear function of part of the error term, φ i. Because the error term is correlated with one of the explanatory variables, assumption three (III. All explanatory variables are uncorrelated with the error term.) is violated, so OLS won't work. Now, knowing that this may be a problem, what can or should be done about it? Kennedy, (chapter 10) describes different approaches to dealing with this problem and, if you're interested, I would be happy to share them with you. IV. Observations of the error term are uncorrelated with each other (no serial correlation). If you're looking at time series data (data collected from the same source in a number of different periods) the error term (e i = Y i - Ŷ i ) in one period should not have any relation to the error term from the previous period. A way to investigate this is to graph the error terms over time and see if there are any patterns or long runs of either positive or negative values. You may at some point see reference made to something called a runs test. This is basically a test to see if the number of simultaneous observations with either positive or negative residuals is suspiciously high. In looking at the California gasoline consumption data, there appears to be serial correlation for some models. V. The error term has a constant variance (no heteroskedasticity).
4 ECON 497: Lecture 4 Page 4 of 4 This means that the errors aren't more spread out for some of the observations than for others. It's tough to describe, but there's a good picture of it on Studenmund, p. 99. (in the third edition). The Studenmund picture shows a scatterplot with an explanatory variable on the horizontal axis, the dependent variable on the vertical axis and the regression line drawn in. The points of the scatterplot are further away from the regression line for larger values of the explanatory variable. Here s another picture Scaterplot of Squared Redsiduals on SQFT Residsq When the independent variable, SQFT, is smaller, the errors tend to be smaller. As SQFT increases, the errors get larger. This may be easier to see if you plot out the squared errors as is done in this picture. Basically, if you graph the squared residuals against all the explanatory variables, the size of the residuals shouldn't depend on the value of the explanatory variables. If the residuals get larger as the explanatory variable gets larger (or smaller) then you have heteroskedasticity. An example of a case in which heteroskedasticity might be a problem is in modeling house prices as a function of the house characteristics. There might be larger variance for the error term for more expensive houses and smaller variance for less expensive houses. A 95% confidence interval for the true value of a house with an estimated value of $40,000 might be [$38,000, $42,000] while the same interval for a house with an estimated value of $2,000,000 might be [$1,900,000, $2,100,000]. Kennedy (pp ) has a very good discussion about the consequences of heteroskedasticity, methods of testing for it and a somewhat vague description of how to correct for it. Kennedy offers four methods of detecting heteroskedasticity: 1. Visual inspection of the residuals
5 ECON 497: Lecture 4 Page 5 of 5 2. The Goldfeld-Quandt test 3. The Breusch-Pagan test 4. The White test The first of these is within your power to do in Excel. The others should be available options in any good software package. To deal with heteroskedasticity, there are really two options. 1. You could run a weighted least squares (rather than an ordinary least squares) regression. 2. You could opt for the zen approach and eliminate heteroskedasticity in a more spiritual way. VI. No explanatory variable is a perfect linear function of any other explanatory variable(s) (no perfect multicollinearity). This just means that no explanatory variable is a linear function of another explanatory variable or other explanatory variables. This means that you can include as explanatory variables X and X 2. You cannot include, for example, temperature in degrees Fahrenheit (F) and in degrees Celsius (C) because F = C. That is, Celsius is a linear function of Fahrenheit. This is why you must exclude one of the dummy variables if there is an exhaustive list of them. If, for example, you have dummy variables for male (M) and female (F) subjects and there are no other genders, then for each observation M + F = 1 or F = 1 - M or M = 1 - F. Because these two variables are linear functions of each other, one of them must be excluded. One way to see if this might be a problem is to generate a matrix of correlation coefficients for all the explanatory variables and the dependent variable. This won't tell you if a large number of variables are linearly related (such as dummy variables for a person's home state, for example) but it will tell you if two variables are linearly related. Alternatively, when you do a linear regression in SPSS, you can ask for Statistics/Collinearity diagnostics. In addition to all the wonderful things you usually get with your regression output, you ll get Variance Inflation Factors (VIF) for each explanatory variable. The larger this is, the greater the likelihood that you have multicollinearity. The VIF is calculated based on a regression of each explanatory variable on all the other explanatory variables, and is equal to 1/(1-R 2 ) from that regression. Kennedy (pp ) has a very good section on multicollinearity. A fun quote from this section:
6 ECON 497: Lecture 4 Page 6 of 6 The OLS estimator in the presence of multicollinearity remains unbiased and in fact is still BLUE. The R2 statistic is unaffected. In fact, since all the CLR assumptions are (strictly speaking) still met, the OLS estimator retains all its desirable properties, as noted in chapter 3. The major undesirable consequence of multicollinearity is that the variances of the OLS estimates of the parameters of the collinear variables are quite large. These high variances arise because in the presence of multicollinearity the OLS estimating procedure is not given enough independent variation in a variable to calculate with confidence the effect it has on the dependent variable. Possible rememdies are later suggested by Kennedy. VII. The error term is normally distributed This is important when generating confidence intervals and doing hypothesis testing in small samples but is less important as sample sizes increases. To see if your residuals are normally distributed, you can generate a histogram of them and see if they appear to be normal. To do this in SPSS you can do a regression, save the residuals and then use Graph/Histogram to make a histogram of the residuals see if they appear to be normally distributed. There are also statistical tests that you can do to see if the residuals are normally distributed. To be totally honest, most of the time if you do a statistical test to determine whether your residuals are normally distributed, the null hypothesis of normality will probably be rejected.
ECON 497 Midterm Spring
ECON 497 Midterm Spring 2009 1 ECON 497: Economic Research and Forecasting Name: Spring 2009 Bellas Midterm You have three hours and twenty minutes to complete this exam. Answer all questions and explain
More informationECON 497: Lecture Notes 10 Page 1 of 1
ECON 497: Lecture Notes 10 Page 1 of 1 Metropolitan State University ECON 497: Research and Forecasting Lecture Notes 10 Heteroskedasticity Studenmund Chapter 10 We'll start with a quote from Studenmund:
More informationLECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity
LECTURE 10 Introduction to Econometrics Multicollinearity & Heteroskedasticity November 22, 2016 1 / 23 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists
More informationEconometrics Part Three
!1 I. Heteroskedasticity A. Definition 1. The variance of the error term is correlated with one of the explanatory variables 2. Example -- the variance of actual spending around the consumption line increases
More informationLECTURE 11. Introduction to Econometrics. Autocorrelation
LECTURE 11 Introduction to Econometrics Autocorrelation November 29, 2016 1 / 24 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists of choosing: 1. correct
More informationKeller: Stats for Mgmt & Econ, 7th Ed July 17, 2006
Chapter 17 Simple Linear Regression and Correlation 17.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will
More informationRegression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.
Regression Analysis BUS 735: Business Decision Making and Research 1 Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn how to estimate
More informationChapter 16. Simple Linear Regression and Correlation
Chapter 16 Simple Linear Regression and Correlation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will
More informationRegression Analysis. BUS 735: Business Decision Making and Research
Regression Analysis BUS 735: Business Decision Making and Research 1 Goals and Agenda Goals of this section Specific goals Learn how to detect relationships between ordinal and categorical variables. Learn
More informationChapter 16. Simple Linear Regression and dcorrelation
Chapter 16 Simple Linear Regression and dcorrelation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will
More informationMaking sense of Econometrics: Basics
Making sense of Econometrics: Basics Lecture 4: Qualitative influences and Heteroskedasticity Egypt Scholars Economic Society November 1, 2014 Assignment & feedback enter classroom at http://b.socrative.com/login/student/
More informationRegression of Inflation on Percent M3 Change
ECON 497 Final Exam Page of ECON 497: Economic Research and Forecasting Name: Spring 2006 Bellas Final Exam Return this exam to me by midnight on Thursday, April 27. It may be e-mailed to me. It may be
More informationSemester 2, 2015/2016
ECN 3202 APPLIED ECONOMETRICS 5. HETEROSKEDASTICITY Mr. Sydney Armstrong Lecturer 1 The University of Guyana 1 Semester 2, 2015/2016 WHAT IS HETEROSKEDASTICITY? The multiple linear regression model can
More informationIris Wang.
Chapter 10: Multicollinearity Iris Wang iris.wang@kau.se Econometric problems Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences?
More informationAssumptions of the error term, assumptions of the independent variables
Petra Petrovics, Renáta Géczi-Papp Assumptions of the error term, assumptions of the independent variables 6 th seminar Multiple linear regression model Linear relationship between x 1, x 2,, x p and y
More information1 Motivation for Instrumental Variable (IV) Regression
ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data
More informationHypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima
Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s
More informationBusiness Statistics. Lecture 9: Simple Regression
Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals
More informationCHAPTER 6: SPECIFICATION VARIABLES
Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero
More informationPBAF 528 Week 8. B. Regression Residuals These properties have implications for the residuals of the regression.
PBAF 528 Week 8 What are some problems with our model? Regression models are used to represent relationships between a dependent variable and one or more predictors. In order to make inference from the
More informationEcon107 Applied Econometrics
Econ107 Applied Econometrics Topics 2-4: discussed under the classical Assumptions 1-6 (or 1-7 when normality is needed for finite-sample inference) Question: what if some of the classical assumptions
More informationChapter 3 Multiple Regression Complete Example
Department of Quantitative Methods & Information Systems ECON 504 Chapter 3 Multiple Regression Complete Example Spring 2013 Dr. Mohammad Zainal Review Goals After completing this lecture, you should be
More informationBasic Business Statistics 6 th Edition
Basic Business Statistics 6 th Edition Chapter 12 Simple Linear Regression Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value of a dependent variable based
More informationMaking sense of Econometrics: Basics
Making sense of Econometrics: Basics Lecture 7: Multicollinearity Egypt Scholars Economic Society November 22, 2014 Assignment & feedback Multicollinearity enter classroom at room name c28efb78 http://b.socrative.com/login/student/
More information2 Prediction and Analysis of Variance
2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering
More informationRockefeller College University at Albany
Rockefeller College University at Albany PAD 705 Handout: Suggested Review Problems from Pindyck & Rubinfeld Original prepared by Professor Suzanne Cooper John F. Kennedy School of Government, Harvard
More informationRegression Models. Chapter 4. Introduction. Introduction. Introduction
Chapter 4 Regression Models Quantitative Analysis for Management, Tenth Edition, by Render, Stair, and Hanna 008 Prentice-Hall, Inc. Introduction Regression analysis is a very valuable tool for a manager
More informationMULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS Page 1 MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level
More informationRegression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur
Regression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur Lecture 10 Software Implementation in Simple Linear Regression Model using
More informationSteps in Regression Analysis
MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4) 2-1 Steps in Regression Analysis 1. Review the literature and develop the theoretical model 2. Specify the model:
More informationChapter 13. Multiple Regression and Model Building
Chapter 13 Multiple Regression and Model Building Multiple Regression Models The General Multiple Regression Model y x x x 0 1 1 2 2... k k y is the dependent variable x, x,..., x 1 2 k the model are the
More informationContest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.
Updated: November 17, 2011 Lecturer: Thilo Klein Contact: tk375@cam.ac.uk Contest Quiz 3 Question Sheet In this quiz we will review concepts of linear regression covered in lecture 2. NOTE: Please round
More informationEcon 510 B. Brown Spring 2014 Final Exam Answers
Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity
More informationIntroduction to Econometrics. Heteroskedasticity
Introduction to Econometrics Introduction Heteroskedasticity When the variance of the errors changes across segments of the population, where the segments are determined by different values for the explanatory
More informationCorrelation Analysis
Simple Regression Correlation Analysis Correlation analysis is used to measure strength of the association (linear relationship) between two variables Correlation is only concerned with strength of the
More informationChapter 14 Student Lecture Notes 14-1
Chapter 14 Student Lecture Notes 14-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter 14 Multiple Regression Analysis and Model Building Chap 14-1 Chapter Goals After completing this
More informationSection 3: Simple Linear Regression
Section 3: Simple Linear Regression Carlos M. Carvalho The University of Texas at Austin McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction
More informationInferences for Regression
Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In
More informationApplied Quantitative Methods II
Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator
More informationSingle and multiple linear regression analysis
Single and multiple linear regression analysis Marike Cockeran 2017 Introduction Outline of the session Simple linear regression analysis SPSS example of simple linear regression analysis Additional topics
More informationReview of Multiple Regression
Ronald H. Heck 1 Let s begin with a little review of multiple regression this week. Linear models [e.g., correlation, t-tests, analysis of variance (ANOVA), multiple regression, path analysis, multivariate
More informationA particularly nasty aspect of this is that it is often difficult or impossible to tell if a model fails to satisfy these steps.
ECON 497: Lecture 6 Page 1 of 1 Metropolitan State University ECON 497: Research and Forecasting Lecture Notes 6 Specification: Choosing the Independent Variables Studenmund Chapter 6 Before we start,
More informationFinQuiz Notes
Reading 10 Multiple Regression and Issues in Regression Analysis 2. MULTIPLE LINEAR REGRESSION Multiple linear regression is a method used to model the linear relationship between a dependent variable
More information:Effects of Data Scaling We ve already looked at the effects of data scaling on the OLS statistics, 2, and R 2. What about test statistics?
MRA: Further Issues :Effects of Data Scaling We ve already looked at the effects of data scaling on the OLS statistics, 2, and R 2. What about test statistics? 1. Scaling the explanatory variables Suppose
More informationFinal Exam - Solutions
Ecn 102 - Analysis of Economic Data University of California - Davis March 19, 2010 Instructor: John Parman Final Exam - Solutions You have until 5:30pm to complete this exam. Please remember to put your
More informationECON 497 Final Exam Page 1 of 12
ECON 497 Final Exam Page of 2 ECON 497: Economic Research and Forecasting Name: Spring 2008 Bellas Final Exam Return this exam to me by 4:00 on Wednesday, April 23. It may be e-mailed to me. It may be
More informationLab 11 - Heteroskedasticity
Lab 11 - Heteroskedasticity Spring 2017 Contents 1 Introduction 2 2 Heteroskedasticity 2 3 Addressing heteroskedasticity in Stata 3 4 Testing for heteroskedasticity 4 5 A simple example 5 1 1 Introduction
More informationMidterm 2 - Solutions
Ecn 102 - Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman Midterm 2 - Solutions You have until 10:20am to complete this exam. Please remember to put
More informationSIMPLE REGRESSION ANALYSIS. Business Statistics
SIMPLE REGRESSION ANALYSIS Business Statistics CONTENTS Ordinary least squares (recap for some) Statistical formulation of the regression model Assessing the regression model Testing the regression coefficients
More informationSTOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Friday, June 5, 009 Examination time: 3 hours
More informationBinary Logistic Regression
The coefficients of the multiple regression model are estimated using sample data with k independent variables Estimated (or predicted) value of Y Estimated intercept Estimated slope coefficients Ŷ = b
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationChapter 8 Heteroskedasticity
Chapter 8 Walter R. Paczkowski Rutgers University Page 1 Chapter Contents 8.1 The Nature of 8. Detecting 8.3 -Consistent Standard Errors 8.4 Generalized Least Squares: Known Form of Variance 8.5 Generalized
More informationStatistics for Managers using Microsoft Excel 6 th Edition
Statistics for Managers using Microsoft Excel 6 th Edition Chapter 13 Simple Linear Regression 13-1 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value of
More informationChapter 4. Regression Models. Learning Objectives
Chapter 4 Regression Models To accompany Quantitative Analysis for Management, Eleventh Edition, by Render, Stair, and Hanna Power Point slides created by Brian Peterson Learning Objectives After completing
More informationL7: Multicollinearity
L7: Multicollinearity Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Introduction ï Example Whats wrong with it? Assume we have this data Y
More informationRef.: Spring SOS3003 Applied data analysis for social science Lecture note
SOS3003 Applied data analysis for social science Lecture note 05-2010 Erling Berge Department of sociology and political science NTNU Spring 2010 Erling Berge 2010 1 Literature Regression criticism I Hamilton
More informationMultiple linear regression S6
Basic medical statistics for clinical and experimental research Multiple linear regression S6 Katarzyna Jóźwiak k.jozwiak@nki.nl November 15, 2017 1/42 Introduction Two main motivations for doing multiple
More informationChapter 4: Regression Models
Sales volume of company 1 Textbook: pp. 129-164 Chapter 4: Regression Models Money spent on advertising 2 Learning Objectives After completing this chapter, students will be able to: Identify variables,
More informationACE 564 Spring Lecture 8. Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information. by Professor Scott H.
ACE 564 Spring 2006 Lecture 8 Violations of Basic Assumptions I: Multicollinearity and Non-Sample Information by Professor Scott H. Irwin Readings: Griffiths, Hill and Judge. "Collinear Economic Variables,
More informationappstats27.notebook April 06, 2017
Chapter 27 Objective Students will conduct inference on regression and analyze data to write a conclusion. Inferences for Regression An Example: Body Fat and Waist Size pg 634 Our chapter example revolves
More informationNotes 11: OLS Theorems ECO 231W - Undergraduate Econometrics
Notes 11: OLS Theorems ECO 231W - Undergraduate Econometrics Prof. Carolina Caetano For a while we talked about the regression method. Then we talked about the linear model. There were many details, but
More informationLECTURE 5. Introduction to Econometrics. Hypothesis testing
LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will
More informationChapter 19: Logistic regression
Chapter 19: Logistic regression Self-test answers SELF-TEST Rerun this analysis using a stepwise method (Forward: LR) entry method of analysis. The main analysis To open the main Logistic Regression dialog
More informationRegression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.
Regression, part II I. What does it all mean? A) Notice that so far all we ve done is math. 1) One can calculate the Least Squares Regression Line for anything, regardless of any assumptions. 2) But, if
More informationEconomics 300. Econometrics Multiple Regression: Extensions and Issues
Economics 300 Econometrics Multiple : Extensions and Dennis C. Plott University of Illinois at Chicago Department of Economics www.dennisplott.com Fall 2014 Dennis C. Plott (UIC) ECON 300 Fall 2014 1 /
More informationLECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit
LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationDiagnostics of Linear Regression
Diagnostics of Linear Regression Junhui Qian October 7, 14 The Objectives After estimating a model, we should always perform diagnostics on the model. In particular, we should check whether the assumptions
More informationEconomics 300. Econometrics Multiple Regression: Extensions and Issues
Economics 300 Econometrics Multiple : Extensions and Dennis C. Plott University of Illinois at Chicago Department of Economics www.dennisplott.com Fall 2014 Dennis C. Plott (UIC) ECON 300 Fall 2014 1 /
More informationLecture 4: Heteroskedasticity
Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan
More informationECNS 561 Multiple Regression Analysis
ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking
More informationRegression, Part I. - In correlation, it would be irrelevant if we changed the axes on our graph.
Regression, Part I I. Difference from correlation. II. Basic idea: A) Correlation describes the relationship between two variables, where neither is independent or a predictor. - In correlation, it would
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
What is Multiple Linear Regression Several independent variables may influence the change in response variable we are trying to study. When several independent variables are included in the equation, the
More informationModelling the Electric Power Consumption in Germany
Modelling the Electric Power Consumption in Germany Cerasela Măgură Agricultural Food and Resource Economics (Master students) Rheinische Friedrich-Wilhelms-Universität Bonn cerasela.magura@gmail.com Codruța
More informationThe Simple Linear Regression Model
The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate
More informationAn overview of applied econometrics
An overview of applied econometrics Jo Thori Lind September 4, 2011 1 Introduction This note is intended as a brief overview of what is necessary to read and understand journal articles with empirical
More informationy response variable x 1, x 2,, x k -- a set of explanatory variables
11. Multiple Regression and Correlation y response variable x 1, x 2,, x k -- a set of explanatory variables In this chapter, all variables are assumed to be quantitative. Chapters 12-14 show how to incorporate
More informationImmigration attitudes (opposes immigration or supports it) it may seriously misestimate the magnitude of the effects of IVs
Logistic Regression, Part I: Problems with the Linear Probability Model (LPM) Richard Williams, University of Notre Dame, https://www3.nd.edu/~rwilliam/ Last revised February 22, 2015 This handout steals
More informationFinal Exam - Solutions
Ecn 102 - Analysis of Economic Data University of California - Davis March 17, 2010 Instructor: John Parman Final Exam - Solutions You have until 12:30pm to complete this exam. Please remember to put your
More information2. Linear regression with multiple regressors
2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More informationappstats8.notebook October 11, 2016
Chapter 8 Linear Regression Objective: Students will construct and analyze a linear model for a given set of data. Fat Versus Protein: An Example pg 168 The following is a scatterplot of total fat versus
More informationLinear Regression with Multiple Regressors
Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution
More informationBig Data Analysis with Apache Spark UC#BERKELEY
Big Data Analysis with Apache Spark UC#BERKELEY This Lecture: Relation between Variables An association A trend» Positive association or Negative association A pattern» Could be any discernible shape»
More informationBasic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler
Basic econometrics Tutorial 3 Dipl.Kfm. Introduction Some of you were asking about material to revise/prepare econometrics fundamentals. First of all, be aware that I will not be too technical, only as
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationChapter 27 Summary Inferences for Regression
Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test
More informationEconomics 308: Econometrics Professor Moody
Economics 308: Econometrics Professor Moody References on reserve: Text Moody, Basic Econometrics with Stata (BES) Pindyck and Rubinfeld, Econometric Models and Economic Forecasts (PR) Wooldridge, Jeffrey
More informationGov 2000: 9. Regression with Two Independent Variables
Gov 2000: 9. Regression with Two Independent Variables Matthew Blackwell Harvard University mblackwell@gov.harvard.edu Where are we? Where are we going? Last week: we learned about how to calculate a simple
More informationRegression Analysis: Exploring relationships between variables. Stat 251
Regression Analysis: Exploring relationships between variables Stat 251 Introduction Objective of regression analysis is to explore the relationship between two (or more) variables so that information
More informationMATH 1150 Chapter 2 Notation and Terminology
MATH 1150 Chapter 2 Notation and Terminology Categorical Data The following is a dataset for 30 randomly selected adults in the U.S., showing the values of two categorical variables: whether or not the
More informationMultiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =
Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =
More informationLecture 5: Omitted Variables, Dummy Variables and Multicollinearity
Lecture 5: Omitted Variables, Dummy Variables and Multicollinearity R.G. Pierse 1 Omitted Variables Suppose that the true model is Y i β 1 + β X i + β 3 X 3i + u i, i 1,, n (1.1) where β 3 0 but that the
More informationLinear Models in Econometrics
Linear Models in Econometrics Nicky Grant At the most fundamental level econometrics is the development of statistical techniques suited primarily to answering economic questions and testing economic theories.
More informationOrdinary Least Squares Regression Explained: Vartanian
Ordinary Least Squares Regression Explained: Vartanian When to Use Ordinary Least Squares Regression Analysis A. Variable types. When you have an interval/ratio scale dependent variable.. When your independent
More informationPractice exam questions
Practice exam questions Nathaniel Higgins nhiggins@jhu.edu, nhiggins@ers.usda.gov 1. The following question is based on the model y = β 0 + β 1 x 1 + β 2 x 2 + β 3 x 3 + u. Discuss the following two hypotheses.
More information1 Correlation and Inference from Regression
1 Correlation and Inference from Regression Reading: Kennedy (1998) A Guide to Econometrics, Chapters 4 and 6 Maddala, G.S. (1992) Introduction to Econometrics p. 170-177 Moore and McCabe, chapter 12 is
More informationPOLI 618 Notes. Stuart Soroka, Department of Political Science, McGill University. March 2010
POLI 618 Notes Stuart Soroka, Department of Political Science, McGill University March 2010 These pages were written originally as my own lecture notes, but are now designed to be distributed to students
More information