2. Linear regression with multiple regressors
|
|
- Jemimah Sherman
- 5 years ago
- Views:
Transcription
1 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions in the multiple regression model Violations of the assumptions (omitted-variable bias, multicollinearity, heteroskedasticity, autocorrelation) 5
2 2.1. The multiple regression model Intuition: A regression model specifies a functional (parametric) relationship between a dependent (endogenous) variable Y and a set of k independent (exogenous) regressors X 1, X 2,..., X k In a first step, we consider the linear multiple regression model 6
3 Definition 2.1: (Multiple linear regression model) The multiple (linear) regression model is given by Y i = β 0 + β 1 X 1i + β 2 X 2i β k X ki + u i, (2.1) i = 1,..., n, where Y i is the i th observation on the dependent variable, X 1i, X 2i,..., X ki are the i th regressors, u i is the stochastic error term. observations on each of the k The population regression line is the relationship that holds between Y and the X s on average: E(Y i X 1i = x 1, X 2i = x 2,..., X ki = x k ) = β 0 +β 1 x β k x k. 7
4 Meaning of the coefficients: The intercept β 0 is the expected value of Y i (for all i = 1,..., n) when all X-regressors equal 0 β 1,..., β k are the slope coefficients on the respective regressors X 1,..., X k β 1, for example, is the expected change in Y i resulting from changing X 1i by one unit, holding constant X 2i,..., X ki (and analogously β 2,..., β k ) Definition 2.2: (Homoskedasticity, Heteroskedasticity) The error term u i is called homoskedastic if the conditional variance of u i given X 1i,..., X ki, Var(u i X 1i,..., X ki ), is constant for i = 1,..., n and does not depend on the values of X 1i,..., X ki. Otherwise, the error term is called heteroskedastic. 8
5 Example 1: (Student performance) Regression of student performance (Y ) in n = 420 USdistricts on distinct school characteristics (factors) Y i : average test score in the i th district (TEST SCORE) X 1i : average class size in the i th district (measured by the student-teacher ratio, STR) X 2i : percentage of English learners in the i th district (PCTEL) Expected signs of the coefficients: β 1 < 0 β 2 < 0 9
6 Example 2: (House prices) Regression of house prices (Y ) recorded for n = 546 houses sold in Windsor (Canada) on distinct housing characteristics Y i : sale price (in Canadian dollars) of the i th house (SALEPRICE) X 1i : lot size (in square feet) of the i th property (LOTSIZE) X 2i : number of bedrooms in the i th house (BEDROOMS) X 3i : number of bathrooms in the i th house (BATHROOMS) X 4i : number of storeys (excluding the basement) in the i th house (STOREYS) Expected signs of the coefficients: β 1, β 2, β 3, β 4 > 0 10
7 2.2. The OLS estimator in multiple regression Now: Estimation of the coefficients β 0, β 1,..., β k in the multiple regression model on the basis of n observations by applying the Ordinary Least Squares (OLS) technique Idea: Let b 0, b 1,..., b k be estimators of β 0, β 1,..., β k We can predict Y i by b 0 + b 1 X 1i b k X ki The prediction error is Y i b 0 b 1 X 1i... b k X ki 11
8 Idea: [continued] The sum of the squared prediction errors over all n observations is n i=1 (Y i b 0 b 1 X 1i... b k X ki ) 2 (2.2) Definition 2.3: (OLS estimators, predicted values, residuals) The OLS estimators ˆβ 0, ˆβ 1,..., ˆβ k are the values of b 0, b 1,..., b k that minimize the sum of squared prediction errors (2.2). The OLS predicted values Ŷ i and residuals û i (for i = 1,..., n) are and Ŷ i = ˆβ 0 + ˆβ 1 X 1i ˆβ k X ki (2.3) û i = Y i Ŷ i. (2.4) 12
9 Remarks: The OLS estimators ˆβ 0, ˆβ 1,..., ˆβ k and the residuals û i are computed from a sample of n observations of (X 1i,..., X ki, Y i ) for i = 1,..., n They are estimators of the unknown true population coefficients β 0, β 1,..., β k and u i There are closed-form formulas for calculating the OLS estimates from the data (see the lectures Econometrics I+II) In this lecture, we use the software-package EViews 13
10 Regression estimation results (EViews) for the student-performance dataset Dependent Variable: TEST_SCORE Method: Least Squares Date: 07/02/12 Time: 16:29 Sample: Included observations: 420 Variable Coefficient Std. Error t-statistic Prob. C STR PCTEL R-squared Mean dependent var Adjusted R-squared S.D. dependent var S.E. of regression Akaike info criterion Sum squared resid Schwarz criterion Log likelihood Hannan-Quinn criter F-statistic Durbin-Watson stat Prob(F-statistic)
11 Predicted values Ŷ i and residuals û i for the student-performance dataset Residual Actual Fitted 15
12 Regression estimation results (EViews) for the house-prices dataset Dependent Variable: SALEPRICE Method: Least Squares Date: 07/02/12 Time: 16:50 Sample: Included observations: 546 Variable Coefficient Std. Error t-statistic Prob. C LOTSIZE BEDROOMS BATHROOMS STOREYS R-squared Mean dependent var Adjusted R-squared S.D. dependent var S.E. of regression Akaike info criterion Sum squared resid 1.80E+11 Schwarz criterion Log likelihood Hannan-Quinn criter F-statistic Durbin-Watson stat Prob(F-statistic)
13 Predicted values Ŷ i and residuals û i for the house-prices dataset 200, , , ,000 80,000 40,000 80,000 40, ,000-80, Residual Actual Fitted 17
14 OLS assumptions in the multiple regression model (2.1): 1. u i has conditional mean zero given X 1i, X 2i,..., X ki : E(u i X 1i, X 2i,..., X ki ) = 0 2. (X 1i, X 2i,..., X ki, Y i ), i = 1,..., n, are independently and identically distributed (i.i.d.) draws from their joint distribution 3. Large outliers are unlikely: X 1i, X 2i,..., X ki and Y i have nonzero finite fourth moments 4. There is no perfect multicollinearity Remarks: Note that we do not assume any specific parametric distribution for the u i The OLS assumptions imply specific distribution results 18
15 Theorem 2.4: (Unbiasedness, consistency, normality) Given the OLS assumptions the following properties of the OLS estimators ˆβ 0, ˆβ 1,..., ˆβ k hold: 1. ˆβ 0, ˆβ 1,..., ˆβ k are unbiased estimators of β 0,..., β k. 2. ˆβ 0, ˆβ 1,..., ˆβ k are consistent estimators of β 0,..., β k. (Convergence in probability) 3. In large samples ˆβ 0, ˆβ 1,..., ˆβ k are jointly normally distributed and each single OLS estimator ˆβ j, j = 0,..., k, is normally distributed with mean β j and variance σ 2ˆβ j, that is ˆβ j N(β j, σ 2ˆβ j ). 19
16 Remarks: In general, the OLS estimators are correlated This correlation among ˆβ 0, ˆβ 1,..., ˆβ k arises from the correlation among the regressors X 1,..., X k The sampling distribution of the OLS estimators will become relevant in Section 3 (hypothesis-testing, confidence intervals) 20
17 2.3. Measures-of-fit in multiple regression Now: Three well-known summary statistics that measure how well the OLS estimates fit the data Standard error of regression (SER): The SER estimates the standard deviation of the error term u i (under the assumption of homoskedasticity): SER = 1 n k 1 n û 2 i i=1 21
18 Standard error of regression: [continued] We denote the sum of squared residuals by SSR n i=1 û 2 i so that SER = SSR n k 1 Given the OLS assumptions and homoskedasticity the squared SER, (SER) 2, is an unbiased estimator of the unknown constant variance of the u i SER is a measure of the spread of the distribution of Y i around the population regression line Both measures, SER and SSR, are reported in the EViews regression output 22
19 R 2 : The R 2 is the fraction of the sample variance of the Y i explained by the regressors Equivalently, the R 2 is 1 minus the fraction of the variance of the Y i not explained by the regressors (i.e. explained by the residuals) Denoting the explained sum of squares (ESS) and the total sum of squares (TSS) by ESS = n i=1 (Ŷ i Ȳ ) 2 and TSS = respectively, we define the R 2 as R 2 = ESS TSS = 1 SSR TSS n i=1 (Y i Ȳ ) 2, 23
20 R 2 : [continued] In multiple regression, the R 2 increases whenever an additional regressor X k+1 is added to the regression model, unless the estimated coefficient ˆβ k+1 is exactly equal to zero Since in practice it is extremely unusual to have exactly ˆβ k+1 = 0, the R 2 generally increases (and never decreases) when an new regressor is added to the regression model An increase in the R 2 due to the inclusion of a new regressor does not necessarily indicate an actually improved fit of the model 24
21 Adjusted R 2 : The adjusted R 2 (in symbols: R 2 ), deflates the conventional R 2 : R 2 = 1 n 1 SSR n k 1TSS It is always true that R 2 < R 2 (why?) When adding a new regressor X k+1 to the model, the R 2 can increase or decrease (why?) The R 2 can be negative (why?) 25
22 2.4. Omitted-variable bias Now: Discussion of a phenomenon that implies violation of the first OLS assumption on Slide 18 This issue is known under the phrasing omitted-variable bias and is extremely relevant in practice Although theoretically easy to grasp, avoiding this specification problem turns out to be a nontrivial task in many empirical applications 26
23 Definition 2.5: (Omitted-variable bias) Consider the multiple regression model in Definition 2.1 on Slide 7. Omitted-variable bias is the bias in the OLS estimator ˆβ j of the coefficient β j (for j = 1,..., k) that arises when the associated regressor X j is correlated with an omitted variable. More precisely, for omitted-variable bias to occur, the following two conditions must hold: 1. X j is correlated with the omitted variable. 2. The omitted variable is a determinant of the dependent variable Y. 27
24 Example: Consider the house-prices dataset (Slides 16, 17) Using the entire set of regressors, we obtain the OLS estimate ˆβ 2 = for the BEDROOMS-coefficient The correlation coefficients between the regressors are as follows: BEDROOMS BATHROOMS LOTSIZE STOREYS BEDROOMS BATHROOMS LOTSIZE STOREYS
25 Example: [continued] There is positive (significant) correlation between the variable BEDROOMS and all other regressors Excluding the other variables from the regression yields the following OLS-estimates: Dependent Variable: SALEPRICE Method: Least Squares Date: 14/02/12 Time: 16:10 Sample: Included observations: 546 Variable Coefficient Std. Error t-statistic Prob. C BEDROOMS R-squared Mean dependent var Adjusted R-squared S.D. dependent var S.E. of regression Akaike info criterion Sum squared resid 3.36E+11 Schwarz criterion Log likelihood Hannan-Quinn criter F-statistic Durbin-Watson stat Prob(F-statistic) The alternative OLS-estimates of the BEDROOMS-coefficient differ substantially 29
26 Intuitive explanation of the omitted-variable bias: Consider the variable LOTSIZE as omitted LOTSIZE is an important variable for explaining SALEPRICE If we omit LOTSIZE in the regression, it will try to enter in the only way it can, namely through its positive correlation with the included variable BEDROOMS The coefficient on BEDROOMS will confound the effect of BED- ROOMS and LOTSIZE on SALEPRICE 30
27 More formal explanation: Omitted-variable bias means that the first OLS assumption on Slide 18 is violated Reasoning: In the multiple regression model the error term u i represents all factors other than the included regressors X 1,..., X k that are determinants of Y i If an omitted variable is correlated with at least one of the included regressors X 1,..., X k, then u i (which contains this factor) is correlated with the set of regressors This implies that E(u i X 1i,..., X ki ) 0 31
28 Important result: In the case of omitted-variable bias the OLS estimators on the corresponding included regressors are biased in finite samples this bias does not vanish in large samples the OLS estimators are inconsistent Solutions to omitted-variable bias: To be discussed in Section 5 32
29 2.5. Multicollinearity Definition 2.6: (Perfect multicollinearity) Consider the multiple regression model in Definition 2.1 on Slide 7. The regressors X 1,..., X k are said to be perfectly multicollinear if one of the regressors is a perfect linear function of the other regressors. Remarks: Under perfect multicollinearity the OLS estimates cannot be calculated due to division by zero in the OLS formulas Perfect multicollinearity often reflects a logical mistake in choosing the regressors or some unrecognized feature in the data set 33
30 Example: (Dummy variable trap) Consider the student-performance dataset Suppose we partition the school districts into the 3 categories (1) rural, (2) suburban, (3) urban We represent the categories by the dummy regressors { 1 if district i is rural RURAL i = 0 otherwise and by SUBURBAN i and URBAN i analogously defined Since each district belongs to one and only one category, we have for each district i: RURAL i + SUBURBAN i + URBAN i = 1 34
31 Example: [continued] Now, let us define the constant regressor X 0 associated with the intercept coefficient β 0 in the multiple regression model on Slide 7 by X 0i 1 for i = 1,... n Then, for i = 1,..., n, the following relationship holds among the regressors: Perfect multicollinearity X 0i = RURAL i + SUBURBAN i + URBAN i To estimate the regression we must exclude either one of the dummy regressors or the constant regressor X 0 (the intercept β 0 ) from the regression 35
32 Theorem 2.7: (Dummy variable trap) Let there be G different categories in the data set represented by G dummy regressors. If 1. each observation i falls into one and only one category, 2. there is an intercept (constant regressor) in the regression, 3. all G dummy regressors are included as regressors, then regression estimation fails because of perfect multicollinearity. Usual remedy: Exclude one of the dummy regressors (G 1 dummy regressors are sufficient) 36
33 Definition 2.8: (Imperfect multicollinearity) Consider the multiple regression model in Definition 2.1 on Slide 7. The regressors X 1,..., X k are said to be imperfectly multicollinear if two or more of the regressors are highly correlated in the sense that there is a linear function of the regressors that is highly correlated with another regressor. Remarks: Imperfect multicollinearity does not pose any (numeric) problems in calculating OLS estimates However, if regressors are imperfectly multicollinear, then the coefficients on at least one individual regressor will be imprecisely estimated 37
34 Remarks: [continued] Techniques for identifying and mitigating imperfect multicollinearity are presented in econometric textbooks (e.g. Hill et al., 2012, pp ) 38
Applied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple
More information6. Assessing studies based on multiple regression
6. Assessing studies based on multiple regression Questions of this section: What makes a study using multiple regression (un)reliable? When does multiple regression provide a useful estimate of the causal
More information10. Time series regression and forecasting
10. Time series regression and forecasting Key feature of this section: Analysis of data on a single entity observed at multiple points in time (time series data) Typical research questions: What is the
More information4. Nonlinear regression functions
4. Nonlinear regression functions Up to now: Population regression function was assumed to be linear The slope(s) of the population regression function is (are) constant The effect on Y of a unit-change
More informationCHAPTER 6: SPECIFICATION VARIABLES
Recall, we had the following six assumptions required for the Gauss-Markov Theorem: 1. The regression model is linear, correctly specified, and has an additive error term. 2. The error term has a zero
More informationContest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.
Updated: November 17, 2011 Lecturer: Thilo Klein Contact: tk375@cam.ac.uk Contest Quiz 3 Question Sheet In this quiz we will review concepts of linear regression covered in lecture 2. NOTE: Please round
More informationLinear Regression with Multiple Regressors
Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution
More information5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1)
5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1) Assumption #A1: Our regression model does not lack of any further relevant exogenous variables beyond x 1i, x 2i,..., x Ki and
More informationLinear Regression with Multiple Regressors
Linear Regression with Multiple Regressors (SW Chapter 6) Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution
More informationChapter 6: Linear Regression With Multiple Regressors
Chapter 6: Linear Regression With Multiple Regressors 1-1 Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS 4. Measures of fit 5. Sampling distribution
More information3. Linear Regression With a Single Regressor
3. Linear Regression With a Single Regressor Econometrics: (I) Application of statistical methods in empirical research Testing economic theory with real-world data (data analysis) 56 Econometrics: (II)
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests
ECON4150 - Introductory Econometrics Lecture 7: OLS with Multiple Regressors Hypotheses tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 7 Lecture outline 2 Hypothesis test for single
More informationThe Simple Regression Model. Part II. The Simple Regression Model
Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square
More information8. Instrumental variables regression
8. Instrumental variables regression Recall: In Section 5 we analyzed five sources of estimation bias arising because the regressor is correlated with the error term Violation of the first OLS assumption
More informationApplied Econometrics. Applied Econometrics Second edition. Dimitrios Asteriou and Stephen G. Hall
Applied Econometrics Second edition Dimitrios Asteriou and Stephen G. Hall MULTICOLLINEARITY 1. Perfect Multicollinearity 2. Consequences of Perfect Multicollinearity 3. Imperfect Multicollinearity 4.
More informationECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors
ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption
More information1 A Non-technical Introduction to Regression
1 A Non-technical Introduction to Regression Chapters 1 and Chapter 2 of the textbook are reviews of material you should know from your previous study (e.g. in your second year course). They cover, in
More informationSimple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationEconometrics 1. Lecture 8: Linear Regression (2) 黄嘉平
Econometrics 1 Lecture 8: Linear Regression (2) 黄嘉平 中国经济特区研究中 心讲师 办公室 : 文科楼 1726 E-mail: huangjp@szu.edu.cn Tel: (0755) 2695 0548 Office hour: Mon./Tue. 13:00-14:00 The linear regression model The linear
More informationECO321: Economic Statistics II
ECO321: Economic Statistics II Chapter 6: Linear Regression a Hiroshi Morita hmorita@hunter.cuny.edu Department of Economics Hunter College, The City University of New York a c 2010 by Hiroshi Morita.
More informationECON3150/4150 Spring 2015
ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2
More information2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0
Introduction to Econometrics Midterm April 26, 2011 Name Student ID MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. (5,000 credit for each correct
More informationPractice Questions for the Final Exam. Theoretical Part
Brooklyn College Econometrics 7020X Spring 2016 Instructor: G. Koimisis Name: Date: Practice Questions for the Final Exam Theoretical Part 1. Define dummy variable and give two examples. 2. Analyze the
More informationThe general linear regression with k explanatory variables is just an extension of the simple regression as follows
3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because
More informationLECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity
LECTURE 10 Introduction to Econometrics Multicollinearity & Heteroskedasticity November 22, 2016 1 / 23 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists
More informationHeteroskedasticity. Part VII. Heteroskedasticity
Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least
More informationIntroduction to Econometrics. Multiple Regression (2016/2017)
Introduction to Econometrics STAT-S-301 Multiple Regression (016/017) Lecturer: Yves Dominicy Teaching Assistant: Elise Petit 1 OLS estimate of the TS/STR relation: OLS estimate of the Test Score/STR relation:
More informationThe Simple Linear Regression Model
The Simple Linear Regression Model Lesson 3 Ryan Safner 1 1 Department of Economics Hood College ECON 480 - Econometrics Fall 2017 Ryan Safner (Hood College) ECON 480 - Lesson 3 Fall 2017 1 / 77 Bivariate
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More informationHeteroscedasticity 1
Heteroscedasticity 1 Pierre Nguimkeu BUEC 333 Summer 2011 1 Based on P. Lavergne, Lectures notes Outline Pure Versus Impure Heteroscedasticity Consequences and Detection Remedies Pure Heteroscedasticity
More informationECON 4230 Intermediate Econometric Theory Exam
ECON 4230 Intermediate Econometric Theory Exam Multiple Choice (20 pts). Circle the best answer. 1. The Classical assumption of mean zero errors is satisfied if the regression model a) is linear in the
More informationAnswers to Problem Set #4
Answers to Problem Set #4 Problems. Suppose that, from a sample of 63 observations, the least squares estimates and the corresponding estimated variance covariance matrix are given by: bβ bβ 2 bβ 3 = 2
More informationIntroduction to Econometrics. Multiple Regression
Introduction to Econometrics The statistical analysis of economic (and related) data STATS301 Multiple Regression Titulaire: Christopher Bruffaerts Assistant: Lorenzo Ricci 1 OLS estimate of the TS/STR
More informationECON3150/4150 Spring 2016
ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression
More informationModel Specification and Data Problems. Part VIII
Part VIII Model Specification and Data Problems As of Oct 24, 2017 1 Model Specification and Data Problems RESET test Non-nested alternatives Outliers A functional form misspecification generally means
More information7. Integrated Processes
7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider
More informationStatistical Inference. Part IV. Statistical Inference
Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided
More informationMultiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =
Economics 130 Lecture 6 Midterm Review Next Steps for the Class Multiple Regression Review & Issues Model Specification Issues Launching the Projects!!!!! Midterm results: AVG = 26.5 (88%) A = 27+ B =
More informationMultiple Regression Analysis
Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple
More informationLecture 5. In the last lecture, we covered. This lecture introduces you to
Lecture 5 In the last lecture, we covered. homework 2. The linear regression model (4.) 3. Estimating the coefficients (4.2) This lecture introduces you to. Measures of Fit (4.3) 2. The Least Square Assumptions
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More information13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity
Outline: Further Issues in Using OLS with Time Series Data 13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process I. Stationary and Weakly Dependent Time Series III. Highly Persistent
More informationThe Multiple Regression Model Estimation
Lesson 5 The Multiple Regression Model Estimation Pilar González and Susan Orbe Dpt Applied Econometrics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 5 Regression model:
More informationMultiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor)
1 Multiple Regression Analysis: Estimation Simple linear regression model: an intercept and one explanatory variable (regressor) Y i = β 0 + β 1 X i + u i, i = 1,2,, n Multiple linear regression model:
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationMotivation for multiple regression
Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope
More informationEconomics 471: Econometrics Department of Economics, Finance and Legal Studies University of Alabama
Economics 471: Econometrics Department of Economics, Finance and Legal Studies University of Alabama Course Packet The purpose of this packet is to show you one particular dataset and how it is used in
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction
More informationOutline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section
Outline I. The Nature of Time Series Data 11. Time Series Analysis II. Examples of Time Series Models IV. Functional Form, Dummy Variables, and Index Basic Regression Numbers Read Wooldridge (2013), Chapter
More informationLinear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons
Linear Regression with 1 Regressor Introduction to Econometrics Spring 2012 Ken Simons Linear Regression with 1 Regressor 1. The regression equation 2. Estimating the equation 3. Assumptions required for
More informationFreeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94
Freeing up the Classical Assumptions () Introductory Econometrics: Topic 5 1 / 94 The Multiple Regression Model: Freeing Up the Classical Assumptions Some or all of classical assumptions needed for derivations
More informationThe F distribution. If: 1. u 1,,u n are normally distributed; and 2. X i is distributed independently of u i (so in particular u i is homoskedastic)
The F distribution If: 1. u 1,,u n are normally distributed; and. X i is distributed independently of u i (so in particular u i is homoskedastic) then the homoskedasticity-only F-statistic has the F q,n-k
More informationECON3150/4150 Spring 2016
ECON3150/4150 Spring 2016 Lecture 6 Multiple regression model Siv-Elisabeth Skjelbred University of Oslo February 5th Last updated: February 3, 2016 1 / 49 Outline Multiple linear regression model and
More informationNonlinear Regression Functions
Nonlinear Regression Functions (SW Chapter 8) Outline 1. Nonlinear regression functions general comments 2. Nonlinear functions of one variable 3. Nonlinear functions of two variables: interactions 4.
More informationChapter 7. Hypothesis Tests and Confidence Intervals in Multiple Regression
Chapter 7 Hypothesis Tests and Confidence Intervals in Multiple Regression Outline 1. Hypothesis tests and confidence intervals for a single coefficie. Joint hypothesis tests on multiple coefficients 3.
More information11. Simultaneous-Equation Models
11. Simultaneous-Equation Models Up to now: Estimation and inference in single-equation models Now: Modeling and estimation of a system of equations 328 Example: [I] Analysis of the impact of advertisement
More informationEastern Mediterranean University Department of Economics ECON 503: ECONOMETRICS I. M. Balcilar. Midterm Exam Fall 2007, 11 December 2007.
Eastern Mediterranean University Department of Economics ECON 503: ECONOMETRICS I M. Balcilar Midterm Exam Fall 2007, 11 December 2007 Duration: 120 minutes Questions Q1. In order to estimate the demand
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More information7. Integrated Processes
7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More informationLecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables
Lecture 8. Using the CLR Model Relation between patent applications and R&D spending Variables PATENTS = No. of patents (in 000) filed RDEP = Expenditure on research&development (in billions of 99 $) The
More informationAbout the seasonal effects on the potential liquid consumption
About the seasonal effects on the potential liquid consumption Lucie Ravelojaona Guillaume Perrez Clément Cousin ENAC 14/01/2013 Consumption raw data Figure : Evolution during one year of different family
More informationReliability of inference (1 of 2 lectures)
Reliability of inference (1 of 2 lectures) Ragnar Nymoen University of Oslo 5 March 2013 1 / 19 This lecture (#13 and 14): I The optimality of the OLS estimators and tests depend on the assumptions of
More informationECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests
ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one
More informationChapter 2: simple regression model
Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.
More informationSTOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Friday, June 5, 009 Examination time: 3 hours
More informationEssential of Simple regression
Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship
More informationProblem set 1: answers. April 6, 2018
Problem set 1: answers April 6, 2018 1 1 Introduction to answers This document provides the answers to problem set 1. If any further clarification is required I may produce some videos where I go through
More informationBrief Sketch of Solutions: Tutorial 3. 3) unit root tests
Brief Sketch of Solutions: Tutorial 3 3) unit root tests.5.4.4.3.3.2.2.1.1.. -.1 -.1 -.2 -.2 -.3 -.3 -.4 -.4 21 22 23 24 25 26 -.5 21 22 23 24 25 26.8.2.4. -.4 - -.8 - - -.12 21 22 23 24 25 26 -.2 21 22
More informationEconometrics I Lecture 3: The Simple Linear Regression Model
Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating
More informationP1.T2. Stock & Watson Chapters 4 & 5. Bionic Turtle FRM Video Tutorials. By: David Harper CFA, FRM, CIPM
P1.T2. Stock & Watson Chapters 4 & 5 Bionic Turtle FRM Video Tutorials By: David Harper CFA, FRM, CIPM Note: This tutorial is for paid members only. You know who you are. Anybody else is using an illegal
More informationWISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A
WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2015-16 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This
More informationLECTURE 11. Introduction to Econometrics. Autocorrelation
LECTURE 11 Introduction to Econometrics Autocorrelation November 29, 2016 1 / 24 ON PREVIOUS LECTURES We discussed the specification of a regression equation Specification consists of choosing: 1. correct
More informationECNS 561 Multiple Regression Analysis
ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking
More informationEcon 427, Spring Problem Set 3 suggested answers (with minor corrections) Ch 6. Problems and Complements:
Econ 427, Spring 2010 Problem Set 3 suggested answers (with minor corrections) Ch 6. Problems and Complements: 1. (page 132) In each case, the idea is to write these out in general form (without the lag
More informationEconometrics. Final Exam. 27thofJune,2008. Timeforcompletion: 2h30min
Econometrics Final Exam 27thofJune,2008 João Valle e Azevedo António José Morgado Tiago Silva Vieira Timeforcompletion: 2h30min Give your answers in the space provided. Usedraftpapertoplanyouranswersbeforewritingthemontheexampaper.
More informationEconometrics Midterm Examination Answers
Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i
More informationARDL Cointegration Tests for Beginner
ARDL Cointegration Tests for Beginner Tuck Cheong TANG Department of Economics, Faculty of Economics & Administration University of Malaya Email: tangtuckcheong@um.edu.my DURATION: 3 HOURS On completing
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution
More informationExercises (in progress) Applied Econometrics Part 1
Exercises (in progress) Applied Econometrics 2016-2017 Part 1 1. De ne the concept of unbiased estimator. 2. Explain what it is a classic linear regression model and which are its distinctive features.
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model
Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory
More informationHypothesis Tests and Confidence Intervals. in Multiple Regression
ECON4135, LN6 Hypothesis Tests and Confidence Intervals Outline 1. Why multipple regression? in Multiple Regression (SW Chapter 7) 2. Simpson s paradox (omitted variables bias) 3. Hypothesis tests and
More informationECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47
ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with
More informationLecture 4: Regression Analysis
Lecture 4: Regression Analysis 1 Regression Regression is a multivariate analysis, i.e., we are interested in relationship between several variables. For corporate audience, it is sufficient to show correlation.
More informationG. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication
G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationECON 366: ECONOMETRICS II. SPRING TERM 2005: LAB EXERCISE #10 Nonspherical Errors Continued. Brief Suggested Solutions
DEPARTMENT OF ECONOMICS UNIVERSITY OF VICTORIA ECON 366: ECONOMETRICS II SPRING TERM 2005: LAB EXERCISE #10 Nonspherical Errors Continued Brief Suggested Solutions 1. In Lab 8 we considered the following
More informationRockefeller College University at Albany
Rockefeller College University at Albany PAD 705 Handout: Suggested Review Problems from Pindyck & Rubinfeld Original prepared by Professor Suzanne Cooper John F. Kennedy School of Government, Harvard
More informationLecture #8 & #9 Multiple regression
Lecture #8 & #9 Multiple regression Starting point: Y = f(x 1, X 2,, X k, u) Outcome variable of interest (movie ticket price) a function of several variables. Observables and unobservables. One or more
More informationEconometrics Honor s Exam Review Session. Spring 2012 Eunice Han
Econometrics Honor s Exam Review Session Spring 2012 Eunice Han Topics 1. OLS The Assumptions Omitted Variable Bias Conditional Mean Independence Hypothesis Testing and Confidence Intervals Homoskedasticity
More information2 Prediction and Analysis of Variance
2 Prediction and Analysis of Variance Reading: Chapters and 2 of Kennedy A Guide to Econometrics Achen, Christopher H. Interpreting and Using Regression (London: Sage, 982). Chapter 4 of Andy Field, Discovering
More informationHypothesis Tests and Confidence Intervals in Multiple Regression
Hypothesis Tests and Confidence Intervals in Multiple Regression (SW Chapter 7) Outline 1. Hypothesis tests and confidence intervals for one coefficient. Joint hypothesis tests on multiple coefficients
More informationEconometrics Problem Set 11
Econometrics Problem Set WISE, Xiamen University Spring 207 Conceptual Questions. (SW 2.) This question refers to the panel data regressions summarized in the following table: Dependent variable: ln(q
More informationMultiple Regression Analysis: Heteroskedasticity
Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing
More informationEconometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in
More informationThe multiple regression model; Indicator variables as regressors
The multiple regression model; Indicator variables as regressors Ragnar Nymoen University of Oslo 28 February 2013 1 / 21 This lecture (#12): Based on the econometric model specification from Lecture 9
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More information