Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
|
|
- Joseph Norton
- 5 years ago
- Views:
Transcription
1 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 1 / 33
2 The multiple linear regression model The major drawback of the bivariate regression model is that the key assumption SLR.4 (zero conditional mean) is often unrealistic. The multiple linear regression model (MLR) allows to control for many other factors which might otherwise be captured in the error term. Thus it is more amenable to ceteris paribus analysis. The model with k independent variables given a sample (y i, x i1,..., x ik ), i = 1,..., n, reads y i = β 0 + β 1 x i1 + β 2 x i2 + + β k x ik + u i (1) y i = x i β + u i with x i = (1, x i1,..., x ik ) and β = (β 0, β 1,..., β k ). The key assumption is E(u i x i ) = 0 i = 1,..., n (3) (2) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 2 / 33
3 MLR Matrix notation Let y be a n 1 vector of observations on y, let X be the data matrix with dimension n (k + 1) and associated parameter vector β R (k+1) 1, and let u be a n 1 vector of disturbances. With these ingredients we can write the model in (1) as y = Xβ + u (4) or more explicitly, y 1 1 x 11 x x 1k β 0 u 1 y 2. = 1 x 21 x x 2k β u 2. y n 1 x n1 x n2... x nk β k u n (5) where Xβ is the systematic and u the stochastic component. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 3 / 33
4 MLR An example We are interested in the effect of education on hourly wage: wage i = β 0 + β 1 educ i + β 2 exper i + u i, i = 1,..., n (6) We control for years of labor market experience (exper). We are still primarily interested in the effect of education. The MLR takes experience out of the error term u. With the SLR we would have to assume exper educ. 1 β 1 measures the effect of educ on wage when exper is held constant. β 2 measures the effect of exper on wage when educ is held constant. 1 denotes statistical independence. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 4 / 33
5 MLR Estimation As in Module 1, we are looking for a coefficient vector ˆβ R (k+1) 1 that minimizes the sum of squared residuals. Formally the problem reads arg min ˆβ n û 2 i = û û = (y X ˆβ) (y X ˆβ) (7) i=1 = y y ˆβ X y y X ˆβ + ˆβ X X ˆβ (8) = y y 2 ˆβ X y + ˆβ X X ˆβ (9) Note that the last step is possible because ˆβ X y = (y X ˆβ) = y X ˆβ. The first-order condition for minimization is û û ˆβ = 2X y + 2X X ˆβ = 0 (10) or written differently X X ˆβ = X y (11) which is called the system of least squares normal equations. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 5 / 33
6 MLR Estimation If X X is non-singular (i.e., there exists an inverse), pre-multiplying both sides of equation (11) with (X X) 1 yields the OLS estimators ˆβ: ˆβ = (X X) 1 X y (12) Important: The matrix X X is non-singular (= invertible) and ˆβ a unique solution to the minimization problem if and only if we have at least n k observations, and the data matrix X has rank (k + 1). The second point is violated if there are linear dependencies among the explanatory variables (i.e., perfect collinearity). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 6 / 33
7 MLR Properties of the OLS estimator The OLS estimator has various important properties that do not depend on any assumptions, but rather arise by how it is constructed. First, substitute y = X ˆβ + û into the system of normal equations (11) to obtain X X ˆβ = X y X X ˆβ = X (X ˆβ + û) X X ˆβ = X X ˆβ + X û 0 = X û (13) A number of important properties can be derived from this condition. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 7 / 33
8 MLR Properties of the OLS estimator 1 The observed values of X are uncorrelated with the residuals û. This follows immediately from (13): X û = 0 iff. X û. Note that this does not mean that X is uncorrelated with u! We have to assume this. 2 The sum of residuals is zero. If there is a constant, the first column of X will be a column of ones. For the first element in the X û to be zero it must hold that ûi = 0. i 3 The sample mean of the residuals is zero. This follows from the previous property: û = n 1 n û = i= The regression hyperplane passes through the means of the observed values X and ȳ. Recall that û = y X ˆβ. Dividing by n gives û = ȳ X ˆβ. From the previous property: û = ȳ X ˆβ = 0, so ȳ = X ˆβ. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 8 / 33
9 MLR Properties of the OLS estimator 5 The predicted values ŷ are uncorrelated with the residuals û. The predicted values are ŷ = X ˆβ. From this we have ŷ û = (X ˆβ) û = ˆβ X û = 0 because X û = 0. 6 The mean of the predicted Y s for the sample will equal the mean of the observed Y s, i.e. ŷ = ȳ. Proof is left as an exercise (use the result in item 4). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 9 / 33
10 MLR Properties of the OLS estimator These properties always hold true, be careful not to infer anything from the residuals about the actual disturbances! So far we know nothing about ˆβ except that it satisfies all of the properties discussed above. We need to make some assumptions about the true model in order to make any inferences regarding β (the true population parameters) from ˆβ (our estimator of the true parameters). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 10 / 33
11 MLR Expected value and variance The assumptions from the bivariate model translate to the multivariate case as follows: Assumption MLR.1 Linear in parameters The population model is linear: y = β X + u. Assumption MLR.2 Random sampling We have a random sample of n observations, {(y i, x i) i = 1,..., n}, that follows the population model in assumption MLR.1. Assumption MLR.3 No perfect collinearity The data matrix X has rank (k + 1). Assumption MLR.4 Zero conditional mean Conditional on the entire matrix X, each error u i has mean zero: E(u i X) = 0 i = 1,..., n. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 11 / 33
12 Finite sample properties Note that we only need assumption MLR.3 (no perfect collinearity) to obtain an OLS estimate ˆβ. Whether this estimate actually makes sense, i.e. is unbiased and representative for the full population, depends on the other assumptions. Especially the zero conditional mean assumption (MLR.4) often poses problems in practice. Requires that, conditional on the observed covariates x i, unobservables are on average orthogonal to the error term u i. Fails in the case of Simultaneity Selection Omitted variables Functional form misspecification Measurement error We will discuss sources and consequences of these cases in Module 6. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 12 / 33
13 Finite sample properties Expected values Theorem: Unbiasedness of OLS Under assumptions MLR.1 through MLR.4, E( ˆβ) = β. (14) In other words, ˆβ is an unbiased estimate for β. Proof. Rewrite the OLS estimator as ˆβ = (X X) 1 X y = (X X) 1 X (Xβ + u) = (X X) 1 (X X)β + (X X) 1 X u (15) Because X X is a square matrix, (X X) 1 (X X) = I, thus ˆβ = β + (X X) 1 X u (16) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 13 / 33
14 Finite sample properties Expected values Taking conditional expectations on both sides of equation (16) gives E( ˆβ X) = β + (X X) 1 X E(u X) (17) By MLR.4, E(u X) = 0, so This completes the proof. E( ˆβ X) = β (18) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 14 / 33
15 Finite sample properties Expected variances Assumption MLR.1 Linear in parameters The population model is linear: y = β X + u. Assumption MLR.2 Random sampling We have a random sample {(y i, x i) i = 1,..., n} that follows the population model: y i = β 0 + β 1x i + u i. Assumption MLR.3 No perfect collinearity The data matrix X has rank (k + 1). Assumption MLR.4 Zero conditional mean Conditional on the entire matrix X, each error u i has zero mean: E(u i X) = 0 i = 1,..., n. Assumption MLR.5 Homoskedasticity and no serial correlation The error u i has the same variance given any values of the covariates, i.e. Var(u i X) = σ 2, i = 1,..., n, and there is no serial correlation between the errors: Cov(u i, u j X) = 0 for all j i. We can write these two assumptions as Var(u X) = σ 2 I. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 15 / 33
16 Finite sample properties Expected variances Assumption MLR.5 requires that errors are homoskedastic and that there is no serial correlation (meaning that errors are not correlated across observations this is especially important if you deal with panel data, but sometimes also in cross-sectional settings). Combining these assumptions, we can write the variance-covariance matrix of the disturbances as σ Var(u X) = E(uu 0 σ X) = = σ2 I (19) σ 2 Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 16 / 33
17 Finite sample properties Expected variances Theorem: Variance-covariance matrix of the OLS estimator Under assumptions MLR.1 through MLR.5, Var( ˆβ) = σ 2 (X X) 1 (20) Proof. See Wooldridge (2013), p For one particular ˆβ j ˆβ, the variance is obtained by multiplying σ 2 by the jth diagonal element of (X X) 1. It can also be written as Var( ˆβ j ) = σ 2 SST j (1 R 2 j ) (21) where SST j = n i=1 (x ij x j ) 2 is the total sample variation in x j and R 2 j is the R-squared from regressing x j on all other independent variables (and including an intercept). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 17 / 33
18 Finite sample properties Expected variances The unbiased estimator of the error variance in the multivariate case is given by ˆσ 2 = u u n k 1 (22) where u u is again the sum of squared residuals. Theorem: Unbiasedness of σ 2 Under assumptions MLR.1 through MLR.5, ˆσ 2 is an unbiased estimate for σ 2. That is, E(ˆσ 2 X) = σ 2 σ 2 > 0 (23) Proof. See Wooldridge (2013), p Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 18 / 33
19 Finite sample properties Gauss-Markov theorem Gauss-Markov Theorem Under assumptions MLR.1 through MLR.5, ˆβ is the best linear unbiased estimator. Proof. See Wooldridge (2013), p The Gauss-Markov theorem translated: OLS is the estimator with the smallest variance amongst all linear unbiased estimators. OLS is BLUE. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 19 / 33
20 Inference Sampling distributions of the OLS estimators Although it is not necessary for the Gauss-Markov theorem to hold, 2 we assume normally distributed disturbances to derive sampling distributions. Assumption MLR.6 Normality of errors Conditional on X, u is distributed as multivariate normal with mean zero and variance-covariance matrix σ 2 I. That is, u Normal(0, σ 2 I). (24) 2 We will show later that, as soon as asymptotics kick in, we don t need the normality assumption anymore for our test statistics to be valid. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 20 / 33
21 Inference Sampling distributions of the OLS estimators Theorem: Normality of ˆβ Under the classical linear model assumptions MLR.1 through MLR.6, ˆβ conditional on X is distributed as multivariate normal with mean β and variance-covariance matrix σ 2 X X 1. That is, ˆβ Normal(β, σ 2 (X X) 1 ) (25) Therefore, ˆβ j β j sd( ˆβ j ) Normal(0, 1) (26) Proof. Wooldridge (2013, p. 113) provides a sketch of the proof for (25). The result in (26) is straightforward; if we substract the mean from a normally distributed random variable and divide by its standard deviation, we get a standard normal variable with mean zero and a standard deviation of 1. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 21 / 33
22 Inference Sampling distributions of the OLS estimators Theorem: Distribution of t-statistics Under assumptions MLR.1 through MLR.6, Proof. Wooldridge (2013), p ˆβ j β j se( ˆβ t n k 1 (27) j ) }{{} t-statistic This is an important result for inference. It says that, when we estimate σ in sd( ˆβ j ) by ˆσ which yields se( ˆβ j ), ( ˆβ j β j )/se( ˆβ j ) is t-distributed with n k 1 degrees of freedom. Note that β j is some hypothesized value. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 22 / 33
23 Inference Testing a single population parameter Pick a significance level and formulate a null-hypothesis (H 0 ): One-sided alternatives: H 0 : β j 0; H 1 : β j > 0. Two-sided alternatives: H 0 : β j = 0; H 1 : β j 0. One-sided alternatives: H 0 : β j > α j; H 1 : β j α j, with α R. t-statistic: t (estimate hypothesized value), according to theorem (27). standard error p-value for t-test: what is the smallest significance level at which H 0 would be rejected? Confidence intervals range of likely values for β: CI ˆβ j ± c se( ˆβ j ) (28) where c is the 97.5 th percentile of the t n k 1 distribution. Economic vs. statistical significance. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 23 / 33
24 Inference Testing multiple linear restrictions The F -test allows to test for multiple hypotheses. Consider a model with k = 4 independent variables: y i = β 0 + β 1 x i1 + β 2 x i2 + β 3 x i3 + β 4 x i4 + u i. Suppose you want to test whether x 1, x 2, and x 3 are jointly insignificant Formulate a null-hypothesis: H 0 : β 1 = β 2 = β 3 = 0 H 1 : H 0 is not true. F -statistic: F (SSR r SSR ur )/q SSR ur /(n k 1) F q,n k 1 (29) where q is the number of restrictions, SSR r is the sum of squared residuals from the restricted model and SSR ur is the SSR from the unrestricted model. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 24 / 33
25 Inference Example We estimate the following model: final i = β 0 + β 1 attend i + β 2 hwrte i + u i (30) where final [10, 39] are final exam points, attend [2, 32] is the number of classes attended out of 32, and hwrte is the percentage of homeworks turned in times reg final attend hwrte Source SS df MS Number of obs = 674 F( 2, 671) = 9.20 Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE = final Coef. Std. Err. t P> t [95% Conf. Interval] attend hwrte _cons Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 25 / 33
26 Asymptotics So far, we have looked at finite sample properties of OLS, i.e., properties that hold independent of how large n is. However, it is also important to know large sample or asymptotic properties of OLS. These are defined as sample size grows without bound. An important result is that even without assuming normality (MLR.6), t and F statistics are approximately t and F distributed. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 26 / 33
27 Asymptotics Consistency In general, an estimator ˆθ is consistent, if it converges in probability 3 to the population parameter θ, that is, ˆθ p θ, or plim ˆθ = θ. Note that unbiasedness does not necessarily imply consistency, and consistency does not automatically imply unbiasedness. Consistency of OLS: OLS is unbiased under assumptions MLR.1 through MLR.4, so ˆβ j is always distributed around β j. The distribution of ˆβ j becomes more and more tightly distributed around β j as the sample size grows. As n, the distribution of ˆβ j collapses to a single point β j. 3 A random variable X n converges to X in probability if for some ε > 0, lim P( X n X ε) = 0 (31) n Note that here the probability converges, not the random variable itself. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 27 / 33
28 Asymptotics Consistency Figure: Sampling distributions of ˆβ 1 [Source: Wooldridge (2013), Figure 5.1]. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 28 / 33
29 Asymptotics Consistency Theorem: Consistency of OLS Under assumptions MLR.1 through MLR.4, the OLS estimator ˆβ is consistent for β. Proof. We show consistency for the bivariate case with one regressor β 1, the general proof for k regressors is given in Wooldridge (2013), p First note that we can also write the OLS estimator ˆβ 1 simply as ˆβ 1 = n i=1 x iy i n i=1 x2 i (32) and after plugging in y i = βx i + u i and some algebra, we obtain ˆβ 1 = β + n 1 n i=1 x iu i n 1 n i=1 x2 i (33) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 29 / 33
30 Asymptotics Consistency As n, we have ˆβ 1 p β + plim n 1 n i=1 x iu i plim n 1 n i=1 x2 i By the law of large numbers, 4 we have plim n 1 n i=1 x i = E(x i ) and plim n 1 n i=1 u i = E(u i ). Since we assume zero conditional mean (MLR.4), which obviously implies E(u i ) = 0, we get ˆβ 1 p β + 0 plim n 1 n i=1 x2 i (34) = β (36) This proofs that ˆβ 1 converges to β in probability. 4 Let X 1,..., X n be some sequence of i.i.d. random variables with arbitrary distribution. The law of large number states that X n = n 1 (X 1,..., X n) E(X n) (35) That is, the sample average converges to the expected value. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 30 / 33
31 Asymptotics Consistency Consistency is related to bias as follows: An estimator ˆθ is consistent iff. it converges to some value θ and the bias, i.e., Bias(θ) = E(ˆθ) θ, converges to zero. Individual estimators in the sequence ˆθ j ˆθ may be biased, but the overall sequence is still consistent if the bias converges to zero. Estimators can be Unbiased but not consistent Biased but consistent Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 31 / 33
32 Asymptotics Normality Another important large sample property of OLS is that ˆβ j is asymptotically normally distributed under assumptions MLR.1 through MLR.5. Formally, ˆβ j β se( ˆβ j ) where Var( ˆβ j ) is the usual OLS standard error. Result stems from the central limit theorem. 5 a Normal(0, 1) (37) We do not need the normality assumption in MLR.6 for our test statistics to be valid, as long as the sample size is reasonably large. All we have to assume is finite variance; Var(u) <. 5 The central limit theorem states that the standardized sums of i.i.d. random variables converges to a normal distribution, irrespective of their own distributions. Let X 1,..., X n be i.i.d. random variables with finite expected value µ and variance σ 2. Then ) ) 1 nσ ( n i=1 X i nµ = 1 n ( n i=1 X µ σ = X µ σ d n N(0, 1) (38) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 32 / 33
33 Literature Main reference: Wooldridge, J. M. (2015). Introductory Econometrics: A Modern Approach, 5th ed., South Western College Publishing. Additional reference: Greene, W. H. (2012). Econometric Analysis, 7th edition, Pearson. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 33 / 33
Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018
Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate
More informationStatistical Inference with Regression Analysis
Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationThe Simple Regression Model. Part II. The Simple Regression Model
Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square
More informationECON3150/4150 Spring 2015
ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2
More informationTHE MULTIVARIATE LINEAR REGRESSION MODEL
THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationThe general linear regression with k explanatory variables is just an extension of the simple regression as follows
3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because
More informationcoefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1
Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple
More informationPractical Econometrics. for. Finance and Economics. (Econometrics 2)
Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationMultiple Linear Regression
Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random
More informationMultiple Linear Regression CIVL 7012/8012
Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationHeteroskedasticity. Part VII. Heteroskedasticity
Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least
More informationLECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit
LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define
More informationECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47
ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with
More informationECNS 561 Multiple Regression Analysis
ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationECON3150/4150 Spring 2016
ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression
More informationα version (only brief introduction so far)
Econometrics I KS Module 8: Panel Data Econometrics Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: June 18, 2018 α version (only brief introduction so far) Alexander
More informationLecture 8: Instrumental Variables Estimation
Lecture Notes on Advanced Econometrics Lecture 8: Instrumental Variables Estimation Endogenous Variables Consider a population model: y α y + β + β x + β x +... + β x + u i i i i k ik i Takashi Yamano
More informationMultiple Regression Analysis
Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators
More informationMultivariate Regression Analysis
Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x
More informationSimple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 16, 2013 Outline Introduction Simple
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationWarwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation
Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Michele Aquaro University of Warwick This version: July 21, 2016 1 / 31 Reading material Textbook: Introductory
More informationEconometrics - 30C00200
Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business
More informationEconometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in
More information2. (3.5) (iii) Simply drop one of the independent variables, say leisure: GP A = β 0 + β 1 study + β 2 sleep + β 3 work + u.
BOSTON COLLEGE Department of Economics EC 228 Econometrics, Prof. Baum, Ms. Yu, Fall 2003 Problem Set 3 Solutions Problem sets should be your own work. You may work together with classmates, but if you
More informationLinear Regression. Junhui Qian. October 27, 2014
Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests
ECON4150 - Introductory Econometrics Lecture 7: OLS with Multiple Regressors Hypotheses tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 7 Lecture outline 2 Hypothesis test for single
More information1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11
Econ 495 - Econometric Review 1 Contents 1 Linear Regression Analysis 4 1.1 The Mincer Wage Equation................. 4 1.2 Data............................. 6 1.3 Econometric Model.....................
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationINTRODUCTION TO BASIC LINEAR REGRESSION MODEL
INTRODUCTION TO BASIC LINEAR REGRESSION MODEL 13 September 2011 Yogyakarta, Indonesia Cosimo Beverelli (World Trade Organization) 1 LINEAR REGRESSION MODEL In general, regression models estimate the effect
More informationBusiness Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM
Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF
More informationLecture 3: Multivariate Regression
Lecture 3: Multivariate Regression Rates, cont. Two weeks ago, we modeled state homicide rates as being dependent on one variable: poverty. In reality, we know that state homicide rates depend on numerous
More informationECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests
ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one
More informationLab 07 Introduction to Econometrics
Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand
More informationECO220Y Simple Regression: Testing the Slope
ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x
More informationEconometrics of Panel Data
Econometrics of Panel Data Jakub Mućk Meeting # 2 Jakub Mućk Econometrics of Panel Data Meeting # 2 1 / 26 Outline 1 Fixed effects model The Least Squares Dummy Variable Estimator The Fixed Effect (Within
More informationInstrumental Variables, Simultaneous and Systems of Equations
Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated
More informationSection I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -
First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III
More informationAnswer all questions from part I. Answer two question from part II.a, and one question from part II.b.
B203: Quantitative Methods Answer all questions from part I. Answer two question from part II.a, and one question from part II.b. Part I: Compulsory Questions. Answer all questions. Each question carries
More informationMultiple Regression Analysis: Heteroskedasticity
Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing
More informationAdvanced Quantitative Methods: ordinary least squares
Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the
More informationMultiple Regression: Inference
Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled
More informationEconometrics Multiple Regression Analysis: Heteroskedasticity
Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 17, 2012 Outline Heteroskedasticity
More informationECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors
ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption
More informationMultiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators
1 2 Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE Hüseyin Taştan 1 1 Yıldız Technical University Department of Economics These presentation notes are based on Introductory
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationEcon 510 B. Brown Spring 2014 Final Exam Answers
Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model
Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory
More informationSimple Linear Regression Model & Introduction to. OLS Estimation
Inside ECOOMICS Introduction to Econometrics Simple Linear Regression Model & Introduction to Introduction OLS Estimation We are interested in a model that explains a variable y in terms of other variables
More informationCourse Econometrics I
Course Econometrics I 4. Heteroskedasticity Martin Halla Johannes Kepler University of Linz Department of Economics Last update: May 6, 2014 Martin Halla CS Econometrics I 4 1/31 Our agenda for today Consequences
More information1 Independent Practice: Hypothesis tests for one parameter:
1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More informationECON Introductory Econometrics. Lecture 13: Internal and external validity
ECON4150 - Introductory Econometrics Lecture 13: Internal and external validity Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 9 Lecture outline 2 Definitions of internal and external
More informationEconometrics of Panel Data
Econometrics of Panel Data Jakub Mućk Meeting # 1 Jakub Mućk Econometrics of Panel Data Meeting # 1 1 / 31 Outline 1 Course outline 2 Panel data Advantages of Panel Data Limitations of Panel Data 3 Pooled
More informationEmpirical Application of Simple Regression (Chapter 2)
Empirical Application of Simple Regression (Chapter 2) 1. The data file is House Data, which can be downloaded from my webpage. 2. Use stata menu File Import Excel Spreadsheet to read the data. Don t forget
More informationThe Simple Regression Model. Simple Regression Model 1
The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising
More informationEconometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage
More informationEconometrics Master in Business and Quantitative Methods
Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid Models with discrete dependent variables and applications of panel data methods in all fields of economics
More informationEconometrics Midterm Examination Answers
Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i
More informationHeteroskedasticity and Autocorrelation
Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationMultiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor)
1 Multiple Regression Analysis: Estimation Simple linear regression model: an intercept and one explanatory variable (regressor) Y i = β 0 + β 1 X i + u i, i = 1,2,, n Multiple linear regression model:
More informationRegression #8: Loose Ends
Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch
More informationEmpirical Economic Research, Part II
Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction
More informationAdvanced Econometrics
Based on the textbook by Verbeek: A Guide to Modern Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna May 16, 2013 Outline Univariate
More informationMotivation for multiple regression
Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope
More informationEconometrics Homework 1
Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 11, 2012 Outline Heteroskedasticity
More informationStatistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression
Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression
More informationOutline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section
Outline I. The Nature of Time Series Data 11. Time Series Analysis II. Examples of Time Series Models IV. Functional Form, Dummy Variables, and Index Basic Regression Numbers Read Wooldridge (2013), Chapter
More informationLab 6 - Simple Regression
Lab 6 - Simple Regression Spring 2017 Contents 1 Thinking About Regression 2 2 Regression Output 3 3 Fitted Values 5 4 Residuals 6 5 Functional Forms 8 Updated from Stata tutorials provided by Prof. Cichello
More informationGraduate Econometrics Lecture 4: Heteroskedasticity
Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model
More informationIntroduction to Estimation Methods for Time Series models. Lecture 1
Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation
More information7 Introduction to Time Series
Econ 495 - Econometric Review 1 7 Introduction to Time Series 7.1 Time Series vs. Cross-Sectional Data Time series data has a temporal ordering, unlike cross-section data, we will need to changes some
More informationEconomics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects
Economics 113 Simple Regression Models Simple Regression Assumptions Simple Regression Derivation Changing Units of Measurement Nonlinear effects OLS and unbiased estimates Variance of the OLS estimates
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More informationLecture 2 Multiple Regression and Tests
Lecture 2 and s Dr.ssa Rossella Iraci Capuccinello 2017-18 Simple Regression Model The random variable of interest, y, depends on a single factor, x 1i, and this is an exogenous variable. The true but
More informationIntermediate Econometrics
Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage
More informationStatistical Inference. Part IV. Statistical Inference
Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided
More informationEssential of Simple regression
Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship
More informationThe Statistical Property of Ordinary Least Squares
The Statistical Property of Ordinary Least Squares The linear equation, on which we apply the OLS is y t = X t β + u t Then, as we have derived, the OLS estimator is ˆβ = [ X T X] 1 X T y Then, substituting
More informationApplied Regression Analysis
Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More information