Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Size: px
Start display at page:

Download "Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018"

Transcription

1 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 1 / 33

2 The multiple linear regression model The major drawback of the bivariate regression model is that the key assumption SLR.4 (zero conditional mean) is often unrealistic. The multiple linear regression model (MLR) allows to control for many other factors which might otherwise be captured in the error term. Thus it is more amenable to ceteris paribus analysis. The model with k independent variables given a sample (y i, x i1,..., x ik ), i = 1,..., n, reads y i = β 0 + β 1 x i1 + β 2 x i2 + + β k x ik + u i (1) y i = x i β + u i with x i = (1, x i1,..., x ik ) and β = (β 0, β 1,..., β k ). The key assumption is E(u i x i ) = 0 i = 1,..., n (3) (2) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 2 / 33

3 MLR Matrix notation Let y be a n 1 vector of observations on y, let X be the data matrix with dimension n (k + 1) and associated parameter vector β R (k+1) 1, and let u be a n 1 vector of disturbances. With these ingredients we can write the model in (1) as y = Xβ + u (4) or more explicitly, y 1 1 x 11 x x 1k β 0 u 1 y 2. = 1 x 21 x x 2k β u 2. y n 1 x n1 x n2... x nk β k u n (5) where Xβ is the systematic and u the stochastic component. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 3 / 33

4 MLR An example We are interested in the effect of education on hourly wage: wage i = β 0 + β 1 educ i + β 2 exper i + u i, i = 1,..., n (6) We control for years of labor market experience (exper). We are still primarily interested in the effect of education. The MLR takes experience out of the error term u. With the SLR we would have to assume exper educ. 1 β 1 measures the effect of educ on wage when exper is held constant. β 2 measures the effect of exper on wage when educ is held constant. 1 denotes statistical independence. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 4 / 33

5 MLR Estimation As in Module 1, we are looking for a coefficient vector ˆβ R (k+1) 1 that minimizes the sum of squared residuals. Formally the problem reads arg min ˆβ n û 2 i = û û = (y X ˆβ) (y X ˆβ) (7) i=1 = y y ˆβ X y y X ˆβ + ˆβ X X ˆβ (8) = y y 2 ˆβ X y + ˆβ X X ˆβ (9) Note that the last step is possible because ˆβ X y = (y X ˆβ) = y X ˆβ. The first-order condition for minimization is û û ˆβ = 2X y + 2X X ˆβ = 0 (10) or written differently X X ˆβ = X y (11) which is called the system of least squares normal equations. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 5 / 33

6 MLR Estimation If X X is non-singular (i.e., there exists an inverse), pre-multiplying both sides of equation (11) with (X X) 1 yields the OLS estimators ˆβ: ˆβ = (X X) 1 X y (12) Important: The matrix X X is non-singular (= invertible) and ˆβ a unique solution to the minimization problem if and only if we have at least n k observations, and the data matrix X has rank (k + 1). The second point is violated if there are linear dependencies among the explanatory variables (i.e., perfect collinearity). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 6 / 33

7 MLR Properties of the OLS estimator The OLS estimator has various important properties that do not depend on any assumptions, but rather arise by how it is constructed. First, substitute y = X ˆβ + û into the system of normal equations (11) to obtain X X ˆβ = X y X X ˆβ = X (X ˆβ + û) X X ˆβ = X X ˆβ + X û 0 = X û (13) A number of important properties can be derived from this condition. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 7 / 33

8 MLR Properties of the OLS estimator 1 The observed values of X are uncorrelated with the residuals û. This follows immediately from (13): X û = 0 iff. X û. Note that this does not mean that X is uncorrelated with u! We have to assume this. 2 The sum of residuals is zero. If there is a constant, the first column of X will be a column of ones. For the first element in the X û to be zero it must hold that ûi = 0. i 3 The sample mean of the residuals is zero. This follows from the previous property: û = n 1 n û = i= The regression hyperplane passes through the means of the observed values X and ȳ. Recall that û = y X ˆβ. Dividing by n gives û = ȳ X ˆβ. From the previous property: û = ȳ X ˆβ = 0, so ȳ = X ˆβ. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 8 / 33

9 MLR Properties of the OLS estimator 5 The predicted values ŷ are uncorrelated with the residuals û. The predicted values are ŷ = X ˆβ. From this we have ŷ û = (X ˆβ) û = ˆβ X û = 0 because X û = 0. 6 The mean of the predicted Y s for the sample will equal the mean of the observed Y s, i.e. ŷ = ȳ. Proof is left as an exercise (use the result in item 4). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 9 / 33

10 MLR Properties of the OLS estimator These properties always hold true, be careful not to infer anything from the residuals about the actual disturbances! So far we know nothing about ˆβ except that it satisfies all of the properties discussed above. We need to make some assumptions about the true model in order to make any inferences regarding β (the true population parameters) from ˆβ (our estimator of the true parameters). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 10 / 33

11 MLR Expected value and variance The assumptions from the bivariate model translate to the multivariate case as follows: Assumption MLR.1 Linear in parameters The population model is linear: y = β X + u. Assumption MLR.2 Random sampling We have a random sample of n observations, {(y i, x i) i = 1,..., n}, that follows the population model in assumption MLR.1. Assumption MLR.3 No perfect collinearity The data matrix X has rank (k + 1). Assumption MLR.4 Zero conditional mean Conditional on the entire matrix X, each error u i has mean zero: E(u i X) = 0 i = 1,..., n. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 11 / 33

12 Finite sample properties Note that we only need assumption MLR.3 (no perfect collinearity) to obtain an OLS estimate ˆβ. Whether this estimate actually makes sense, i.e. is unbiased and representative for the full population, depends on the other assumptions. Especially the zero conditional mean assumption (MLR.4) often poses problems in practice. Requires that, conditional on the observed covariates x i, unobservables are on average orthogonal to the error term u i. Fails in the case of Simultaneity Selection Omitted variables Functional form misspecification Measurement error We will discuss sources and consequences of these cases in Module 6. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 12 / 33

13 Finite sample properties Expected values Theorem: Unbiasedness of OLS Under assumptions MLR.1 through MLR.4, E( ˆβ) = β. (14) In other words, ˆβ is an unbiased estimate for β. Proof. Rewrite the OLS estimator as ˆβ = (X X) 1 X y = (X X) 1 X (Xβ + u) = (X X) 1 (X X)β + (X X) 1 X u (15) Because X X is a square matrix, (X X) 1 (X X) = I, thus ˆβ = β + (X X) 1 X u (16) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 13 / 33

14 Finite sample properties Expected values Taking conditional expectations on both sides of equation (16) gives E( ˆβ X) = β + (X X) 1 X E(u X) (17) By MLR.4, E(u X) = 0, so This completes the proof. E( ˆβ X) = β (18) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 14 / 33

15 Finite sample properties Expected variances Assumption MLR.1 Linear in parameters The population model is linear: y = β X + u. Assumption MLR.2 Random sampling We have a random sample {(y i, x i) i = 1,..., n} that follows the population model: y i = β 0 + β 1x i + u i. Assumption MLR.3 No perfect collinearity The data matrix X has rank (k + 1). Assumption MLR.4 Zero conditional mean Conditional on the entire matrix X, each error u i has zero mean: E(u i X) = 0 i = 1,..., n. Assumption MLR.5 Homoskedasticity and no serial correlation The error u i has the same variance given any values of the covariates, i.e. Var(u i X) = σ 2, i = 1,..., n, and there is no serial correlation between the errors: Cov(u i, u j X) = 0 for all j i. We can write these two assumptions as Var(u X) = σ 2 I. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 15 / 33

16 Finite sample properties Expected variances Assumption MLR.5 requires that errors are homoskedastic and that there is no serial correlation (meaning that errors are not correlated across observations this is especially important if you deal with panel data, but sometimes also in cross-sectional settings). Combining these assumptions, we can write the variance-covariance matrix of the disturbances as σ Var(u X) = E(uu 0 σ X) = = σ2 I (19) σ 2 Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 16 / 33

17 Finite sample properties Expected variances Theorem: Variance-covariance matrix of the OLS estimator Under assumptions MLR.1 through MLR.5, Var( ˆβ) = σ 2 (X X) 1 (20) Proof. See Wooldridge (2013), p For one particular ˆβ j ˆβ, the variance is obtained by multiplying σ 2 by the jth diagonal element of (X X) 1. It can also be written as Var( ˆβ j ) = σ 2 SST j (1 R 2 j ) (21) where SST j = n i=1 (x ij x j ) 2 is the total sample variation in x j and R 2 j is the R-squared from regressing x j on all other independent variables (and including an intercept). Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 17 / 33

18 Finite sample properties Expected variances The unbiased estimator of the error variance in the multivariate case is given by ˆσ 2 = u u n k 1 (22) where u u is again the sum of squared residuals. Theorem: Unbiasedness of σ 2 Under assumptions MLR.1 through MLR.5, ˆσ 2 is an unbiased estimate for σ 2. That is, E(ˆσ 2 X) = σ 2 σ 2 > 0 (23) Proof. See Wooldridge (2013), p Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 18 / 33

19 Finite sample properties Gauss-Markov theorem Gauss-Markov Theorem Under assumptions MLR.1 through MLR.5, ˆβ is the best linear unbiased estimator. Proof. See Wooldridge (2013), p The Gauss-Markov theorem translated: OLS is the estimator with the smallest variance amongst all linear unbiased estimators. OLS is BLUE. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 19 / 33

20 Inference Sampling distributions of the OLS estimators Although it is not necessary for the Gauss-Markov theorem to hold, 2 we assume normally distributed disturbances to derive sampling distributions. Assumption MLR.6 Normality of errors Conditional on X, u is distributed as multivariate normal with mean zero and variance-covariance matrix σ 2 I. That is, u Normal(0, σ 2 I). (24) 2 We will show later that, as soon as asymptotics kick in, we don t need the normality assumption anymore for our test statistics to be valid. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 20 / 33

21 Inference Sampling distributions of the OLS estimators Theorem: Normality of ˆβ Under the classical linear model assumptions MLR.1 through MLR.6, ˆβ conditional on X is distributed as multivariate normal with mean β and variance-covariance matrix σ 2 X X 1. That is, ˆβ Normal(β, σ 2 (X X) 1 ) (25) Therefore, ˆβ j β j sd( ˆβ j ) Normal(0, 1) (26) Proof. Wooldridge (2013, p. 113) provides a sketch of the proof for (25). The result in (26) is straightforward; if we substract the mean from a normally distributed random variable and divide by its standard deviation, we get a standard normal variable with mean zero and a standard deviation of 1. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 21 / 33

22 Inference Sampling distributions of the OLS estimators Theorem: Distribution of t-statistics Under assumptions MLR.1 through MLR.6, Proof. Wooldridge (2013), p ˆβ j β j se( ˆβ t n k 1 (27) j ) }{{} t-statistic This is an important result for inference. It says that, when we estimate σ in sd( ˆβ j ) by ˆσ which yields se( ˆβ j ), ( ˆβ j β j )/se( ˆβ j ) is t-distributed with n k 1 degrees of freedom. Note that β j is some hypothesized value. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 22 / 33

23 Inference Testing a single population parameter Pick a significance level and formulate a null-hypothesis (H 0 ): One-sided alternatives: H 0 : β j 0; H 1 : β j > 0. Two-sided alternatives: H 0 : β j = 0; H 1 : β j 0. One-sided alternatives: H 0 : β j > α j; H 1 : β j α j, with α R. t-statistic: t (estimate hypothesized value), according to theorem (27). standard error p-value for t-test: what is the smallest significance level at which H 0 would be rejected? Confidence intervals range of likely values for β: CI ˆβ j ± c se( ˆβ j ) (28) where c is the 97.5 th percentile of the t n k 1 distribution. Economic vs. statistical significance. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 23 / 33

24 Inference Testing multiple linear restrictions The F -test allows to test for multiple hypotheses. Consider a model with k = 4 independent variables: y i = β 0 + β 1 x i1 + β 2 x i2 + β 3 x i3 + β 4 x i4 + u i. Suppose you want to test whether x 1, x 2, and x 3 are jointly insignificant Formulate a null-hypothesis: H 0 : β 1 = β 2 = β 3 = 0 H 1 : H 0 is not true. F -statistic: F (SSR r SSR ur )/q SSR ur /(n k 1) F q,n k 1 (29) where q is the number of restrictions, SSR r is the sum of squared residuals from the restricted model and SSR ur is the SSR from the unrestricted model. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 24 / 33

25 Inference Example We estimate the following model: final i = β 0 + β 1 attend i + β 2 hwrte i + u i (30) where final [10, 39] are final exam points, attend [2, 32] is the number of classes attended out of 32, and hwrte is the percentage of homeworks turned in times reg final attend hwrte Source SS df MS Number of obs = 674 F( 2, 671) = 9.20 Model Prob > F = Residual R-squared = Adj R-squared = Total Root MSE = final Coef. Std. Err. t P> t [95% Conf. Interval] attend hwrte _cons Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 25 / 33

26 Asymptotics So far, we have looked at finite sample properties of OLS, i.e., properties that hold independent of how large n is. However, it is also important to know large sample or asymptotic properties of OLS. These are defined as sample size grows without bound. An important result is that even without assuming normality (MLR.6), t and F statistics are approximately t and F distributed. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 26 / 33

27 Asymptotics Consistency In general, an estimator ˆθ is consistent, if it converges in probability 3 to the population parameter θ, that is, ˆθ p θ, or plim ˆθ = θ. Note that unbiasedness does not necessarily imply consistency, and consistency does not automatically imply unbiasedness. Consistency of OLS: OLS is unbiased under assumptions MLR.1 through MLR.4, so ˆβ j is always distributed around β j. The distribution of ˆβ j becomes more and more tightly distributed around β j as the sample size grows. As n, the distribution of ˆβ j collapses to a single point β j. 3 A random variable X n converges to X in probability if for some ε > 0, lim P( X n X ε) = 0 (31) n Note that here the probability converges, not the random variable itself. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 27 / 33

28 Asymptotics Consistency Figure: Sampling distributions of ˆβ 1 [Source: Wooldridge (2013), Figure 5.1]. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 28 / 33

29 Asymptotics Consistency Theorem: Consistency of OLS Under assumptions MLR.1 through MLR.4, the OLS estimator ˆβ is consistent for β. Proof. We show consistency for the bivariate case with one regressor β 1, the general proof for k regressors is given in Wooldridge (2013), p First note that we can also write the OLS estimator ˆβ 1 simply as ˆβ 1 = n i=1 x iy i n i=1 x2 i (32) and after plugging in y i = βx i + u i and some algebra, we obtain ˆβ 1 = β + n 1 n i=1 x iu i n 1 n i=1 x2 i (33) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 29 / 33

30 Asymptotics Consistency As n, we have ˆβ 1 p β + plim n 1 n i=1 x iu i plim n 1 n i=1 x2 i By the law of large numbers, 4 we have plim n 1 n i=1 x i = E(x i ) and plim n 1 n i=1 u i = E(u i ). Since we assume zero conditional mean (MLR.4), which obviously implies E(u i ) = 0, we get ˆβ 1 p β + 0 plim n 1 n i=1 x2 i (34) = β (36) This proofs that ˆβ 1 converges to β in probability. 4 Let X 1,..., X n be some sequence of i.i.d. random variables with arbitrary distribution. The law of large number states that X n = n 1 (X 1,..., X n) E(X n) (35) That is, the sample average converges to the expected value. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 30 / 33

31 Asymptotics Consistency Consistency is related to bias as follows: An estimator ˆθ is consistent iff. it converges to some value θ and the bias, i.e., Bias(θ) = E(ˆθ) θ, converges to zero. Individual estimators in the sequence ˆθ j ˆθ may be biased, but the overall sequence is still consistent if the bias converges to zero. Estimators can be Unbiased but not consistent Biased but consistent Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 31 / 33

32 Asymptotics Normality Another important large sample property of OLS is that ˆβ j is asymptotically normally distributed under assumptions MLR.1 through MLR.5. Formally, ˆβ j β se( ˆβ j ) where Var( ˆβ j ) is the usual OLS standard error. Result stems from the central limit theorem. 5 a Normal(0, 1) (37) We do not need the normality assumption in MLR.6 for our test statistics to be valid, as long as the sample size is reasonably large. All we have to assume is finite variance; Var(u) <. 5 The central limit theorem states that the standardized sums of i.i.d. random variables converges to a normal distribution, irrespective of their own distributions. Let X 1,..., X n be i.i.d. random variables with finite expected value µ and variance σ 2. Then ) ) 1 nσ ( n i=1 X i nµ = 1 n ( n i=1 X µ σ = X µ σ d n N(0, 1) (38) Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 32 / 33

33 Literature Main reference: Wooldridge, J. M. (2015). Introductory Econometrics: A Modern Approach, 5th ed., South Western College Publishing. Additional reference: Greene, W. H. (2012). Econometric Analysis, 7th edition, Pearson. Alexander Ahammer (JKU) Module 2: Multivariate Linear Regression 33 / 33

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018 Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate

More information

Statistical Inference with Regression Analysis

Statistical Inference with Regression Analysis Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Multiple Regression Analysis. Part III. Multiple Regression Analysis Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

The Simple Regression Model. Part II. The Simple Regression Model

The Simple Regression Model. Part II. The Simple Regression Model Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

THE MULTIVARIATE LINEAR REGRESSION MODEL

THE MULTIVARIATE LINEAR REGRESSION MODEL THE MULTIVARIATE LINEAR REGRESSION MODEL Why multiple regression analysis? Model with more than 1 independent variable: y 0 1x1 2x2 u It allows : -Controlling for other factors, and get a ceteris paribus

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

The general linear regression with k explanatory variables is just an extension of the simple regression as follows 3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because

More information

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1

coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1 Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple

More information

Practical Econometrics. for. Finance and Economics. (Econometrics 2)

Practical Econometrics. for. Finance and Economics. (Econometrics 2) Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Heteroskedasticity. Part VII. Heteroskedasticity

Heteroskedasticity. Part VII. Heteroskedasticity Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least

More information

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define

More information

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47

ECON2228 Notes 2. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 47 ECON2228 Notes 2 Christopher F Baum Boston College Economics 2014 2015 cfb (BC Econ) ECON2228 Notes 2 2014 2015 1 / 47 Chapter 2: The simple regression model Most of this course will be concerned with

More information

ECNS 561 Multiple Regression Analysis

ECNS 561 Multiple Regression Analysis ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

ECON3150/4150 Spring 2016

ECON3150/4150 Spring 2016 ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression

More information

α version (only brief introduction so far)

α version (only brief introduction so far) Econometrics I KS Module 8: Panel Data Econometrics Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: June 18, 2018 α version (only brief introduction so far) Alexander

More information

Lecture 8: Instrumental Variables Estimation

Lecture 8: Instrumental Variables Estimation Lecture Notes on Advanced Econometrics Lecture 8: Instrumental Variables Estimation Endogenous Variables Consider a population model: y α y + β + β x + β x +... + β x + u i i i i k ik i Takashi Yamano

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Simple Linear Regression: The Model

Simple Linear Regression: The Model Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 5 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 44 Outline of Lecture 5 Now that we know the sampling distribution

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 16, 2013 Outline Introduction Simple

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Lecture 4: Multivariate Regression, Part 2

Lecture 4: Multivariate Regression, Part 2 Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above

More information

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation Michele Aquaro University of Warwick This version: July 21, 2016 1 / 31 Reading material Textbook: Introductory

More information

Econometrics - 30C00200

Econometrics - 30C00200 Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business

More information

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in

More information

2. (3.5) (iii) Simply drop one of the independent variables, say leisure: GP A = β 0 + β 1 study + β 2 sleep + β 3 work + u.

2. (3.5) (iii) Simply drop one of the independent variables, say leisure: GP A = β 0 + β 1 study + β 2 sleep + β 3 work + u. BOSTON COLLEGE Department of Economics EC 228 Econometrics, Prof. Baum, Ms. Yu, Fall 2003 Problem Set 3 Solutions Problem sets should be your own work. You may work together with classmates, but if you

More information

Linear Regression. Junhui Qian. October 27, 2014

Linear Regression. Junhui Qian. October 27, 2014 Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests

ECON Introductory Econometrics. Lecture 7: OLS with Multiple Regressors Hypotheses tests ECON4150 - Introductory Econometrics Lecture 7: OLS with Multiple Regressors Hypotheses tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 7 Lecture outline 2 Hypothesis test for single

More information

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11

1 Linear Regression Analysis The Mincer Wage Equation Data Econometric Model Estimation... 11 Econ 495 - Econometric Review 1 Contents 1 Linear Regression Analysis 4 1.1 The Mincer Wage Equation................. 4 1.2 Data............................. 6 1.3 Econometric Model.....................

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

INTRODUCTION TO BASIC LINEAR REGRESSION MODEL

INTRODUCTION TO BASIC LINEAR REGRESSION MODEL INTRODUCTION TO BASIC LINEAR REGRESSION MODEL 13 September 2011 Yogyakarta, Indonesia Cosimo Beverelli (World Trade Organization) 1 LINEAR REGRESSION MODEL In general, regression models estimate the effect

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

Lecture 3: Multivariate Regression

Lecture 3: Multivariate Regression Lecture 3: Multivariate Regression Rates, cont. Two weeks ago, we modeled state homicide rates as being dependent on one variable: poverty. In reality, we know that state homicide rates depend on numerous

More information

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one

More information

Lab 07 Introduction to Econometrics

Lab 07 Introduction to Econometrics Lab 07 Introduction to Econometrics Learning outcomes for this lab: Introduce the different typologies of data and the econometric models that can be used Understand the rationale behind econometrics Understand

More information

ECO220Y Simple Regression: Testing the Slope

ECO220Y Simple Regression: Testing the Slope ECO220Y Simple Regression: Testing the Slope Readings: Chapter 18 (Sections 18.3-18.5) Winter 2012 Lecture 19 (Winter 2012) Simple Regression Lecture 19 1 / 32 Simple Regression Model y i = β 0 + β 1 x

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 2 Jakub Mućk Econometrics of Panel Data Meeting # 2 1 / 26 Outline 1 Fixed effects model The Least Squares Dummy Variable Estimator The Fixed Effect (Within

More information

Instrumental Variables, Simultaneous and Systems of Equations

Instrumental Variables, Simultaneous and Systems of Equations Chapter 6 Instrumental Variables, Simultaneous and Systems of Equations 61 Instrumental variables In the linear regression model y i = x iβ + ε i (61) we have been assuming that bf x i and ε i are uncorrelated

More information

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem -

Section I. Define or explain the following terms (3 points each) 1. centered vs. uncentered 2 R - 2. Frisch theorem - First Exam: Economics 388, Econometrics Spring 006 in R. Butler s class YOUR NAME: Section I (30 points) Questions 1-10 (3 points each) Section II (40 points) Questions 11-15 (10 points each) Section III

More information

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b. B203: Quantitative Methods Answer all questions from part I. Answer two question from part II.a, and one question from part II.b. Part I: Compulsory Questions. Answer all questions. Each question carries

More information

Multiple Regression Analysis: Heteroskedasticity

Multiple Regression Analysis: Heteroskedasticity Multiple Regression Analysis: Heteroskedasticity y = β 0 + β 1 x 1 + β x +... β k x k + u Read chapter 8. EE45 -Chaiyuth Punyasavatsut 1 topics 8.1 Heteroskedasticity and OLS 8. Robust estimation 8.3 Testing

More information

Advanced Quantitative Methods: ordinary least squares

Advanced Quantitative Methods: ordinary least squares Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the

More information

Multiple Regression: Inference

Multiple Regression: Inference Multiple Regression: Inference The t-test: is ˆ j big and precise enough? We test the null hypothesis: H 0 : β j =0; i.e. test that x j has no effect on y once the other explanatory variables are controlled

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 17, 2012 Outline Heteroskedasticity

More information

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors ECON4150 - Introductory Econometrics Lecture 6: OLS with Multiple Regressors Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 6 Lecture outline 2 Violation of first Least Squares assumption

More information

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators

Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE. Sampling Distributions of OLS Estimators 1 2 Multiple Regression Analysis: Inference MULTIPLE REGRESSION ANALYSIS: INFERENCE Hüseyin Taştan 1 1 Yıldız Technical University Department of Economics These presentation notes are based on Introductory

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econ 510 B. Brown Spring 2014 Final Exam Answers Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory

More information

Simple Linear Regression Model & Introduction to. OLS Estimation

Simple Linear Regression Model & Introduction to. OLS Estimation Inside ECOOMICS Introduction to Econometrics Simple Linear Regression Model & Introduction to Introduction OLS Estimation We are interested in a model that explains a variable y in terms of other variables

More information

Course Econometrics I

Course Econometrics I Course Econometrics I 4. Heteroskedasticity Martin Halla Johannes Kepler University of Linz Department of Economics Last update: May 6, 2014 Martin Halla CS Econometrics I 4 1/31 Our agenda for today Consequences

More information

1 Independent Practice: Hypothesis tests for one parameter:

1 Independent Practice: Hypothesis tests for one parameter: 1 Independent Practice: Hypothesis tests for one parameter: Data from the Indian DHS survey from 2006 includes a measure of autonomy of the women surveyed (a scale from 0-10, 10 being the most autonomous)

More information

ECON Introductory Econometrics. Lecture 16: Instrumental variables

ECON Introductory Econometrics. Lecture 16: Instrumental variables ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental

More information

ECON Introductory Econometrics. Lecture 13: Internal and external validity

ECON Introductory Econometrics. Lecture 13: Internal and external validity ECON4150 - Introductory Econometrics Lecture 13: Internal and external validity Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 9 Lecture outline 2 Definitions of internal and external

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 1 Jakub Mućk Econometrics of Panel Data Meeting # 1 1 / 31 Outline 1 Course outline 2 Panel data Advantages of Panel Data Limitations of Panel Data 3 Pooled

More information

Empirical Application of Simple Regression (Chapter 2)

Empirical Application of Simple Regression (Chapter 2) Empirical Application of Simple Regression (Chapter 2) 1. The data file is House Data, which can be downloaded from my webpage. 2. Use stata menu File Import Excel Spreadsheet to read the data. Don t forget

More information

The Simple Regression Model. Simple Regression Model 1

The Simple Regression Model. Simple Regression Model 1 The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising

More information

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage

More information

Econometrics Master in Business and Quantitative Methods

Econometrics Master in Business and Quantitative Methods Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid Models with discrete dependent variables and applications of panel data methods in all fields of economics

More information

Econometrics Midterm Examination Answers

Econometrics Midterm Examination Answers Econometrics Midterm Examination Answers March 4, 204. Question (35 points) Answer the following short questions. (i) De ne what is an unbiased estimator. Show that X is an unbiased estimator for E(X i

More information

Heteroskedasticity and Autocorrelation

Heteroskedasticity and Autocorrelation Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Multiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor)

Multiple Regression Analysis: Estimation. Simple linear regression model: an intercept and one explanatory variable (regressor) 1 Multiple Regression Analysis: Estimation Simple linear regression model: an intercept and one explanatory variable (regressor) Y i = β 0 + β 1 X i + u i, i = 1,2,, n Multiple linear regression model:

More information

Regression #8: Loose Ends

Regression #8: Loose Ends Regression #8: Loose Ends Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #8 1 / 30 In this lecture we investigate a variety of topics that you are probably familiar with, but need to touch

More information

Empirical Economic Research, Part II

Empirical Economic Research, Part II Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction

More information

Advanced Econometrics

Advanced Econometrics Based on the textbook by Verbeek: A Guide to Modern Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna May 16, 2013 Outline Univariate

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

Econometrics Homework 1

Econometrics Homework 1 Econometrics Homework Due Date: March, 24. by This problem set includes questions for Lecture -4 covered before midterm exam. Question Let z be a random column vector of size 3 : z = @ (a) Write out z

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 11, 2012 Outline Heteroskedasticity

More information

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression

More information

Outline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section

Outline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section Outline I. The Nature of Time Series Data 11. Time Series Analysis II. Examples of Time Series Models IV. Functional Form, Dummy Variables, and Index Basic Regression Numbers Read Wooldridge (2013), Chapter

More information

Lab 6 - Simple Regression

Lab 6 - Simple Regression Lab 6 - Simple Regression Spring 2017 Contents 1 Thinking About Regression 2 2 Regression Output 3 3 Fitted Values 5 4 Residuals 6 5 Functional Forms 8 Updated from Stata tutorials provided by Prof. Cichello

More information

Graduate Econometrics Lecture 4: Heteroskedasticity

Graduate Econometrics Lecture 4: Heteroskedasticity Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model

More information

Introduction to Estimation Methods for Time Series models. Lecture 1

Introduction to Estimation Methods for Time Series models. Lecture 1 Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation

More information

7 Introduction to Time Series

7 Introduction to Time Series Econ 495 - Econometric Review 1 7 Introduction to Time Series 7.1 Time Series vs. Cross-Sectional Data Time series data has a temporal ordering, unlike cross-section data, we will need to changes some

More information

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects Economics 113 Simple Regression Models Simple Regression Assumptions Simple Regression Derivation Changing Units of Measurement Nonlinear effects OLS and unbiased estimates Variance of the OLS estimates

More information

Regression #3: Properties of OLS Estimator

Regression #3: Properties of OLS Estimator Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with

More information

Lecture 2 Multiple Regression and Tests

Lecture 2 Multiple Regression and Tests Lecture 2 and s Dr.ssa Rossella Iraci Capuccinello 2017-18 Simple Regression Model The random variable of interest, y, depends on a single factor, x 1i, and this is an exogenous variable. The true but

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage

More information

Statistical Inference. Part IV. Statistical Inference

Statistical Inference. Part IV. Statistical Inference Part IV Statistical Inference As of Oct 5, 2017 Sampling Distributions of the OLS Estimator 1 Statistical Inference Sampling Distributions of the OLS Estimator Testing Against One-Sided Alternatives Two-Sided

More information

Essential of Simple regression

Essential of Simple regression Essential of Simple regression We use simple regression when we are interested in the relationship between two variables (e.g., x is class size, and y is student s GPA). For simplicity we assume the relationship

More information

The Statistical Property of Ordinary Least Squares

The Statistical Property of Ordinary Least Squares The Statistical Property of Ordinary Least Squares The linear equation, on which we apply the OLS is y t = X t β + u t Then, as we have derived, the OLS estimator is ˆβ = [ X T X] 1 X T y Then, substituting

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information