Advanced Quantitative Methods: ordinary least squares

Size: px
Start display at page:

Download "Advanced Quantitative Methods: ordinary least squares"

Transcription

1 Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012

2

3 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the independent variables also known as explanatory variables also known as regressors or predictors (or factors, carriers) X is sometimes called the design matrix (or factor space) y is regressed on X The error term ε is sometimes called a disturbance: ε = y Xβ. The difference between the observed and predicted dependent variable is called the residual: e = y ŷ = y Xˆβ.

4 y i = β 0 +β 1 x 1 +β 2 x 2 + +ε i y = Xβ +ε ε N(0,σ 2 ) y N(Xβ,σ 2 )

5 Components Two components of the model: y N(µ,σ 2 ) Stochastic µ = Xβ Systematic Generalised version (not necessarily linear): y f(µ,α) Stochastic µ = g(x,β) Systematic (King 1998, 8)

6 Components y f(µ,α) Stochastic µ = g(x,β) Systematic Stochastic component: varies over repeated (hypothetical) observations on the same unit. Systematic component: varies across units, but constant given X. (King 1998, 8)

7 Uncertainty y f(µ,α) Stochastic µ = g(x,β) Systematic Two types of uncertainty: Estimation uncertainty: lack of knowledge about α and β; can be reduced by increasing n. Fundamental uncertainty: represented by stochastic component and exists independent of researcher.

8 Ordinary Least Squares (OLS) For the linear model, the most popular method of estimation is (OLS). ˆβ OLS are those estimates of β that minimize the sum of squared residuals: e e. OLS is the best linear unbiased estimator (BLUE).

9 Assumptions: specification Linear in parameters (i.e. f(xβ) = Xβ and E(y) = Xβ) No extraneous variables in X No omitted independent variables Parameters to be estimated are constant Number of parameters is less than the number of cases, k < n Note that this does not imply that you cannot include non-linearly transformed variables, e.g. y i = β 0 +β 1 x i +β 2 xi 2 +ε i can be estimated with OLS.

10 Assumptions: errors Errors have an expected value of zero, E(ε X) = 0 Errors are normally distributed, ε N(0,σ 2 ) Errors have a constant variance, var(ε X) = σ 2 < Errors are not autocorrelated, cov(ε i,ε j X) = 0 i j Errors and X are uncorrelated, cov(x,ε) = 0

11 Assumptions: regressors X varies X is of full column rank (note: requires k < n) No measurement error in X No endogenous variables in X

12 y = Xβ +ε y 1 y 2. y n n 1 1 x 12 x 13 x 1k 1 x 22 x 23 x 2k = x n2 x n3 x nk n k β 1 β 2. β k k 1 ε 1 ε 2 +. ε n n 1 y 1 y 2. y n n 1 β 1 +β 2 x 12 +β 3 x β k x 1k β 1 +β 2 x 22 +β 3 x β k x 2k =. β 1 +β 2 x n2 +β 3 x n3 + +β k x nk n 1 ε 1 ε 2 +. ε n n 1

13 Deriving ˆβ OLS y = Xβ +ε y = Xˆβ +e ˆβ OLS = argmin ˆβ e e = argmin(y Xˆβ) (y Xˆβ) ˆβ

14 Deriving ˆβ OLS e e = (y Xˆβ) (y Xˆβ) = (y (Xˆβ) )(y Xˆβ) = y y (Xˆβ) y y Xˆβ +(Xˆβ) Xˆβ = y y ˆβ X y y Xˆβ + ˆβ X Xˆβ = y y 2y Xˆβ + ˆβ X Xˆβ

15 Deriving ˆβ OLS ˆβ OLS = argmin ˆβ (e e) ˆβ OLS = 0 e e = (y y 2y Xˆβ + ˆβ X Xˆβ) ˆβ OLS = 0

16 Deriving ˆβ OLS (y y 2y Xˆβ + ˆβ X Xˆβ) ˆβ OLS = 0 2X Xˆβ OLS 2X y = 0 2X Xˆβ OLS = 2X y X Xˆβ OLS = X y (X X) 1 X Xˆβ OLS = (X X) 1 X y ˆβ OLS = (X X) 1 X y

17 X Xˆβ OLS = X y 1 x 12 x 13 x 1k 1 x 22 x 23 x 2k x n2 x n3 x nk 1 x 12 x 13 x 1k 1 x 22 x 23 x 2k x n2 x n3 x nk 1 x 12 x 13 x 1k 1 x 22 x 23 x 2k = x n2 x n3 x nk n k n k n k ˆβ 1 ˆβ 2. ˆβ k y 1 y 2. y n k 1 n 1 (ˆβ here refers to OLS estimates.)

18 X Xˆβ OLS = X y x 12 x 22 x n2 x 13 x 23 x n x 1k x 2k x nk k n 1 x 12 x 13 x 1k 1 x 22 x 23 x 2k x n2 x n3 x nk x 12 x 22 x n2 = x 13 x 23 x n x 1k x 2k x nk n k k n ˆβ 1 ˆβ 2. ˆβ k y 1 y 2. y n k 1 n 1

19 X Xˆβ OLS = X y n xi2 xik xi2 (xi2 ) 2 xi2 x ik xi3 xi3 x i2 xi3 x ik xik xik x i2 (xik ) 2 k k ˆβ 1 ˆβ 2 ˆβ 3. ˆβ k k 1 yi xi2 y i xi3 = y i. xik y i k 1 ( refers to n i.)

20 X Xˆβ OLS = X y So this can be seen as a set of linear equations to solve: ˆβ 1 n+ ˆβ 2 xi2 + + ˆβ k xik = y i ˆβ 1 xi2 + ˆβ 2 (xi2 ) ˆβ k xi2 x ik = x i2 y i ˆβ 1 xi3 + ˆβ 2 xi3 x i2 + + ˆβ k xi3 x ik = x i3 y i ˆβ 1 xik + ˆβ 2 xik x i2 + + ˆβ k (xik ) 2 = x ik y i.

21 X Xˆβ OLS = X y When there is only one independent variable, this reduces to: ˆβ 1 n+ ˆβ 2 xi = y i ˆβ 1 xi + ˆβ 2 x 2 i = x i y i ˆβ 1 = yi ˆβ 2 xi n = nȳ ˆβ 2 n x n (ȳ ˆβ 2 x) x i + ˆβ 2 x 2 i = x i y i = ȳ ˆβ 2 x nȳ x ˆβ 2 n x 2 + ˆβ 2 x 2 i = x i y i xi y i nȳ x ˆβ 2 = x 2 i n x 2

22 Simple regression y i = β 0 +β 1 x i +ε i If β 1 = 0 and the only regressor is the intercept, then ˆβ 0 = ȳ. If β 0 = 0, so that there is no intercept and one explanatory variable x, then ˆβ 1 =. n i x i y i n i x 2 i If there is an intercept and one explanatory variable, then ˆβ 1 = n i (x i x)(y i ȳ) n i (x i x) 2

23 Demeaning If the observations are expressed as deviations from their means, i.e.: y i = y i ȳ x i = x i x, then ˆβ 1 = n i x i y n i x 2 i i The intercept can be estimated as ȳ ˆβ 1 x.

24 Standardized variables The coefficient on a variable is interpreted as the effect on y given a one unit increase in x and thus the interpretation is dependent on the scale of measurement. An option can be to standardize the variables: y = y ȳ σy 2 x = x x σ 2 x y = X β +ε and interpret the coefficients as the effect on y expressed in standard deviations as x increase by one standard deviation.

25 Standardized variables library(arm) summary(standardize(lm(y ~ x)), binary.inputs="leave.alone") See also Andrew Gelman (2007), Scaling regression inputs by dividing by two standard deviations.

26 Exercise: teenage gambling library(faraway) data(teengamb) summary(teengamb) Using matrix formulas, 1 regress gamble on a constant 2 regress gamble on sex 3 regress gamble on sex, status, income, verbal

27 Exercise: teenage gambling 1 Regress gamble on sex, status, income, verbal 2 Which observation has the largest residual? 3 Compute mean and median of residuals 4 Compute correlation between residuals and fitted values 5 Compute correlation between residuals and income 6 All other predictors held constant, what would be the difference in predicted expenditure on gambling between males and females?

28 Deriving var(ˆβ OLS ) var(ˆβ OLS ) = E[(ˆβ OLS E(ˆβ OLS ))(ˆβ OLS E(ˆβ OLS )) ] = E[(ˆβ OLS β)(ˆβ OLS β) ]

29 Deriving var(ˆβ OLS ) ˆβ OLS β = (X X) 1 X y β = (X X) 1 X (Xβ +ε) β = (X X) 1 X Xβ +(X X) 1 X ε β = β +(X X) 1 X ε β = (X X) 1 X ε

30 Deriving var(ˆβ OLS ) var(ˆβ OLS ) = E[(ˆβ OLS β)(ˆβ OLS β) ] = E[((X X) 1 X ε)((x X) 1 X ε) ] = E[(X X) 1 X εε X(X X) 1 ] = (X X) 1 X E[εε ]X(X X) 1 = (X X) 1 X σ 2 IX(X X) 1 = σ 2 (X X) 1 X X(X X) 1 = σ 2 (X X) 1

31 Standard errors var(ˆβ OLS ) = σ 2 (X X) 1 The standard errors of ˆβ OLS are then the square root of the diagonal of this matrix. In the simple case where y = β 0 +β 1 x, this gives var(ˆβ 1 ) = σ 2 n i (x i x) 2 Note how an increase in variance in x leads a decrease in the standard error of ˆβ 1.

32 e = y Xˆβ OLS = y X(X X) 1 X y = (I X(X X) 1 X )y = My = M(Xβ +ε) = (I X(X X) 1 X )Xβ +Mε = Xβ X(X X) 1 X Xβ +Mε = Xβ Xβ +Mε = Mε

33 e e = (Mε) (Mε) = ε M Mε = ε Mε E[e e X] = E[ε Mε X] ε Mε is a scalar, so if you see it as a 1 1 matrix, it is equal to its trace, therefore E[e e X] = E[tr(ε Mε X] = E[tr(Mεε ) X] = tr(me[εε X]) = tr(mσ 2 I) = σ 2 tr(m)

34 tr(m) = tr(i X(X X) 1 X ) = tr(i) tr(x(x X) 1 X ) = n tr(x X(X X) 1 ) = n k E(e e X) = (n k)σ 2 ˆσ 2 = e e n k

35 OLS in R n <- dim(x)[1] k <- dim(x)[2] b.hat <- solve(t(x) %*% X) %*% t(x) %*% y e <- y - X %*% b.hat s2.hat <- 1/(n-k) * t(e) %*% e v.hat <- s2.hat * solve(t(x) %*% X) Or: summary(lm(y ~ X))

36 Exercise: US wages library(faraway) data(uswages) summary(uswages) 1 Using matrix formulas, regress wage on educ, exper and race. 2 Interpret the results 3 Plot residuals against fitted values and against educ 4 Repeat with log(wage) as dependent variable

37 Unbiasedness of ˆβ OLS Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS From deriving var(ˆβ OLS ) we have: ˆβ OLS β = (X X) 1 X y β = (X X) 1 X (Xβ +ε) β = (X X) 1 X Xβ +(X X) 1 X ε β = β +(X X) 1 X ε β = (X X) 1 X ε Since we assume E(ε) = 0: E(ˆβ OLS β) = (X X) 1 X E(ε) = 0 Therefore, ˆβ OLS is an unbiased estimator of β.

38 Efficiency of ˆβ OLS Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS The Gauss-Markov theorem states that there is no linear unbiased estimator of β that has a smaller sampling variance than ˆβ OLS, i.e. ˆβ OLS is BLUE. An estimator is linear iff it can be expressed as a linear function of the data on the dependent variable: ˆβlinear j = n i f(x ij )y i, which is the case for OLS: ˆβ OLS = (X X) 1 X y = Cy (Wooldridge, pp , )

39 Gauss-Markov Theorem Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS ˆβ OLS = (X X) 1 X y = Cy Imagine we have another linear estimator: ˆβ = Wy, where W = C+D. ˆβ = (C+D)y = Cy+Dy = ˆβ OLS +D(Xβ +ε) = ˆβ OLS +DXβ +Dε

40 Gauss-Markov Theorem Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS E(ˆβ ) = E(ˆβ OLS +DXβ +Dε) = E(ˆβ OLS )+DXβ +E(Dε) = β +DXβ +0 DXβ = 0 DX = 0, because both ˆβ OLS and ˆβ are unbiased, so E(ˆβ OLS ) = E(ˆβ ) = β. (Hayashi, pp )

41 Gauss-Markov Theorem Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS ˆβ = ˆβ OLS +Dε ˆβ β = ˆβ OLS +Dε β = (C+D)ε var(ˆβ ) = E(ˆβ β)e(ˆβ β) = E((C+D)ε)E((C+D)ε) = (C+D)E(εε )(C+D) = σ 2 (C+D)(C+D) = σ 2 (CC +DC +CD +DD ) = σ 2 ((X X) 1 +DD ) σ 2 (X X) 1

42 Consistency of ˆβ OLS Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS ˆβ OLS = (X X) 1 X y = (X X) 1 X (Xβ +ε) = β +(X X) 1 X ε = β +( 1 n X X) 1 ( 1 n X ε) 1 lim n n X X = Q = lim (1 n n X X) 1 = Q 1 E( 1 n X ε) = E( 1 n n x i ε i ) = 0 i

43 Consistency of ˆβ OLS Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS var( 1 n X ε) = E( 1 n X ε( 1 n X ε) ) = 1 n X E(εε )X 1 n = ( σ2 X n )(X n ) lim n var(1 n X ε) = 0Q = 0

44 Consistency of ˆβ OLS Unbiasedness of ˆβ OLS Efficiency of ˆβ OLS Consistency of ˆβ OLS ˆβ OLS = β +( 1 n X X) 1 ( 1 n X ε) lim (1 n n X X) 1 = Q 1 E( 1 n X ε) = 0 lim var(1 n n X ε) = 0 imply that the sampling distribution of ˆβ OLS collapses to β as n becomes very large, i.e.: plim ˆβ OLS = β n

45 Sums of squares SST Total sum of squares (y i ȳ) 2 SSE Explained sum of squares (ŷ i ȳ) 2 SSR Residual sum of squares ei 2 = (ŷ i y i ) 2 = e e The key to remember is that SST = SSE + SSR Sometimes instead of explained and residual, regression and error are used, respectively, so that the abbreviations are swapped (!).

46 Defined in terms of sums of squares: = SSE SST = 1 SSR SST (yi ŷ i ) 2 = 1 (yi ȳ) 2 Interpretation: the proportion of the variation in y that is explained linearly by the independent variables A much over-used statistic: it may not be what we are interested in at all

47 in matrix algebra y y = (ŷ+e) (ŷ+e) = ŷ ŷ+ŷ e+e ŷ+e e = ŷ ŷ+2ˆβ X e+e e = ŷ ŷ+e e = 1 e e y y = ŷ ŷ y y (Hayashi, p. 20)

48 When a model has no intercept, it is possible for to lie outside the interval (0, 1) rises with the addition of more explanatory variables. For this reason we often report the adjusted : 1 (1 ) n 1 n k The values from different y samples cannot be compared

49 Exercise: US wages library(faraway) data(uswages) summary(uswages) 1 Using matrix formulas, regress wage on educ, exper and race. 2 What proportion of the variance in wage is explained by these three variables?

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

The Simple Regression Model. Part II. The Simple Regression Model

The Simple Regression Model. Part II. The Simple Regression Model Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Multiple Regression Analysis. Part III. Multiple Regression Analysis Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant

More information

Introduction to Estimation Methods for Time Series models. Lecture 1

Introduction to Estimation Methods for Time Series models. Lecture 1 Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation

More information

1. The OLS Estimator. 1.1 Population model and notation

1. The OLS Estimator. 1.1 Population model and notation 1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology

More information

Quantitative Methods I: Regression diagnostics

Quantitative Methods I: Regression diagnostics Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear

More information

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

The general linear regression with k explanatory variables is just an extension of the simple regression as follows 3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because

More information

STAT5044: Regression and Anova. Inyoung Kim

STAT5044: Regression and Anova. Inyoung Kim STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

Lecture Notes on Different Aspects of Regression Analysis

Lecture Notes on Different Aspects of Regression Analysis Andreas Groll WS 2012/2013 Lecture Notes on Different Aspects of Regression Analysis Department of Mathematics, Workgroup Financial Mathematics, Ludwig-Maximilians-University Munich, Theresienstr. 39,

More information

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Please answer all of the questions and show your work Clearly indicate your final answer to each question If you think a question is

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information

Econometrics Master in Business and Quantitative Methods

Econometrics Master in Business and Quantitative Methods Econometrics Master in Business and Quantitative Methods Helena Veiga Universidad Carlos III de Madrid Models with discrete dependent variables and applications of panel data methods in all fields of economics

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

REGRESSION ANALYSIS AND INDICATOR VARIABLES

REGRESSION ANALYSIS AND INDICATOR VARIABLES REGRESSION ANALYSIS AND INDICATOR VARIABLES Thesis Submitted in partial fulfillment of the requirements for the award of degree of Masters of Science in Mathematics and Computing Submitted by Sweety Arora

More information

The Multiple Regression Model Estimation

The Multiple Regression Model Estimation Lesson 5 The Multiple Regression Model Estimation Pilar González and Susan Orbe Dpt Applied Econometrics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 5 Regression model:

More information

EC3062 ECONOMETRICS. THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation. (1) y = β 0 + β 1 x β k x k + ε,

EC3062 ECONOMETRICS. THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation. (1) y = β 0 + β 1 x β k x k + ε, THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation (1) y = β 0 + β 1 x 1 + + β k x k + ε, which can be written in the following form: (2) y 1 y 2.. y T = 1 x 11... x 1k 1

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 16, 2013 Outline Introduction Simple

More information

The Aitken Model. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 41

The Aitken Model. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 41 The Aitken Model Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 41 The Aitken Model (AM): Suppose where y = Xβ + ε, E(ε) = 0 and Var(ε) = σ 2 V for some σ 2 > 0 and some known

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Linear Regression. Junhui Qian. October 27, 2014

Linear Regression. Junhui Qian. October 27, 2014 Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent

More information

Simple Linear Regression: The Model

Simple Linear Regression: The Model Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random

More information

ECON3150/4150 Spring 2015

ECON3150/4150 Spring 2015 ECON3150/4150 Spring 2015 Lecture 3&4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo January 29, 2015 1 / 67 Chapter 4 in S&W Section 17.1 in S&W (extended OLS assumptions) 2

More information

Multiple Regression Model: I

Multiple Regression Model: I Multiple Regression Model: I Suppose the data are generated according to y i 1 x i1 2 x i2 K x ik u i i 1...n Define y 1 x 11 x 1K 1 u 1 y y n X x n1 x nk K u u n So y n, X nxk, K, u n Rks: In many applications,

More information

Measuring the fit of the model - SSR

Measuring the fit of the model - SSR Measuring the fit of the model - SSR Once we ve determined our estimated regression line, we d like to know how well the model fits. How far/close are the observations to the fitted line? One way to do

More information

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

Chapter 2 The Simple Linear Regression Model: Specification and Estimation Chapter The Simple Linear Regression Model: Specification and Estimation Page 1 Chapter Contents.1 An Economic Model. An Econometric Model.3 Estimating the Regression Parameters.4 Assessing the Least Squares

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Reference: Davidson and MacKinnon Ch 2. In particular page

Reference: Davidson and MacKinnon Ch 2. In particular page RNy, econ460 autumn 03 Lecture note Reference: Davidson and MacKinnon Ch. In particular page 57-8. Projection matrices The matrix M I X(X X) X () is often called the residual maker. That nickname is easy

More information

FIRST MIDTERM EXAM ECON 7801 SPRING 2001

FIRST MIDTERM EXAM ECON 7801 SPRING 2001 FIRST MIDTERM EXAM ECON 780 SPRING 200 ECONOMICS DEPARTMENT, UNIVERSITY OF UTAH Problem 2 points Let y be a n-vector (It may be a vector of observations of a random variable y, but it does not matter how

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage

More information

Sample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them.

Sample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them. Sample Problems 1. True or False Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them. (a) The sample average of estimated residuals

More information

Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = The errors are uncorrelated with common variance:

Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = The errors are uncorrelated with common variance: 8. PROPERTIES OF LEAST SQUARES ESTIMATES 1 Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = 0. 2. The errors are uncorrelated with common variance: These assumptions

More information

Applied Econometrics (QEM)

Applied Econometrics (QEM) Applied Econometrics (QEM) The Simple Linear Regression Model based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #2 The Simple

More information

2. Regression Review

2. Regression Review 2. Regression Review 2.1 The Regression Model The general form of the regression model y t = f(x t, β) + ε t where x t = (x t1,, x tp ), β = (β 1,..., β m ). ε t is a random variable, Eε t = 0, Var(ε t

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

Linear Regression Model. Badr Missaoui

Linear Regression Model. Badr Missaoui Linear Regression Model Badr Missaoui Introduction What is this course about? It is a course on applied statistics. It comprises 2 hours lectures each week and 1 hour lab sessions/tutorials. We will focus

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA March 6, 2017 KC Border Linear Regression II March 6, 2017 1 / 44 1 OLS estimator 2 Restricted regression 3 Errors in variables 4

More information

The Simple Regression Model. Simple Regression Model 1

The Simple Regression Model. Simple Regression Model 1 The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising

More information

Econometrics I Lecture 3: The Simple Linear Regression Model

Econometrics I Lecture 3: The Simple Linear Regression Model Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating

More information

The Linear Regression Model

The Linear Regression Model The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general

More information

Econ 583 Final Exam Fall 2008

Econ 583 Final Exam Fall 2008 Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random

More information

Non-Spherical Errors

Non-Spherical Errors Non-Spherical Errors Krishna Pendakur February 15, 2016 1 Efficient OLS 1. Consider the model Y = Xβ + ε E [X ε = 0 K E [εε = Ω = σ 2 I N. 2. Consider the estimated OLS parameter vector ˆβ OLS = (X X)

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Introductory Econometrics

Introductory Econometrics Introductory Econometrics Violation of basic assumptions Heteroskedasticity Barbara Pertold-Gebicka CERGE-EI 16 November 010 OLS assumptions 1. Disturbances are random variables drawn from a normal distribution.

More information

Lecture 14 Simple Linear Regression

Lecture 14 Simple Linear Regression Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of

More information

Chapter 2 Multiple Regression I (Part 1)

Chapter 2 Multiple Regression I (Part 1) Chapter 2 Multiple Regression I (Part 1) 1 Regression several predictor variables The response Y depends on several predictor variables X 1,, X p response {}}{ Y predictor variables {}}{ X 1, X 2,, X p

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,

More information

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors: Wooldridge, Introductory Econometrics, d ed. Chapter 3: Multiple regression analysis: Estimation In multiple regression analysis, we extend the simple (two-variable) regression model to consider the possibility

More information

Ordinary Least Squares Regression

Ordinary Least Squares Regression Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section

More information

TMA4255 Applied Statistics V2016 (5)

TMA4255 Applied Statistics V2016 (5) TMA4255 Applied Statistics V2016 (5) Part 2: Regression Simple linear regression [11.1-11.4] Sum of squares [11.5] Anna Marie Holand To be lectured: January 26, 2016 wiki.math.ntnu.no/tma4255/2016v/start

More information

4 Multiple Linear Regression

4 Multiple Linear Regression 4 Multiple Linear Regression 4. The Model Definition 4.. random variable Y fits a Multiple Linear Regression Model, iff there exist β, β,..., β k R so that for all (x, x 2,..., x k ) R k where ε N (, σ

More information

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A ) Econ 60 Matrix Differentiation Let a and x are k vectors and A is an k k matrix. a x a x = a = a x Ax =A + A x Ax x =A + A x Ax = xx A We don t want to prove the claim rigorously. But a x = k a i x i i=

More information

Making sense of Econometrics: Basics

Making sense of Econometrics: Basics Making sense of Econometrics: Basics Lecture 2: Simple Regression Egypt Scholars Economic Society Happy Eid Eid present! enter classroom at http://b.socrative.com/login/student/ room name c28efb78 Outline

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

ECON3150/4150 Spring 2016

ECON3150/4150 Spring 2016 ECON3150/4150 Spring 2016 Lecture 4 - The linear regression model Siv-Elisabeth Skjelbred University of Oslo Last updated: January 26, 2016 1 / 49 Overview These lecture slides covers: The linear regression

More information

Simple Linear Regression Model & Introduction to. OLS Estimation

Simple Linear Regression Model & Introduction to. OLS Estimation Inside ECOOMICS Introduction to Econometrics Simple Linear Regression Model & Introduction to Introduction OLS Estimation We are interested in a model that explains a variable y in terms of other variables

More information

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X =

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X = The Gauss-Markov Linear Model y Xβ + ɛ y is an n random vector of responses X is an n p matrix of constants with columns corresponding to explanatory variables X is sometimes referred to as the design

More information

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018 Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate

More information

Lecture 4 Multiple linear regression

Lecture 4 Multiple linear regression Lecture 4 Multiple linear regression BIOST 515 January 15, 2004 Outline 1 Motivation for the multiple regression model Multiple regression in matrix notation Least squares estimation of model parameters

More information

14 Multiple Linear Regression

14 Multiple Linear Regression B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 14 Multiple Linear Regression 14.1 The multiple linear regression model In simple linear regression, the response variable y is expressed in

More information

The Statistical Property of Ordinary Least Squares

The Statistical Property of Ordinary Least Squares The Statistical Property of Ordinary Least Squares The linear equation, on which we apply the OLS is y t = X t β + u t Then, as we have derived, the OLS estimator is ˆβ = [ X T X] 1 X T y Then, substituting

More information

Econometrics - 30C00200

Econometrics - 30C00200 Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business

More information

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2) RNy, econ460 autumn 04 Lecture note Orthogonalization and re-parameterization 5..3 and 7.. in HN Orthogonalization of variables, for example X i and X means that variables that are correlated are made

More information

Need for Several Predictor Variables

Need for Several Predictor Variables Multiple regression One of the most widely used tools in statistical analysis Matrix expressions for multiple regression are the same as for simple linear regression Need for Several Predictor Variables

More information

Advanced Quantitative Methods: specification and multicollinearity

Advanced Quantitative Methods: specification and multicollinearity Advanced Quantitative Methods: Specification and multicollinearity University College Dublin February 16, 2011 1 Specification 2 3 4 Omitted variables Nonlinearities Outline 1 Specification 2 3 4 Omitted

More information

Inference in Regression Analysis

Inference in Regression Analysis Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value

More information

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression

More information

Statistical View of Least Squares

Statistical View of Least Squares Basic Ideas Some Examples Least Squares May 22, 2007 Basic Ideas Simple Linear Regression Basic Ideas Some Examples Least Squares Suppose we have two variables x and y Basic Ideas Simple Linear Regression

More information

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects Economics 113 Simple Regression Models Simple Regression Assumptions Simple Regression Derivation Changing Units of Measurement Nonlinear effects OLS and unbiased estimates Variance of the OLS estimates

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

The Classical Linear Regression Model

The Classical Linear Regression Model The Classical Linear Regression Model ME104: Linear Regression Analysis Kenneth Benoit August 14, 2012 CLRM: Basic Assumptions 1. Specification: Relationship between X and Y in the population is linear:

More information

Applied Econometrics (QEM)

Applied Econometrics (QEM) Applied Econometrics (QEM) based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #3 1 / 42 Outline 1 2 3 t-test P-value Linear

More information

Chapter 5 Matrix Approach to Simple Linear Regression

Chapter 5 Matrix Approach to Simple Linear Regression STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:

More information

Econometrics. Andrés M. Alonso. Unit 1: Introduction: The regression model. Unit 2: Estimation principles. Unit 3: Hypothesis testing principles.

Econometrics. Andrés M. Alonso. Unit 1: Introduction: The regression model. Unit 2: Estimation principles. Unit 3: Hypothesis testing principles. Andrés M. Alonso andres.alonso@uc3m.es Unit 1: Introduction: The regression model. Unit 2: Estimation principles. Unit 3: Hypothesis testing principles. Unit 4: Heteroscedasticity in the regression model.

More information

STAT 540: Data Analysis and Regression

STAT 540: Data Analysis and Regression STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

ECNS 561 Multiple Regression Analysis

ECNS 561 Multiple Regression Analysis ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking

More information

Empirical Economic Research, Part II

Empirical Economic Research, Part II Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction

More information

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define

More information

Estimating Estimable Functions of β. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 17

Estimating Estimable Functions of β. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 17 Estimating Estimable Functions of β Copyright c 202 Dan Nettleton (Iowa State University) Statistics 5 / 7 The Response Depends on β Only through Xβ In the Gauss-Markov or Normal Theory Gauss-Markov Linear

More information

Simple Linear Regression Analysis

Simple Linear Regression Analysis LINEAR REGRESSION ANALYSIS MODULE II Lecture - 6 Simple Linear Regression Analysis Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Prediction of values of study

More information

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0:

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0: Linear Regression 1 Single Explanatory Variable Assume (y is not necessarily normal) where Examples: y» F; Ey = + x Vary = ¾ 2 ) y = + x + u Eu = 0 Varu = ¾ 2 Exu = 0: 1. School performance as a function

More information