Econometrics I. Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius

Size: px
Start display at page:

Download "Econometrics I. Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius"

Transcription

1 Econometrics I Introduction to Regression Analysis (Outline of the course) Marijus Radavi ius Vilnius 2012

2 Content 1 Introduction Motivation Terminology Least squares method Least squares problem Linear regression Least squares for linear regression Geometrical interpretation of the least squares Regression function and prediction Denition of regression function Best linear prediction and Gaussian systems Gauss-Markov model Gauss-Markov conditions Properties of least squares estimator Gaussian Regression Denition of Gaussian Regression model Hypothesis testing When Gaussian-Markov model does not hold 16 1

3 1 Introduction 1.1 Motivation Econometrics is based upon the development of statistical methods for estimating economic relationships, testing economic theories,and evaluating and implementing government and business policy. The most common application of econometrics is the forecasting of such important macroeconomic variables as interest rates, ination rates, and gross domestic product. While forecasts of economic indicators are highly visible and are often widely published, econometric methods can be used in economic areas that have nothing to do with macroeconomic forecasting. For example, we will study the eects of political campaign expenditures on voting outcomes. We will consider the eect of school spending on student performance in the eld of education. The general purpose of course "Econometrics I" is to teach students to apply least squares method, construct and select regression models, as well as to develop statistical perception and ability to formalize real (economic) problems. Let us cite William H.Greene ([2], p.7): "The linear regression model is the single most useful tool in the econometrician's kit. Though to an increasing degree in the contemporary literature, it is often only the departure point for the full analysis, it remains the device used to begin almost all empirical research." 1.2 Terminology If a variable y of interest is inuenced by variables x 1,..., x k, this can be expressed mathematically by equation y = f(x), x := (x 1,..., x k ), (1) where f is some (usually) unknown function and A is the transposition of a matrix A. (Thus x is a column-vector). Actually, in economics, as opposed to natural sciences, any economic index is inuenced or related to lots of indices and factors. It is impossible to take them all into account and to measure. Thus, in practice, equation (1) is not valid, however it can hold approximately provided x contains all essential variables and the impact of the rest is comparatively small. This leads to the following model, called the regression model: y = f(x) + ε endogenious variable = regression function }{{} response = trend, signal }{{} systematic component 2 + errors }{{} + noise }{{} random component (2)

4 The variables y and x have several dierent names. According to old-fashioned tradition, y is called the dependent variable, while x is called the independent variable. The variable y is also called the explained variable, the response (variable), the predicted variable, or the regressand. The variable x is also called the explanatory variable, the control variable, the predictor (variable), the regressor, or the covariate. The terms dependent variable and independent variable are still rather frequently used in econometrics. But be aware that the label "independent" here does not refer to the statistical notion of independence between random variables. The terms explained and explanatory variables are probably the most descriptive. Response and control are used mostly in the experimental sciences (biometrics, technometrics), where the variable x is under the experimenter's control. The systematic component describes the expected values of endogenous variable y, i.e., "typical" or expected behavior of a (economic) system under conditions determined by the exogenous variable x while the random component ε describes uncertainty or random deviation of the system from its "typical" value. Degree of "uncertainty" or extent of random deviation is expressed by corresponding variances and covariances. When predicting y by x, the error term ε in (2) is omitted since it is "unpredictable", and we come back to equation (1). Hence, the function f determines the prediction rule of y through x. It is called the regression function (of y by x). Thus, the general problem of econometric analysis can be stated as follows: On the basis of available data to specify (identify) structure of the systematic component and covariance structure of the random component. To choose convenient parametrization for the specied structures and to estimate the unknown parameters. To make econometric inference and to give (economic) interpretation of the results. When dealing with structures it is important to be familiar with a dierent forms of their representations: symbolic specication, analytical formulas and available data representation in matrix language, as well as their graphical representation (graphical models). 3

5 2 Least squares method 2.1 Least squares problem The least squares method is based on 1. Available data D = D N (x, y) := {(x(t), y(t)), t = 1,..., N}, (3) where y(t) R and x(t) R k (R k is the k-dimensional euclidean space) are observations of explained variable y and k-dimensional explanatory variable x, respectively, in case t, N being a sample size. 2. A given class of possible F regression functions (prediction rules) f : R k R, f is a measurable function. 3. Goodness-of-t measure dened as (empirical) quadratic risk, i.e. average losses with the quadratic loss function (u v) 2, u, v R, ˆR(f) = ˆR 2 (f D) = N [y(t) f(x(t))] 2. (4) t=1 It is a special case of a risk with a general loss function l(u; v), u, v R, ˆR l (f) = ˆR l (f D) = N l (y(t); f(x(t))). (5) t=1 Least squares problem: Find ˆf F, ˆf = ˆf( D), which minimizes the risk ˆR(f D) in the class F. In short, ˆf := arg min ˆR(f D). (6) f F Example 1. Suppose F = {f : f const}. Then ˆf(x) ȳ := (1/N) N t=1 y(t), i.e. the average value of y's a well-known central characteristic of data. Example 2. Suppose F is the same as in the previous example, but the risk is dened with absolute losses l 1 (u; v) := u v, u, v R. Hence ˆR 1 (f) := ˆR l1 (f D) := N y(t) f(x(t)) (7) t=1 and the problem is to nd ˆf 1 := arg min f F ˆR 1 (f D). (8) 4

6 It is straightforward to check that ˆf 1 (x) median(y(t), t = 1,..., N), i.e. another well-known central characteristic of data. Example 3. (Multiple regression) Suppose F = F L where F L := {f : f(x) = β 0 + β x, β 0 R, β R k }. (9) is the class of all (continuous) linear mapping f : R k R. Example 3 describes the main object of interest of this course. 2.2 Linear regression Let D N (x, y) be data given in (3). Multiple linear regression model (with an intercept) (MLR) for D N (x, y) is determined by equations y(t) = β 0 + β x(t) + ε(t), t = 1,..., N. (10) Sometimes it is natural to assume that an expected value of y when all x's are zeros is also zero. This implies that intercept β 0 = 0. Multiple linear regression model without an intercept (MLR0) for D N (x, y) is determined by equations y(t) = β x(t) + ε(t), t = 1,..., N. (11) In matrix form, models (10) and (11) are written as follows: and Y = X β + E, (12) Y = Xβ + E, (13) respectively. Here X is the design matrix in the model (MLR0) X := ( X 1 X 2... X k ), (14) X i := ( x i (1), x i (2),..., x i (N) ), i = 1,..., k, (15) and X is the design matrix in the model (MLR) X := ( 1 N X ) = ( ) 1 N X 1 X 2... X k, (16) 1 N := ( ) 1, 1,..., 1 R }{{} N, (17) β := N times ( β 0, β ) R k+1. (18) 5

7 2.3 Least squares for linear regression Consider (MLR0) model in matrix form (13). For f F L with f(x) = f β (x) := β x, let us denote ˆR(β) := ˆR(f β ). Then the problem is ˆR(β) = ˆR(β D) = 1 N Y Xβ 2 min β R k. (19) The least squares estimator (LSE) ˆβ of β (in model (13)) (assuming that X X is not degenerate) is given by the equation Similarly, in model (12), LSE of β is equal to ˆβ = ( X X ) 1 X Y, (20) ˆ β = ( ˆβ0, ˆβ ) ( ) 1 = X X X Y. (21) More transparent representation of ˆ β is the following ˆβ 0 = ȳ ˆβ x, (22) ˆβ = (Ẋ ) 1Ẋ Ẋ Ẏ. (23) Here ȳ is the average of y's, x is the vector of averages of exploratory variables x, Ẋ := ( Ẋ 1 Ẋ 2... Ẋ k ), (24) Ẏ and Ẋi are vectors of centered values of the response variable y and i-th explanatory variable x i, respectively: Ẏ := Ẋ i := (. y(1) ȳ, y(2) ȳ,..., y(n) ȳ) (25) ( ), x i (1) x i, x i (2) x i,..., x i (N) x i i = 1,..., k, (26) 2.4 Geometrical interpretation of the least squares It follows from (22) that, in the (MLR) model (the model with the intercept!), the hyperplane determined by the least squares estimator ˆf(x) = ˆβ 0 + ˆβ x of the regression function f goes through the point ( x, ȳ), i.e., ˆf( x) = ȳ. It is clear, that the best linear prediction (least squares approximation) of Y is Ŷ = X ˆβ. (27) Let L = L X := span{x 1,..., X k } = {Xβ, β R k } (28) 6

8 denote the linear subspace generated by N-dimensional column-vectors X 1,..., X k. Then Ŷ = ΠY, (29) where Π = Π X denotes the orthogonal projector onto the linear subspace L X. In the case of (MLR) model, formulas analogous to (28) and (29) are valid for centered variables as well. It suces all vectors and matrices in (28) and (29) replace by their centered versions. Pitagor's theorem implies Y 2 = Ŷ 2 + Y Ŷ 2, (30) and Ẏ 2 = Ẏ 2 + Ê 2. (31) Here Ê := Y Ŷ = Ẏ Ẏ stands for the vector of the least squares residuals. Thus, we derive variance analysis identity or the formula of variance decomposition: σ 2 y = σ 2 ŷ + σ 2ˆε. (32) Here σ 2 y is the total variance of y. σŷ 2 is the variance of ŷ, the predicted values of y. It represents the part of variation of y explained by variation of explanatory variables x using the tted MLR model. σ2ˆε is the variance of the residuals ˆε. It represents the part of variation of y which can not be explained by variation of explanatory variables x using a MLR model. The variance analysis identity is often presented via sum of squares of the corresponding terms in (32) (i.e., in non-normalized way) Here SST = SSM + SSR. SST = Ẏ 2 is the total sum of squares. SSM = Ẏ 2 is the model sum of squares, i.e., sum of squares of the (centered) least squares approximations (predictions). SSR = Ê 2 is the residual sum of squares. 7

9 R-square or determination coecient R 2 := SSM SST = σ2 ŷ σ 2 y = 1 σ2ˆε σ 2 y = 1 SSR SST [0, 1] (33) represents the (relative) part of the variance of y explained by variation of explanatory variables x via the tted MLR model. R 2 = 1 implies that MLR model ts the data precisely without any errors (all residuals are zeros). R 2 = 0 implies that ˆβ = 0 k, i.e. the explanatory variables x explains nothing and the best prediction of y's is a constant ȳ. If (MLR0) model is applied, the corresponding formulas yields noncentered R- square or noncentered determination coecient R 2 0. Using geometrical interpretation Ẏ and Ê it is easy to derive the following convenient formulas ˆβ (Ẋ ) Ẏ ˆẎ R 2 Ẏ = Ẏ Ẏ = Ẏ = ˆẎ 2, (34) 2 Ẏ 2 ( ) ˆβ X Y Ŷ 2 R0 2 = = Y Y Y. (35) 2 Notice that Ẏ prediction ˆẎ = Ẋ ˆβ, ˆẎ Ẏ = ˆẎ ˆẎ = ˆẎ 2, the denominator Ẏ Ẏ = Ẏ 2, and Ẋ Ẏ has been already calculated in ˆβ formula (23). 8

10 3 Regression function and prediction We would like to investigate properties of the least squares estimator. The rst step in this direction could be replacing the empirical characteristics with the corresponding theoretical characteristics (the plug-in principal) and to study (theoretical) properties they imply. This enables us to formalize the denition of the regression function in a natural way. 3.1 Denition of regression function It is clear that the theoretical substitute of the (empirical) quadratic risk ˆR(f) (see (4)) is the corresponding theoretical quadratic risk where (x, y), x R k, y R is a random pair. R(f) = E [y f(x)] 2, (36) Denition 1. (Regression function) A measurable function f, f : R k R is called a regression function of y with respect to x R k if it minimizes the quadratic risk R(f) in the class of all measurable functions f : R k R. Proposition 1. Let (x, y) be a random pair, Ey 2 <. Then for any measurable function f : R k R with Ef 2 (x) < the following equality holds: Here, for brevity, E [y f(x)] 2 = E [(y f (x)] 2 + E [f(x) f (x)] 2. (37) denotes the conditional expectation of y given x. f (x) := E [y x] (38) Corollary 1. The measurable function E [y x] is the regression function. Denition 2. (Regression function) The regression function of y with respect to x R k is the conditional expectation E [y x] of y given x. The correlation ratio η 2 is dened as the maximal correlation between y and any measurable function of x: η 2 := max f F COR2 (y, f(x)). (39) From Proposition 1, the denition of η 2 (39) and (38), we derive that η 2 = COR 2 (y, f (x)) = VAR(f (x)). (40) VAR(y) It is clear from (39) and (40) that η 2 [0, 1], η 2 = 1 implies y f (x) almost surely and η 2 = 0 implies y const = E(y) almost surely. 9

11 3.2 Best linear prediction and Gaussian systems When seeking the best prediction rule one can impose some (reasonable) restrictions on these rules thus replacing the set of all measurable rules with some its proper subset. The choice F = F L with F L being the class of all linear mappings dened in (9) yields the best linear prediction rule f L := arg min f F L R(f). (41) Actually, f L is the least squares prediction rule and hence it can be expressed through the mean and the covariance matrix of the random vector (x, y) as follows: f L (x) := E(y) + COV(y, x) (COV(x, x)) 1 (x E(x)). (42) Moreover, the variance of the best linear prediction error ε := y f L (x) is given by VAR(ε) = E(y f L (x)) 2 = VAR(y) COV(y, x) (COV(x, x)) 1 COV(x, y). (43) is dened as the max- The theoretical (or expected) determination coecient RE 2 imal correlation between y and any linear function of x: R 2 E := max f F L COR 2 (y, f(x)). (44) Similarly as in case of (empirical) determination coecient (R-sqaure) R 2 it is established that RE 2 = COR 2 (y, f L (x)) = VAR(f L(x)). (45) VAR(y) It is obvious that η 2 RE 2 and generally the equality does not hold. However, it does hold if (x, y) is a Gaussian system. Denition 3. (Gaussian system) A pair (ξ, η) of random vectors ξ and η is called Gaussian system i the distribution of the random vector (ξ, η ) is Gaussian. Proposition 2. (Regression function for Gaussian system) Suppose (x, y) is a Gaussian system. Then E[y x] := E(y) + COV(y, x) (COV(x, x)) 1 (x E(x)). (46) and, for VAR[y x] := E [(y E[y x]) 2 x], VAR[y x] = VAR(y) COV(y, x) (COV(x, x)) 1 COV(x, y). (47) In other words, for Gaussian system (x, y) the regression function E[y x] = f L (x) and η 2 = R 2 E. 10

12 4 Gauss-Markov model Gauss-Markov model (GMm) is though simple but rather general model which ensures good properties of the least squares estimator of the linear regression coecients. 4.1 Gauss-Markov conditions Let D N (x, y) be given data. It is said that the data D N (x, y) satisfy Gauss-Markov model (GMm) if the following assumptions are fullled: (GM1) Model specication. (MLR0) model specication holds Y = Xβ + E, Y R N, β R k. (48) In fact, this is not a real restriction, rather a denition of the parameters β and the errors E. (GM2) Mean and covariances of errors. 1. Mean EE = 0 N, 2. Covariance COV(E, E) = σ 2 I N. Here σ 2 > 0 is an additional (to β) unknown parameter of the model, I N is the identity matrix of order N. (GM3) Independence of regressors and errors. The design matrix X and the error vector E are independent. In particular, this assumption is satised if X is non-random. (GM4) Uniqueness. The matrix X X is (almost surely) non-degenerate, i.e. rank(x) = k. This condition ensures that there exists the unique least squares estimator ˆβ of β. (GM5) The set of unknown regression coecients. The set of possible values of the unknown parameters β (denoted B) is the whole Euclidean space R k. A rather frequent alternative to B = R k is B = B A where B A := {β R k : Aβ = a}, A is a given m k matrix, rank(a) = m < k, a R m is a given vector. Remark. The Gauss-Markov model is a partial (or semiparametric) probabilistic model. It determines only moments of the observations up to second order, i.e., only means, variances and covariances. Thus, the Gauss-Markov model does not alow to calculate any probabilities. One can only to estimate probabilities by making use of Chebyshev-type inequalities. 11

13 4.2 Properties of least squares estimator In this section we will discuss the following properties of the least squares estimator ˆβ in the Gauss-Markov model: unbiasedness, consistency, eciency. Unbiasedness. The least squares estimator ˆβ in the Gauss-Markov model is unbiased. Consistency. consistent if The least squares estimator ˆβ in the Gauss-Markov model is ( (X tr X ) 1) 0, N. (49) Here tr(a) stands for the trace of a quadratic matrix A (i.e., the sum of the diagonal elements of A). Eciency. Let A be a given m n matrix, θ := Aβ. Gauss-Markov Theorem. Suppose that Gauss-Markov model holds. The estimator ˆθ := A ˆβ of θ where ˆβ is the least squares estimator of β is ecient in the class of all linear unbiased estimators of θ. The covariance matrix of the least squares estimator ˆβ in the Gauss-Markov model is cov ( ˆβ, ˆβ) = σ 2 ( X X ) 1. (50) In the Gauss-Markov model with the intercept β 0, the covariance matrix of the least squares estimator ˆβ of regression coecients β is cov ( ˆβ, ˆβ) = σ2 ( ) ( 1 cov(x, x) = σ 2 Ẋ Ẋ ) 1, (51) N where cov(x, x) is the empirical k k covariance matrix of the explanatory variables x, Ẋ is the design matrix of centered values of the explanatory variables x. 12

14 5 Gaussian Regression The Gauss-Markov model is only a partial (or semiparametric) probabilistic model. It determines only moments of the observed data up to second order, i.e., only means, variances and covariances. Hence, the Gauss-Markov model does not alow to calculate any probabilities and herewith to use likelihoods, construct condence intervals and to test hypotheses. A natural additional condition to the Gauss-Markov model leads to the Gaussian regression model (GRm), a complete probabilistic (regression) model. 5.1 Denition of Gaussian Regression model Let us assume that for (x, y) observation data D = D N (x, y) := {(x(t), y(t)), t = 1,..., N} Gauss-Markov model is valid with the condition (GM2) replaced with (tightened to) the condition (GR2). (GR2) Distribution of errors. The vector of random errors E is normally distributed with zero means and a spherical covariance matrix: E Normal N (0 N, σ 2 I N ), σ 2 > 0. (52) The brief mathematical form to write the Gaussian regression model (without making use of the unobservable error vector E) is the following: [Y X] Normal N (Xβ, σ 2 I N ). Let ˆβ ML and ˆσ ML 2 denote the maximum likelihood estimators (MLe) of the unknown GRm parameters β and σ 2, respectively. Proposition 1. (Maximum likelihood estimator) Suppose that data D N (x, y) satisfy Gaussian regression model. Then ˆβ ML = ˆβ where ˆβ is the least squares estimator of β and ˆσ ML 2 is the empirical variance of the residuals Ê := Y X ˆβ ˆσ 2 ML := 1 N N (y(t) ˆβ x(t)) 2 = 1 Y X N ˆβ 2. (53) t=1 The general asymptotic theory of maximum likelihood estimation implies that ˆβ ML and ˆσ ML 2 are asymptotically (as N ) ecient (in the class of all statistics). 13

15 5.2 Hypothesis testing Consider a general linear hypothesis testing problem H 0 : Aβ = a versus H 1 : Aβ a. (54) Here A is a given m k matrix, rank(a) = m, a R m is a given vector. Many problems important from the econometric viewpoint can be expressed as a general linear hypothesis testing problem. The standard ones are the following two problems: Testing statistical insignicance of a regression model. Assume Gauss- Markov model without intercept. The testing problem states H 0 : β = 0 k versus H 1 : β 0 k. (55) It follows from (54) with A = I k and a = 0 k. Testing statistical insignicance of separate regressors. For each i {1,..., k}, the testing problem states H (i) 0 : β i = 0 versus H (i) 1 : β 0. (56) For testing (54), the following criteria can be used: likelihood ratio, Fisher, and Student. It is convenient to introduce a new parameter θ := Aβ. Denote ˆθ := A ˆβ and s 2 := N N k ˆσ2 ML = 1 N k N (y(t) ˆβ x(t)) 2 = 1 Y X N k ˆβ 2. (57) t=1 The statistics s 2 is an unbiased estimator of σ 2 Fisher criterion. The general form of Fisher statistic is T F := m 1 (θ ˆθ) (ĈOV(ˆθ, ˆθ) ) 1 (θ ˆθ) (58) Here ĈOV(ˆθ, ˆθ) is an (unbiased) estimator of COV(ˆθ, ˆθ) obtained by estimating σ 2 with s 2, ĈOV(ˆθ, ˆθ) = s 2 A ( X X ) 1 A. (59) Thus, T F := (Aβ A ˆβ) (A ( X X ) 1 1 A ) (Aβ A ˆβ) ms 2. (60) Under the valid H 0, the Fisher statistic T F has the Fisher distribution with (N k, m) degrees of freedom. 14

16 The critical set of Fisher criterion for a signicance level α is K F α := {T F > F (N k,m) (1 α)} (61) Here F (N k,m) (p) denotes p-quantile of the Fisher distribution with (N k, m) degrees of freedom. Student criterion. If m = 1, i.e. θ R (is a scalar), the Fisher statistic T F := (θ ˆθ) 2 ĈOV(ˆθ, ˆθ) can be replaced by the corresponding Student statistic T S := θ ˆθ ĈOV(ˆθ, ˆθ) (62) which preserves information about the direction of deviations from the null hypothesis H 0. It is easy to derive that T S := A(β ˆβ) sλ A, λ 2 A := A ( X X ) 1 A, (63) and the critical set of Student criterion for a signicance level α is K S α := { T S > t N k (1 α/2)}. (64) Here t N k (p) denotes p-quantile of the Student distribution with N k degrees of freedom. Formula (64) is a consequence of well-known fact that the Student statistic T S, under the valid H 0, has the Student distribution with N k degrees of freedom. Likelihood ratio (LR) criterion. This criterion is usually introduced in standard course on mathematical statitic. Here we would like only to stress that the Fisher criterion in the Gaussian regression model is asymptotically equivalent to the LR criterion (for the same testing problem) when N. Remark 1. Condence intervals and regions for the parameters β and θ can be constructed by well-known relation between criteria of hypotheses testing and construction of condence regions or condence intervals. 15

17 6 When Gaussian-Markov model does not hold In this section, we discuss problems which arise when some assumptions of the Gauss-Markov model are violated. We try to answer the following questions: What are consequences of violation of a certain assumption of the Gauss- Markov model? How can these violations be detected (indentied)? What are remedies for undesirable consequences of these violations? This covers the following topics: Regresion model misspecication: violation of (GM1) includes nonlinearity of regression model, omission of relevant and inclusion of irrelevant regressors. Correlated and heteroscedastic errors: this item discusses violations of (GM2), generalized and feasible least squares method, transformations of the response variable. Stochastic regressors: violation of (GM3) is usually related to violations of exogeneity, e.g., errors in regressors. Multicolinearity: violation of (GM4) is usually referred to as multicolinearity, which, in a strict sense, rather rarely occur in practice but often takes latent forms yet leading to computational instability and statistical inconsistency. 16

18 References 1. Creel, M. (2006) Econometrics (version 0.80, 2006) 2. Greene, W.H. Econometric Analysis. 5th ed. Upper Saddle River, New Jersey: Prentice-Hall, Hill, R. C.; Griths, W. E.; Lim, G. C. (2011) Principles of Econometrics (4th ed.), John Wiley & Sons, Inc. 4. Peracchi, F. (2001) Econometrics, John Wiley & Sons, Inc. 5. Verbeek, M. (2004) A Guide to Modern Econometrics (2nd ed.), John Wiley & Sons, Inc. 6. Wooldridge, J.M. (2006) Introductory Econometrics: A Modern Approach (3rd edition.), South-Western College Publishing, Cincinnati, Ohio. 17

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Linear Regression. Junhui Qian. October 27, 2014

Linear Regression. Junhui Qian. October 27, 2014 Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency

More information

Lecture 1: OLS derivations and inference

Lecture 1: OLS derivations and inference Lecture 1: OLS derivations and inference Econometric Methods Warsaw School of Economics (1) OLS 1 / 43 Outline 1 Introduction Course information Econometrics: a reminder Preliminary data exploration 2

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation. Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 16, 2013 Outline Introduction Simple

More information

Introduction to Estimation Methods for Time Series models. Lecture 1

Introduction to Estimation Methods for Time Series models. Lecture 1 Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation

More information

Applied Econometrics (QEM)

Applied Econometrics (QEM) Applied Econometrics (QEM) based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #3 1 / 42 Outline 1 2 3 t-test P-value Linear

More information

Applied Econometrics (QEM)

Applied Econometrics (QEM) Applied Econometrics (QEM) The Simple Linear Regression Model based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #2 The Simple

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

The Simple Regression Model. Part II. The Simple Regression Model

The Simple Regression Model. Part II. The Simple Regression Model Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square

More information

Introduction to Linear Regression Analysis

Introduction to Linear Regression Analysis Introduction to Linear Regression Analysis Samuel Nocito Lecture 1 March 2nd, 2018 Econometrics: What is it? Interaction of economic theory, observed data and statistical methods. The science of testing

More information

Statistical Inference with Regression Analysis

Statistical Inference with Regression Analysis Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

1. The OLS Estimator. 1.1 Population model and notation

1. The OLS Estimator. 1.1 Population model and notation 1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology

More information

Advanced Econometrics

Advanced Econometrics Based on the textbook by Verbeek: A Guide to Modern Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna May 16, 2013 Outline Univariate

More information

Homoskedasticity. Var (u X) = σ 2. (23)

Homoskedasticity. Var (u X) = σ 2. (23) Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This

More information

Econometrics I Lecture 3: The Simple Linear Regression Model

Econometrics I Lecture 3: The Simple Linear Regression Model Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating

More information

Averaging Estimators for Regressions with a Possible Structural Break

Averaging Estimators for Regressions with a Possible Structural Break Averaging Estimators for Regressions with a Possible Structural Break Bruce E. Hansen University of Wisconsin y www.ssc.wisc.edu/~bhansen September 2007 Preliminary Abstract This paper investigates selection

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE Chapter 6. Panel Data Joan Llull Quantitative Statistical Methods II Barcelona GSE Introduction Chapter 6. Panel Data 2 Panel data The term panel data refers to data sets with repeated observations over

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model

Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory

More information

Advanced Quantitative Methods: ordinary least squares

Advanced Quantitative Methods: ordinary least squares Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 3 Jakub Mućk Econometrics of Panel Data Meeting # 3 1 / 21 Outline 1 Fixed or Random Hausman Test 2 Between Estimator 3 Coefficient of determination (R 2

More information

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22 MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

LECTURE 2 LINEAR REGRESSION MODEL AND OLS SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

STAT5044: Regression and Anova. Inyoung Kim

STAT5044: Regression and Anova. Inyoung Kim STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:

More information

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Multiple Regression Analysis. Part III. Multiple Regression Analysis Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant

More information

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors: Wooldridge, Introductory Econometrics, d ed. Chapter 3: Multiple regression analysis: Estimation In multiple regression analysis, we extend the simple (two-variable) regression model to consider the possibility

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley Panel Data Models James L. Powell Department of Economics University of California, Berkeley Overview Like Zellner s seemingly unrelated regression models, the dependent and explanatory variables for panel

More information

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

ECONOMETRICS. Bruce E. Hansen. c2000, 2001, 2002, 2003, University of Wisconsin

ECONOMETRICS. Bruce E. Hansen. c2000, 2001, 2002, 2003, University of Wisconsin ECONOMETRICS Bruce E. Hansen c2000, 200, 2002, 2003, 2004 University of Wisconsin www.ssc.wisc.edu/~bhansen Revised: January 2004 Comments Welcome This manuscript may be printed and reproduced for individual

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 6 Jakub Mućk Econometrics of Panel Data Meeting # 6 1 / 36 Outline 1 The First-Difference (FD) estimator 2 Dynamic panel data models 3 The Anderson and Hsiao

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information

Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation

Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 9: Asymptotics III(MLE) 1 / 20 Jensen

More information

Econ 583 Final Exam Fall 2008

Econ 583 Final Exam Fall 2008 Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random

More information

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria SOLUTION TO FINAL EXAM Friday, April 12, 2013. From 9:00-12:00 (3 hours) INSTRUCTIONS:

More information

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Tak Wai Chau February 20, 2014 Abstract This paper investigates the nite sample performance of a minimum distance estimator

More information

Estadística II Chapter 4: Simple linear regression

Estadística II Chapter 4: Simple linear regression Estadística II Chapter 4: Simple linear regression Chapter 4. Simple linear regression Contents Objectives of the analysis. Model specification. Least Square Estimators (LSE): construction and properties

More information

The regression model with one stochastic regressor (part II)

The regression model with one stochastic regressor (part II) The regression model with one stochastic regressor (part II) 3150/4150 Lecture 7 Ragnar Nymoen 6 Feb 2012 We will finish Lecture topic 4: The regression model with stochastic regressor We will first look

More information

Simple and Multiple Linear Regression

Simple and Multiple Linear Regression Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where

More information

New York University Department of Economics. Applied Statistics and Econometrics G Spring 2013

New York University Department of Economics. Applied Statistics and Econometrics G Spring 2013 New York University Department of Economics Applied Statistics and Econometrics G31.1102 Spring 2013 Text: Econometric Analysis, 7 h Edition, by William Greene (Prentice Hall) Optional: A Guide to Modern

More information

ECNS 561 Multiple Regression Analysis

ECNS 561 Multiple Regression Analysis ECNS 561 Multiple Regression Analysis Model with Two Independent Variables Consider the following model Crime i = β 0 + β 1 Educ i + β 2 [what else would we like to control for?] + ε i Here, we are taking

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Multiple Linear Regression CIVL 7012/8012

Multiple Linear Regression CIVL 7012/8012 Multiple Linear Regression CIVL 7012/8012 2 Multiple Regression Analysis (MLR) Allows us to explicitly control for many factors those simultaneously affect the dependent variable This is important for

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Applied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013

Applied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013 Applied Regression Chapter 2 Simple Linear Regression Hongcheng Li April, 6, 2013 Outline 1 Introduction of simple linear regression 2 Scatter plot 3 Simple linear regression model 4 Test of Hypothesis

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

The Linear Regression Model

The Linear Regression Model The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Linear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77

Linear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77 Linear Regression Chapter 3 September 27, 2016 Chapter 3 September 27, 2016 1 / 77 1 3.1. Simple linear regression 2 3.2 Multiple linear regression 3 3.3. The least squares estimation 4 3.4. The statistical

More information

The regression model with one fixed regressor cont d

The regression model with one fixed regressor cont d The regression model with one fixed regressor cont d 3150/4150 Lecture 4 Ragnar Nymoen 27 January 2012 The model with transformed variables Regression with transformed variables I References HGL Ch 2.8

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Economic modelling and forecasting

Economic modelling and forecasting Economic modelling and forecasting 2-6 February 2015 Bank of England he generalised method of moments Ole Rummel Adviser, CCBS at the Bank of England ole.rummel@bankofengland.co.uk Outline Classical estimation

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 2 Jakub Mućk Econometrics of Panel Data Meeting # 2 1 / 26 Outline 1 Fixed effects model The Least Squares Dummy Variable Estimator The Fixed Effect (Within

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Business Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Model Evaluation and Selection Predictive Ability of a Model: Denition and Estimation We aim at achieving a balance between parsimony

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

Rewrap ECON November 18, () Rewrap ECON 4135 November 18, / 35

Rewrap ECON November 18, () Rewrap ECON 4135 November 18, / 35 Rewrap ECON 4135 November 18, 2011 () Rewrap ECON 4135 November 18, 2011 1 / 35 What should you now know? 1 What is econometrics? 2 Fundamental regression analysis 1 Bivariate regression 2 Multivariate

More information

Lecture 4: Heteroskedasticity

Lecture 4: Heteroskedasticity Lecture 4: Heteroskedasticity Econometric Methods Warsaw School of Economics (4) Heteroskedasticity 1 / 24 Outline 1 What is heteroskedasticity? 2 Testing for heteroskedasticity White Goldfeld-Quandt Breusch-Pagan

More information

Large Sample Properties & Simulation

Large Sample Properties & Simulation Large Sample Properties & Simulation Quantitative Microeconomics R. Mora Department of Economics Universidad Carlos III de Madrid Outline Large Sample Properties (W App. C3) 1 Large Sample Properties (W

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

1. The Multivariate Classical Linear Regression Model

1. The Multivariate Classical Linear Regression Model Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The

More information

Chapter 2: simple regression model

Chapter 2: simple regression model Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.

More information

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y Predictor or Independent variable x Model with error: for i = 1,..., n, y i = α + βx i + ε i ε i : independent errors (sampling, measurement,

More information

Econometrics - 30C00200

Econometrics - 30C00200 Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

Empirical Economic Research, Part II

Empirical Economic Research, Part II Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?

More information

3. Linear Regression With a Single Regressor

3. Linear Regression With a Single Regressor 3. Linear Regression With a Single Regressor Econometrics: (I) Application of statistical methods in empirical research Testing economic theory with real-world data (data analysis) 56 Econometrics: (II)

More information

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that Linear Regression For (X, Y ) a pair of random variables with values in R p R we assume that E(Y X) = β 0 + with β R p+1. p X j β j = (1, X T )β j=1 This model of the conditional expectation is linear

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory

More information

Linear regression COMS 4771

Linear regression COMS 4771 Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.

More information

Statistics 910, #5 1. Regression Methods

Statistics 910, #5 1. Regression Methods Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 1 Jakub Mućk Econometrics of Panel Data Meeting # 1 1 / 31 Outline 1 Course outline 2 Panel data Advantages of Panel Data Limitations of Panel Data 3 Pooled

More information

Quantitative Methods I: Regression diagnostics

Quantitative Methods I: Regression diagnostics Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear

More information

AMS-207: Bayesian Statistics

AMS-207: Bayesian Statistics Linear Regression How does a quantity y, vary as a function of another quantity, or vector of quantities x? We are interested in p(y θ, x) under a model in which n observations (x i, y i ) are exchangeable.

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information