Advanced Econometrics I
|
|
- Carmella Morris
- 6 years ago
- Views:
Transcription
1 Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim
2 1. Introduction Introduction & CLRM, Autumn Term
3 What is econometrics? Econometrics = economic statistics economic theory mathematics 1 Important notions: Data perceived as realizations of random variables Parameters are real numbers, not random variables Joint distributions of random variables depend on parameters A model a set of restrictions on the joint distribution of variables 1 according to Ragnar Frisch Introduction & CLRM, Autumn Term
4 Motivating Example Mincer equation - The influence of schooling on wages ln(w AGE i ) = β 1 +β 2 S i +β 3 T ENURE i +β 4 EXP R i +ε i Notation: Logarithm of the wage rate: ln(w AGE i ) Years of schooling: S i Experience in the current job: T ENURE i Experience in the labor market: EXP R i Estimation of the parameters β k, where β 2 is return to schooling Introduction & CLRM, Autumn Term
5 The importance of relationships 2 Relationship among variables is what empirical analysis is all about Consider: a dependent variable y i, and a vector of k explanatory variables, x i. Let r i = (y i, x i ) ; i = 1,..., N be i.i.d. Our interest is in the relationship between y i and the explanatory variables with the ultimate agenda of: 2 this section draws from Angrist, 2009 Introduction & CLRM, Autumn Term
6 Description - How does y i usually vary with x i? Prediction - Can we use x i to forecast y i? Causality - What is the effect of elements of x i on y i? Generally we look for relationships that hold on average. On average relationships among variables are summarised by the Conditional Expectation Function (CEF), where E[y i x i ] h(x i ) y i = h(x i ) + ε i E[ε i x i ] 0. Introduction & CLRM, Autumn Term
7 General regression equations Generalization: y i = β 1 x i1 + β 2 x i β K x ik + ε i Index for observations i = 1, 2,..., n and regressors k = 1, 2,..., K y i = β x i + ε i (1x1) (1xK) (Kx1) (1x1) β = β 1 β 2. and x i = x i1 x i2. β K x ik Introduction & CLRM, Autumn Term
8 The key problem of econometrics: We deal with non-experimental data Unobservable variables, interdependence, endogeneity, causality Examples: - Ability bias in Mincer equation - Reverse causality problem if unemployment is regressed on liberalization index - Causal effect on police force and crime is not an independent outcome - Simultaneity problem in demand price equation Introduction & CLRM, Autumn Term
9 2. The CLRM: Parameter Estimation by OLS Hayashi p. 6/15-18 Introduction & CLRM, Autumn Term
10 Classical linear regression model (CLRM) y i = β 1 x i1 + β 2 x i β K x ik + ε i = x i β + ε i (1xK) (Kx1) y i : Dependent variable, observed x i = (x i1, x i2,..., x ik ): Explanatory variables, observed β = (β 1, β 2,..., β K ): Unknown parameters ε i : Disturbance component, unobserved b = (b 1, b 2,..., b K ) estimator of β e i = y i x ib: Estimated residual Introduction & CLRM, Autumn Term
11 For convenience we introduce matrix notation y = X β + ε (nx1) (nxk) (Kx1) (nx1) y 1 y 2. y n = 1 x 12 x x 1K 1. x x n2... x nk β 1 β 2. β K + ε 1 ε 2. ε n Introduction & CLRM, Autumn Term
12 Writing extensively: A system of linear equations y 1 = β 1 + β 2 x β K x 1K + ε 1 y 2 = β 1 + β 2 x β K x 2K + ε 2. y n = β 1 + β 2 x n β K x nk + ε n Introduction & CLRM, Autumn Term
13 We estimate the linear model and choose b such that SSR is minimized Obtain an estimator b of β by minimizing the SSR (sum of squared residuals): argmins(b) = argmin n i=1 e2 i = argmin n i=1 (y i x i b)2 {b} Differentiation with respect to b 1, b 2,..., b K FOC s: (1) S(b) b 1! = 0 e i = 0 (2) S(b) b 2! = 0 e i x i2 = 0. (K) S(b) b K! = 0 e i x ik = 0 FOC s can be conveniently written in matrix notation X e = 0 Introduction & CLRM, Autumn Term
14 The system of K equations is solved by matrix algebra X e = X (y Xb) = X y X Xb = 0 Premultiplying by (X X) 1 : (X X) 1 X y (X X) 1 X Xb = 0 (X X) 1 X y Ib = 0 OLS-estimator: b = (X X) 1 X y Alternatively: b = ( 1 n X X ) 1 1 n X y = ( 1 n n i=1 x ix i) 1 1 n n i=1 x iy i Introduction & CLRM, Autumn Term
15 Zoom into the matrices X X and X y b = ( 1 n X X ) 1 1 n X y = ( 1 n n i=1 x ix i) 1 1 n n i=1 x iy i n i=1 x ix i = x 2 i1 xi1 x i2 xi1 x i3... xi1 x ik xi1 x i2 x 2 i2 xi2 x ik xi1 x ik xi2 x ik... x 2 ik n i=1 x iy i = xi1 y i xi2 y i xi3 y i. xik y i Introduction & CLRM, Autumn Term
16 3. Assumptions of the CLRM Hayashi p Introduction & CLRM, Autumn Term
17 The four core assumptions of CLRM 1.1 Linearity y i = x i β + ε i 1.2 Strict exogeneity E(ε i X) = 0 E(ε i ) = 0 and Cov(ε i, x ik ) = E(ε i x ik ) = No exact multicollinearity, P (rank(x) = k) = 1 No linear dependencies in the data matrix 1.4 Spherical disturbances: V ar(ε i X) = E(ε 2 i X) = σ2 Cov(ε i, ε j X) = 0; E(ε i ε j X) = 0 E(ε i ) = σi 2 and Cov(ε i, ε j ) = 0 by LTE (see Hayashi p. 18) Introduction & CLRM, Autumn Term
18 Interpreting the parameters β of different types of linear equations Linear model y i = β 1 + β 2 x i β K x ik + ε i : A one unit increase in the independent variable x ik increases the dependent variable by β k units Semi-log form log(y i ) = β 1 + β 2 x i β K x ik + ε i : A one unit increase in the independent variable increases the dependent variable approximately by 100 β k percent Log linear model log(y i ) = β 1 log(x i1 ) + β 2 log(x i2 ) β K log(x ik ) + ε i : A one percent increase in x ik increases the dependent variable y i approximately by β k percent Introduction & CLRM, Autumn Term
19 Some important laws Law of Total Expectation (LTE): E X [E Y X (Y X)] = E Y (Y ) Double Expectation Theorem (DET): E X [E Y X (g(y ) X)] = E Y (g(y )) Law of Iterated Expectations (LIE): E Z X [E Y X,Z (Y X, Z) X] = E Y X (Y X) Introduction & CLRM, Autumn Term
20 Some important laws (continued) Generalized DET: E X [E Y X (g(x, Y )) X] = E X,Y (g(x, Y )) Linearity of Conditional Expectations: E Y X [g(x)y X] = g(x)e Y X [Y X] Introduction & CLRM, Autumn Term
21 4. Finite sample properties of the OLS estimator Hayashi p Introduction & CLRM, Autumn Term
22 Finite sample properties of b = (X X) 1 X y 1. E(b) = β: Unbiasedness of the estimator Holds for any sample size Holds under assumptions V ar(b X) = σ 2 (X X) 1 : Conditional variance of b Conditional variance depends on the data Holds under assumptions V ar(ˆβ X) V ar(b X) ˆβ is any other linear unbiased estimator of β Holds under assumptions Introduction & CLRM, Autumn Term
23 Some key results from mathematical statistics z = (nx1) z 1 z 2. z n A = (mxn) a 11 a a 1n a 21 a a m1 a m2... a mn A new random variable: v = (mx1) A (mxn) z (nx1) E(v) (mx1) = E(v 1 ) E(v 2 ). E(v m ) = AE(z) V ar(v) (mxm) = AV ar(z)a Introduction & CLRM, Autumn Term
24 The OLS estimator s unbiasedness E(b) = β E(b β) = 0 sampling error b β = (X X) 1 X y β = (X X) 1 X (Xβ + ε) β = (X X) 1 X Xβ + (X X) 1 X ε β = β + (X X) 1 X ε β = (X X) 1 X ε E(b β X) = (X X) 1 X E(ε X) = 0 under assumption 1.2 E X (E(b X)) = E X (β) = E(b) by the LTE Introduction & CLRM, Autumn Term
25 We show that V ar(b X) = σ 2 (X X) 1 V ar(b X) = V ar(b β X) = V ar((x X) 1 X ε X) = V ar(aε X) = AV ar(ε X)A = Aσ 2 I n A = σ 2 AI n A = σ 2 AA = σ 2 (X X) 1 X X(X X) 1 = σ 2 (X X) 1 Note: β non-random b β sampling error A = (X X) 1 X V ar(ε X) = σ 2 I n Introduction & CLRM, Autumn Term
26 Sketch of the proof of the Gauss Markov theorem V ar( ˆβ X) V ar(b X) V ar( ˆβ X) = V ar(ˆβ β X) = V ar[(d + A)ε X] = (D + A)V ar(ε X)(D + A ) = σ 2 (D + A)(D + A ) = σ 2 (DD + AD + DA + AA ) = σ 2 [DD + (X X) 1 ] σ 2 (X X) 1 = V ar(b X) where C is a function of X ˆβ = Cy D = C A A (X X) 1 X Details of proof: Hayashi pages Introduction & CLRM, Autumn Term
27 The OLS estimator is BLUE - OLS is the best estimator Holds under the Gauss Markov theorem V ar( ˆβ X) V ar(b X) - OLS is linear Holds under assumption OLS is unbiased Holds under assumption Introduction & CLRM, Autumn Term
28 5. Hypothesis Testing under Normality Hayashi p Introduction & CLRM, Autumn Term
29 Hypothesis testing Economic theory provides hypotheses about parameters If theory is right testable implications But: Hypotheses can t be tested without distributional assumptions about ε Distributional assumption: Normality assumption about the conditional distribution of ε X MV N(0, σ 2 I n ) [Assumption 1.5] Introduction & CLRM, Autumn Term
30 Some facts from multivariate statistics Vector of random variables: x = (x 1, x 2,..., x n ) Expectation vector: E(x) = µ = (µ 1, µ 2,..., µ n ) = (E(x 1 ), E(x 2 ),..., E(x n )) Variance-covariance matrix: V ar(x) = Σ = V ar(x 1 ) Cov(x 1, x 2 )... Cov(x 1, x n ) Cov(x 1, x 2 ). V ar(x 2 ).... Cov(x 1, x n )... V ar(x n ) y = c + Ax; c, A non-random vector/matrix E(y) = (E(y 1 ), E(y 2 ),..., E(y n )) = c + Aµ V ar(y) = AΣA x MV N(µ, Σ) y = c + Ax MV N(c + Aµ, AΣA ) Introduction & CLRM, Autumn Term
31 Application of the facts from multivariate statistics and the assumptions b }{{ β } = (X X) 1 X ε sampling error Assuming ε X MV N(0, σ 2 I n ) b β X MV N ( (X X) 1 X E(ε X), (X X) 1 X σ 2 I n X(X X) 1) b β X MV N ( 0, σ 2 (X X) 1) Note that V ar(b X) = σ 2 (X X) 1 OLS-estimator conditionally normally distributed if ε X is multivariate normal Introduction & CLRM, Autumn Term
32 Testing hypothesis about individual parameters (t-test) Null hypothesis: H 0 : β k = β k, β k a hypothesized value, a real number Under assumption 1.5 and ε X MV N(0, σ 2 I n ) alternative hypothesis: H A : β k β k If H 0 is true E(b k ) = β k Test statistic: t k = b k β k σ 2 [(X X) 1 ] kk N(0, 1) Note: [(X X) 1 ] kk is the k-th row k-th column element of (X X) 1 Introduction & CLRM, Autumn Term
33 Nuisance parameter σ 2 can be estimated σ 2 = E(ε 2 i X) = V ar(ε i X) = E(ε 2 i ) = V ar(ε i) We don t know ε i but we use the estimator e i = y i x i b ˆσ 2 = 1 n n i=1 (e i 1 n n i=1 e i) 2 = 1 n n i=1 e2 i = 1 ne e ˆσ 2 is a biased estimator: E(ˆσ 2 X) = n K n σ2 Introduction & CLRM, Autumn Term
34 An unbiased estimator of σ 2 For s 2 = 1 n K n i=1 e2 i = 1 n K e e we get an unbiased estimator E(s 2 X) = 1 n K E(e e X) = σ 2 E ( E(s 2 X) ) = E(s 2 ) = σ 2 Using this provides an unbiased estimator of V ar(b X) = σ 2 (X X) 1 : V ar(b X) = s 2 (X X) 1 t-statistic under H 0 : t k = b k β k [ ] V ar(b X) kk = b k β k SE(b k ) = b k β k [ V ar(bk X) ] t(n K) Introduction & CLRM, Autumn Term
35 Decision rule for the t-test 1. H 0 : β k = β k, is often β k = 0 H A : β k β k 2. Given β k, OLS-estimate b k and s 2, we compute t k = b k β k SE(b k ) 3. Fix significance level α of two-sided test 4. Fix non-rejection and rejection regions decision Remark: σ2 [(X X) 1 ] kk : standard deviation b k X s2 [(X X) 1 ] kk : standard error b k X Introduction & CLRM, Autumn Term
36 Testing joint hypotheses (F-test/Wald test) Write hypothesis as: H 0 : R β = r (#r x K) (K x 1) (#r x 1) R: matrix of real numbers r: number of restrictions Replacing the β = (β 1, β 2,..., β k ) by estimator b = (b 1, b 2,..., b K ) : R b = r Introduction & CLRM, Autumn Term
37 Definition of the F-test statistic Properties of R b: R E(b X) = Rβ = r R V ar(b X)R = R σ 2 (X X) 1 R Rb = r MV N(Rβ, R σ 2 (X X) 1 R ) Using some additional important facts from multivariate statistics z = (z 1, z 2,..., z m ) MV N(µ, Ω) (z µ) Ω 1 (z µ) χ 2 (m) Result applied: Wald statistic (Rb r) [σ 2 R(X X) 1 R ] 1 (Rb r) χ 2 (#r) Introduction & CLRM, Autumn Term
38 Properties of the F-test statistic Replace σ 2 by its unbiased estimate s 2 = 1 dividing by #r: n K n i=1 e2 i = 1 n K e e and F -ratio: F = (Rb r) [R(X X) 1 R ] 1 (Rb r)/#r (e e)/(n K) = (Rb r) [R V ar(b X)R ] 1 (Rb r)/#r F (#r, n K) Note: F-test is one-sided Proof: see Hayashi p. 41 Introduction & CLRM, Autumn Term
39 Decision rule of the F-test 1. Specify H 0 in the form Rβ = r and H A : Rβ r. 2. Calculate F-statistic. 3. Look up entry in the table of the F-distribution for #r and n K at given significance level. 4. Null is not rejected on the significance level α for F less than F α (#r, n K) Introduction & CLRM, Autumn Term
40 Alternative representation of the F-statistic Minimization of the unrestricted sum of squared residuals: min n i=1 (y i x i b)2 SSR U Minimization of the restricted sum of squared residuals: F -ratio: min n i=1 (y i x i b) 2 SSR R F = (SSR R SSR U )/#r SSR U /(n K) Introduction & CLRM, Autumn Term
41 6. Confidence intervals and goodness of fit measures Hayashi p. 38/20 Introduction & CLRM, Autumn Term
42 Duality of t-test and confidence interval Under H 0 : β k = β k t k = b k β k SE(b k ) t(n K) Probability for non-rejection: ( ) P tα 2 (n K) t k tα 2 (n K) = 1 α tα 2 (n K) tα 2 (n K) t k lower critical value upper critical value random variable (value of test statistic) 1 α fixed number ( ) P b k SE(b k )tα 2 (n K) β k b k + SE(b k )tα 2 (n K) = 1 α Introduction & CLRM, Autumn Term
43 The confidence interval Confidence interval for β k : ( ) P b k SE(b k )tα 2 (n K) β k b k + SE(b k )tα 2 (n K) = 1 α The confidence bounds are random variables! b k SE(b k )tα 2 (n K): lower bound b k + SE(b k )tα 2 (n K): upper bound Wrong Interpretation: True parameter β k lies with probability 1 α within the bounds of the confidence interval Problem: Confidence bounds are not fixed; they are random! H 0 is rejected at significance level α if the hypothesized value does not lie within the confidence bounds of the 1 α interval. Introduction & CLRM, Autumn Term
44 Coefficient of determination: uncentered R 2 Measure of the variability of the dependent variable: yi 2 = y y Decomposition of y y: y y = (ŷ + e) (ŷ + e) = ŷ ŷ + 2ŷe + e e = ŷ ŷ + e e R 2 uc 1 e e y y A good model explains much and therefore the residual variation is very small compared to the explained variation. Introduction & CLRM, Autumn Term
45 Coefficient of determination: centered R 2 and R 2 adj Use centered R 2 if there is a constant in the model (x i1 = 1) n i=1 (y i y) 2 = n i=1 (ŷ i y) 2 + n i=1 e2 i R 2 c 1 ni=1 e 2 i ni=1 (y i y) 2 1 SSR SST Note, that R 2 uc and R 2 c lie both in the interval [0, 1] but describe different models. They are not comparable! R 2 adj is constructed with a penalty for heavy parametrization: Radj 2 = 1 SSR/(n K) SST/(n 1) = 1 n 1 SSR n K SST The R 2 adj is an accepted model selection criterion Introduction & CLRM, Autumn Term
46 Alternative goodness of fit measures Akaike criterion (AIC): Schwarz criterion (SBC): log ( ) SSR n + 2K n log ( SSR n ) + log(n)k n Note: Both criteria include a penalty term for heavy parametrization Select model with smallest AIC/SBC Introduction & CLRM, Autumn Term
LECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationEconometrics I Lecture 3: The Simple Linear Regression Model
Econometrics I Lecture 3: The Simple Linear Regression Model Mohammad Vesal Graduate School of Management and Economics Sharif University of Technology 44716 Fall 1397 1 / 32 Outline Introduction Estimating
More information1. The OLS Estimator. 1.1 Population model and notation
1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationApplied Econometrics (QEM)
Applied Econometrics (QEM) based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #3 1 / 42 Outline 1 2 3 t-test P-value Linear
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationThe Multiple Regression Model Estimation
Lesson 5 The Multiple Regression Model Estimation Pilar González and Susan Orbe Dpt Applied Econometrics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 5 Regression model:
More information7. Introduction to Large Sample Theory
7. Introuction to Large Samle Theory Hayashi. 88-97/109-133 Avance Econometrics I, Autumn 2010, Large-Samle Theory 1 Introuction We looke at finite-samle roerties of the OLS estimator an its associate
More informationAdvanced Quantitative Methods: ordinary least squares
Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationThe Simple Regression Model. Part II. The Simple Regression Model
Part II The Simple Regression Model As of Sep 22, 2015 Definition 1 The Simple Regression Model Definition Estimation of the model, OLS OLS Statistics Algebraic properties Goodness-of-Fit, the R-square
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationProblem Set #6: OLS. Economics 835: Econometrics. Fall 2012
Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.
More informationMotivation for multiple regression
Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope
More informationLECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit
LECTURE 6 Introduction to Econometrics Hypothesis testing & Goodness of fit October 25, 2016 1 / 23 ON TODAY S LECTURE We will explain how multiple hypotheses are tested in a regression model We will define
More informationLinear Regression. Junhui Qian. October 27, 2014
Linear Regression Junhui Qian October 27, 2014 Outline The Model Estimation Ordinary Least Square Method of Moments Maximum Likelihood Estimation Properties of OLS Estimator Unbiasedness Consistency Efficiency
More informationIntroduction to Estimation Methods for Time Series models. Lecture 1
Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationPractical Econometrics. for. Finance and Economics. (Econometrics 2)
Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationIntroduction to Econometrics
Introduction to Econometrics Lecture 3 : Regression: CEF and Simple OLS Zhaopeng Qu Business School,Nanjing University Oct 9th, 2017 Zhaopeng Qu (Nanjing University) Introduction to Econometrics Oct 9th,
More informationHypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima
Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s
More informationEconomics 620, Lecture 2: Regression Mechanics (Simple Regression)
1 Economics 620, Lecture 2: Regression Mechanics (Simple Regression) Observed variables: y i ; x i i = 1; :::; n Hypothesized (model): Ey i = + x i or y i = + x i + (y i Ey i ) ; renaming we get: y i =
More informationLECTURE 5 HYPOTHESIS TESTING
October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More information3. Linear Regression With a Single Regressor
3. Linear Regression With a Single Regressor Econometrics: (I) Application of statistical methods in empirical research Testing economic theory with real-world data (data analysis) 56 Econometrics: (II)
More informationEconometrics I. Professor William Greene Stern School of Business Department of Economics 3-1/29. Part 3: Least Squares Algebra
Econometrics I Professor William Greene Stern School of Business Department of Economics 3-1/29 Econometrics I Part 3 Least Squares Algebra 3-2/29 Vocabulary Some terms to be used in the discussion. Population
More informationREVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b)
REVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b) Explain/Obtain the variance-covariance matrix of Both in the bivariate case (two regressors) and
More informationEconometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018
Econometrics I KS Module 1: Bivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: March 12, 2018 Alexander Ahammer (JKU) Module 1: Bivariate
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationGreene, Econometric Analysis (7th ed, 2012)
EC771: Econometrics, Spring 2012 Greene, Econometric Analysis (7th ed, 2012) Chapters 2 3: Classical Linear Regression The classical linear regression model is the single most useful tool in econometrics.
More informationRegression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,
Regression Analysis The multiple linear regression model with k explanatory variables assumes that the tth observation of the dependent or endogenous variable y t is described by the linear relationship
More informationIntroductory Econometrics
Introductory Econometrics Violation of basic assumptions Heteroskedasticity Barbara Pertold-Gebicka CERGE-EI 16 November 010 OLS assumptions 1. Disturbances are random variables drawn from a normal distribution.
More informationApplied Econometrics (QEM)
Applied Econometrics (QEM) The Simple Linear Regression Model based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #2 The Simple
More informationThe regression model with one fixed regressor cont d
The regression model with one fixed regressor cont d 3150/4150 Lecture 4 Ragnar Nymoen 27 January 2012 The model with transformed variables Regression with transformed variables I References HGL Ch 2.8
More informationEmpirical Economic Research, Part II
Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction
More informationThe regression model with one stochastic regressor (part II)
The regression model with one stochastic regressor (part II) 3150/4150 Lecture 7 Ragnar Nymoen 6 Feb 2012 We will finish Lecture topic 4: The regression model with stochastic regressor We will first look
More informationReference: Davidson and MacKinnon Ch 2. In particular page
RNy, econ460 autumn 03 Lecture note Reference: Davidson and MacKinnon Ch. In particular page 57-8. Projection matrices The matrix M I X(X X) X () is often called the residual maker. That nickname is easy
More information1. The Multivariate Classical Linear Regression Model
Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationThe Multiple Linear Regression Model
Short Guides to Microeconometrics Fall 2016 Kurt Schmidheiny Unversität Basel The Multiple Linear Regression Model 1 Introduction The multiple linear regression model and its estimation using ordinary
More informationThe regression model with one stochastic regressor.
The regression model with one stochastic regressor. 3150/4150 Lecture 6 Ragnar Nymoen 30 January 2012 We are now on Lecture topic 4 The main goal in this lecture is to extend the results of the regression
More informationThe Statistical Property of Ordinary Least Squares
The Statistical Property of Ordinary Least Squares The linear equation, on which we apply the OLS is y t = X t β + u t Then, as we have derived, the OLS estimator is ˆβ = [ X T X] 1 X T y Then, substituting
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More informationTopic 4: Model Specifications
Topic 4: Model Specifications Advanced Econometrics (I) Dong Chen School of Economics, Peking University 1 Functional Forms 1.1 Redefining Variables Change the unit of measurement of the variables will
More informationMultiple Regression Analysis
Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationThe general linear regression with k explanatory variables is just an extension of the simple regression as follows
3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because
More information3 Multiple Linear Regression
3 Multiple Linear Regression 3.1 The Model Essentially, all models are wrong, but some are useful. Quote by George E.P. Box. Models are supposed to be exact descriptions of the population, but that is
More informationStatistical Inference with Regression Analysis
Introductory Applied Econometrics EEP/IAS 118 Spring 2015 Steven Buck Lecture #13 Statistical Inference with Regression Analysis Next we turn to calculating confidence intervals and hypothesis testing
More informationEconomics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects
Economics 113 Simple Regression Models Simple Regression Assumptions Simple Regression Derivation Changing Units of Measurement Nonlinear effects OLS and unbiased estimates Variance of the OLS estimates
More informationSimple Linear Regression
Simple Linear Regression Christopher Ting Christopher Ting : christophert@smu.edu.sg : 688 0364 : LKCSB 5036 January 7, 017 Web Site: http://www.mysmu.edu/faculty/christophert/ Christopher Ting QF 30 Week
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationChapter 2 Multiple Regression I (Part 1)
Chapter 2 Multiple Regression I (Part 1) 1 Regression several predictor variables The response Y depends on several predictor variables X 1,, X p response {}}{ Y predictor variables {}}{ X 1, X 2,, X p
More informationThe Linear Regression Model
The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general
More informationcoefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1
Review - Interpreting the Regression If we estimate: It can be shown that: where ˆ1 r i coefficients β ˆ+ βˆ x+ βˆ ˆ= 0 1 1 2x2 y ˆβ n n 2 1 = rˆ i1yi rˆ i1 i= 1 i= 1 xˆ are the residuals obtained when
More informationBasic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = The errors are uncorrelated with common variance:
8. PROPERTIES OF LEAST SQUARES ESTIMATES 1 Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = 0. 2. The errors are uncorrelated with common variance: These assumptions
More informationApplied Econometrics (MSc.) Lecture 3 Instrumental Variables
Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:
More informationEconometrics. Andrés M. Alonso. Unit 1: Introduction: The regression model. Unit 2: Estimation principles. Unit 3: Hypothesis testing principles.
Andrés M. Alonso andres.alonso@uc3m.es Unit 1: Introduction: The regression model. Unit 2: Estimation principles. Unit 3: Hypothesis testing principles. Unit 4: Heteroscedasticity in the regression model.
More informationIntermediate Econometrics
Intermediate Econometrics Markus Haas LMU München Summer term 2011 15. Mai 2011 The Simple Linear Regression Model Considering variables x and y in a specific population (e.g., years of education and wage
More informationEconometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012
Econometric Methods Prediction / Violation of A-Assumptions Burcu Erdogan Universität Trier WS 2011/2012 (Universität Trier) Econometric Methods 30.11.2011 1 / 42 Moving on to... 1 Prediction 2 Violation
More informationEcon 510 B. Brown Spring 2014 Final Exam Answers
Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity
More informationPANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1
PANEL DATA RANDOM AND FIXED EFFECTS MODEL Professor Menelaos Karanasos December 2011 PANEL DATA Notation y it is the value of the dependent variable for cross-section unit i at time t where i = 1,...,
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationECON 4160, Autumn term Lecture 1
ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least
More informationInstrumental Variables
Università di Pavia 2010 Instrumental Variables Eduardo Rossi Exogeneity Exogeneity Assumption: the explanatory variables which form the columns of X are exogenous. It implies that any randomness in the
More informationEconometrics - 30C00200
Econometrics - 30C00200 Lecture 11: Heteroskedasticity Antti Saastamoinen VATT Institute for Economic Research Fall 2015 30C00200 Lecture 11: Heteroskedasticity 12.10.2015 Aalto University School of Business
More informationStatistics and Econometrics I
Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,
More informationECON 3150/4150 spring term 2014: Exercise set for the first seminar and DIY exercises for the first few weeks of the course
ECON 3150/4150 spring term 2014: Exercise set for the first seminar and DIY exercises for the first few weeks of the course Ragnar Nymoen 13 January 2014. Exercise set to seminar 1 (week 6, 3-7 Feb) This
More informationMa 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA
Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA March 6, 2017 KC Border Linear Regression II March 6, 2017 1 / 44 1 OLS estimator 2 Restricted regression 3 Errors in variables 4
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model
Wooldridge, Introductory Econometrics, 4th ed. Chapter 2: The simple regression model Most of this course will be concerned with use of a regression model: a structure in which one or more explanatory
More informationAnswers to Problem Set #4
Answers to Problem Set #4 Problems. Suppose that, from a sample of 63 observations, the least squares estimates and the corresponding estimated variance covariance matrix are given by: bβ bβ 2 bβ 3 = 2
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationEC3062 ECONOMETRICS. THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation. (1) y = β 0 + β 1 x β k x k + ε,
THE MULTIPLE REGRESSION MODEL Consider T realisations of the regression equation (1) y = β 0 + β 1 x 1 + + β k x k + ε, which can be written in the following form: (2) y 1 y 2.. y T = 1 x 11... x 1k 1
More informationSTA 302f16 Assignment Five 1
STA 30f16 Assignment Five 1 Except for Problem??, these problems are preparation for the quiz in tutorial on Thursday October 0th, and are not to be handed in As usual, at times you may be asked to prove
More informationBrief Suggested Solutions
DEPARTMENT OF ECONOMICS UNIVERSITY OF VICTORIA ECONOMICS 366: ECONOMETRICS II SPRING TERM 5: ASSIGNMENT TWO Brief Suggested Solutions Question One: Consider the classical T-observation, K-regressor linear
More informationØkonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning
Økonomisk Kandidateksamen 2004 (I) Econometrics 2 Rettevejledning This is a closed-book exam (uden hjælpemidler). Answer all questions! The group of questions 1 to 4 have equal weight. Within each group,
More informationLinear Models and Estimation by Least Squares
Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationMultiple Regression Model: I
Multiple Regression Model: I Suppose the data are generated according to y i 1 x i1 2 x i2 K x ik u i i 1...n Define y 1 x 11 x 1K 1 u 1 y y n X x n1 x nk K u u n So y n, X nxk, K, u n Rks: In many applications,
More informationIn the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)
RNy, econ460 autumn 04 Lecture note Orthogonalization and re-parameterization 5..3 and 7.. in HN Orthogonalization of variables, for example X i and X means that variables that are correlated are made
More informationTMA4255 Applied Statistics V2016 (5)
TMA4255 Applied Statistics V2016 (5) Part 2: Regression Simple linear regression [11.1-11.4] Sum of squares [11.5] Anna Marie Holand To be lectured: January 26, 2016 wiki.math.ntnu.no/tma4255/2016v/start
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationEstadística II Chapter 4: Simple linear regression
Estadística II Chapter 4: Simple linear regression Chapter 4. Simple linear regression Contents Objectives of the analysis. Model specification. Least Square Estimators (LSE): construction and properties
More informationMultivariate Regression Analysis
Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x
More informationSummer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.
Summer School in Statistics for Astronomers V June 1 - June 6, 2009 Regression Mosuk Chow Statistics Department Penn State University. Adapted from notes prepared by RL Karandikar Mean and variance Recall
More informationQuantitative Methods I: Regression diagnostics
Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear
More information