Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,
|
|
- Katrina Pamela Lawson
- 6 years ago
- Views:
Transcription
1 Regression Analysis The multiple linear regression model with k explanatory variables assumes that the tth observation of the dependent or endogenous variable y t is described by the linear relationship y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T, where the x ti s, t = 1,..., T, i = 1,..., k, are the independent, exogenous, or explanatory variables. If there is a constant in the regression, then, for example, x t1 = 1 for all t. 1
2 To compactify the notation, define y = β = y 1 y 2. y T β 1 β 2. β k, X =, ϵ = ϵ 1 ϵ 2. ϵ T x 11 x 12 x 1k x 21 x 22 x 2k x T 1 x T 2 x T k,, so that y = Xβ + ϵ. We assume that T > k, and X has full column rank, i.e., rank(x) = k. Then we know that also rank(x X) = rank(x) = k. If there is a constant in the regression, then the first column of X is just a column of ones. 2
3 For the disturbance or error process {ϵ t } T t=1, we assume that (i) E(ϵ X) = 0 (ii) The covariance matrix of the errors is Cov(ϵ X) = σ 2 I T = σ σ σ 2, i.e., the errors are homoskedastic (all have the same variance σ 2 ) and uncorrelated, i.e., E(ϵ t ϵ τ ) = We also frequently assume { σ 2 t = τ 0 t τ. ϵ t iid N(0, σ 2 ), i.e., ϵ N(0, σ 2 I T ), (1) where I T is the identity matrix of dimension T. 3
4 The normality assumption may often been acceptable for monthly returns, but is less adequate for higher frequencies such as weekly or daily. A simple test for normality is provided by the Jarque Bera test (JB), which is JB = T (Ŝ2 6 + ( K ) 3) 2 24 asy χ 2 (2), (2) where Ŝ = T 1 t (R it R i ) 3 T σ i 3, K 1 = t (R it R i ) 4 σ i 4 (3) are the sample skewness and sample kurtosis, respectively. Note that both measures are location and scale free, i.e., not affected by a linear transformation R i = a + br i. 4
5 The skewness checks for asymmetries (recall that the normal distribution is symmetric and has skewness zero). The kurtosis, due to the fourth moment involved, gives very large values a higher weight and thus checks whether the tails of the empirical distribution are thicker than those of the normal, i.e., whether the probability of very large observations (of either sign) is more likely than under normality (which has kurtosis equal to three). A kurtosis significantly larger than three indicates excess kurtosis, i.e., larger probability of large losses and gains. Such a density typically has also more weight in the center and less weight in the shoulders of the distribution. Considerable excess kurtosis is typical for returns at daily or weekly frequencies, which is clearly of great importance for risk management. 5
6 0.35 density with excess kurtosis normal density density return Normality is rejected at the 5% level if JB exceeds the right tail critical value of the χ 2 (2) distribution, which is approximately given by
7 19 DAX stocks, monthly data, January 1995 December 2004, index is the DAX i Company skewness kurtosis JB 1 Allianz BASF BayHypo BMW Bayer CommBank Conti DtBank Eon Henkel Linde Lufthansa MAN RWE Schering Siemens Thyssen TUI VW
8 ,, and indicates significance at the 10%, 5%, and 1% level, respectively. 8
9 Ordinary Least Squares (OLS) Estimation The OLS estimator is defined to be the vector β which minimizes the residual sum of squares, S( β) = (y X β) (y X β) = y y 2y X β + β X X β = ϵ ϵ, where ϵ = y X β. To minimize the sum of squares, we calculate S( β) β = 2Xy + 2X X β = 0, β = (X X) 1 X y. (4) 9
10 For example, consider the single index model (SIM), R it = α i + β i R Mt + ϵ it, t = 1,..., T. (5) In this case, X = 1 R M1 1 R M2.. 1 R MT, y = R i1 R i2. R it, (6) X X = X y = T T R Mt t=1 T t=1 T T R Mt t=1 T R it t=1 RMt 2 t=1 R Mt R it, (7), (8) 10
11 (X X) 1 = [ T RMt RMt R 2 Mt ] 1 [ R 2 Mt R Mt ] = = 1 T R Mt T T RMt 2 ( R Mt ) 2 [ ] R 2 M R M R M 1, s 2 M where R M = T 1 R Mt, R 2 M = T 1 R 2 Mt, and s 2 M = R2 M R2 M = T 1 (R Mt R M ) 2. Hence [ αi β i ] = 1 T = [ R 2 M R M R M 1 s 2 M [ R 2 M R M R M 1 s 2 M ] ] [ [ ] Rit RMt R it R i R M R i ]. 11
12 That is β i = R MR i R i R M s 2 M = T 1 t (R it R i )(R Mt R M ) s 2 M = s im s 2 M. α i = R2 MR i R M R M R i s 2 M = R i(r 2 M R 2 M) + R M (R M R i R M R i ) s 2 M = R i R MR i R M R i s 2 R M M = R i β i R M. 12
13 Goodness of Fit The coefficient of determination is often used to informally judge how well the model explains the observed variations in the endogenous variable. This is based on 1 ϵ ϵ = (y X β)(y X β) where ŷ = X β. = y y 2 β X y + β X X β = y y ŷ ŷ, (9) 1 β X y = βx X(X X) 1 X y = β X X β. 13
14 If there is a constant term in the regression, the residuals sum to zero, and we have y = ŷ, so that s 2 y = s 2 ŷ + s2 ϵ, (10) and the R 2 is R 2 = s2 ŷ s 2 y = 1 s2 ϵ s 2, 0 R 2 1. y In the context of the single index model (SIM), the coefficient of determination can be interpreted as an estimate of the fraction of variability of asset i that is due to movements of the market (proportion of nondiversifiable risk). The remaining part of the variability is then due to nonmarket factors and can be diversified away. If there is no constant term in the regression, we can still define a R 2, but this does not have the same interpretation, as (10) is no longer valid. 14
15 That is, (9) still holds, but it cannot be interpreted in terms of a variance decomposition. 15
16 19 DAX stocks, monthly data, January 1995 December 2004, index is the DAX i Company βi R 2 1 Allianz BASF BayHypo BMW Bayer CommBank Conti DtBank Eon Henkel Linde Lufthansa MAN RWE Schering Siemens Thyssen TUI VW
17 Problem of R 2 is that it does not take into account the number of regressors. The adjusted R 2 is R adj 2 = 1 s2 ϵ T 1 s 2 y T k = 1 (1 R2 ) T 1 T k, which may decrease with the inclusion of a new regressor, and may even become negative (so the term R squared may appear misleading). However, we may want to compare different factor models for returns by applying criteria which are directly related to the prospective use of the model, i.e., out of sample forecasting of the correlation structure. For example, a simple approach is to calculate global minimum variance portfolios over the out of sample period and subsequently compare the realized portfolio variances implied by different models. 17
18 Distribution of β We have β = (X X) 1 X y = (X X) 1 X (Xβ + ϵ) = β + (X X) 1 X ϵ, where ϵ t N(0, σ 2 I T ). It follows that β is normally distributed with mean β (unbiasedness) and covariance matrix (X X) 1 X Cov(ϵ) }{{} =σ 2 I T X(X X) 1 = σ 2 (X X) 1, i.e., β N(β, σ 2 (X X) 1 ). 18
19 Estimation of σ 2 An unbiased estimator of σ 2 is σ 2 = ϵ ϵ T k. (11) Further results (under normality of the errors) are: ϵ ϵ σ 2 χ2 (T k), (12) a χ 2 distribution with T k degrees of freedom. β and σ 2 are independent. 19
20 Testing Hypotheses about β Consider a hypothesis about a single element of β, i.e., H 0 : β i = β 0 i. The variance of β i is σ 2 βi = σ 2 (X X) 1 ii, and, under H 0, β i β 0 i σ 2 βi N(0, 1). (13) As we don t know σ 2, we can use σ 2 instead, but this changes the distribution of (13), due to the sampling variance of σ 2. We have seen that (T k) σ 2 /σ 2 χ 2 (T k). 20
21 Thus, dividing the N(0, 1) variable in (13) by σ2 /σ 2 produces a Student s t distribution with T k degrees of freedom, i.e., the t statistic t = β i βi 0 t(t k), σ 2 βi where σ 2 βi = σ 2 (X X) 1 ii. A symmetric (1 α) confidence interval can be calculated via β i ± t 1 α/2,t k σ 2 β i, (14) where t 1 α/2,t k is the 1 α/2 quantile of the t distribution with T k degrees of freedom. The decision rule for a two sided t test is then to reject the null at level α if the (absolute value of the) t statistic exceeds the (1 α/2) quantile of the Student s t distribution with T k degrees of freedom. 21
22 In the two sided t test above, both positive and negative deviations from the null can lead to a rejection of H 0. That is, we test H 0 : β i = βi 0 H 1 : β i βi 0. against the alternative If we are only interested in deviations in one direction, for example H 1 : β i > βi 0, we refer to a one sided test. The corresponding null hypothesis is H 0 : β i β 0 i and we reject H 0 if t > t 1 α,t k, (15) where t 1 α,t k is the (1 α) quantile of t distribution with T k degrees of freedom. 22
23 Next, consider a hypothesis of the form H 0 : Rβ = r, where R is a q k matrix of rank q. For example, if, in the single index model, we want to test α i = 0 and β i = 1 for asset i, R = [ ] [ 0, r = 1 ]. If H 0 is true, then R β r N(0, σ 2 R(X X) 1 R ). Therefore W = (R β r) [R(X X) 1 R ] 1 (R β r) σ 2 χ 2 (q). 23
24 As before, we cannot use this directly, as σ 2 unknown, but we can use: is Let Y 1 χ 2 (ν 1 ), and Y 2 χ 2 (ν 2 ), and let Y 1 and Y 2 be independent. Then the ratio F = Y 1/ν 1 Y 2 /ν 2 = Y 1 Y 2 ν 2 ν 1 F (ν 1, ν 2 ), i.e., F has an F distribution with ν 1 degrees of freedom in the numerator and ν 2 degrees of freedom in the denominator. Use the fact that (T k) σ 2 /σ 2 χ 2 (T k), and independence of β and σ 2. The test statistic F = (R β r) [R(X X) 1 R ] 1 (R β r) q σ 2 F (q, T k). 24
25 Many alternative representations exist for the F statistic. For example, we can consider the restricted least squares problem, i.e., the minimization of the RSS under the constraint that R β R = r. Working through the first order conditions associated with the Lagrangian L = (y Xβ) (y Xβ) + λ (Rβ r), (16) the restricted estimator is β R = β (X X) 1 R [R(X X) 1 R ] 1 (R β r). Calculations show that ϵ R ϵ R = ϵ ϵ+(r β r) [R(X X) 1 R ] 1 (R β r). 25
26 Therefore, three alternative forms of F with appealing interpretation are F = ϵ R ϵ R ϵ ϵ q σ 2 = (R 2 R 2 R )/q (1 R 2 )/(T k) = ( ϵ R ϵ R ϵ ϵ)/q ϵ ϵ/(t k). = ( β R β)(x X)( β R β) q σ 2. 26
27 Optimality Properties of β 1) Gauss-Markov Theorem The OLS estimator is the Best Linear Unbiased Estimator (BLUE), i.e., for any other unbiased linear estimator β of β, we have that Cov( β) Cov( β) is a positive semidefinite matrix. To see this, let be unbiased for β. Define so that β = Ay B = A (X X) 1 X, β = [B + (X X) 1 X ]y = [B + (X X) 1 X ](Xβ + ϵ) = (BX + I k )β + (B + (X X) 1 X )ϵ. 27
28 Unbiasedness requires BX = 0, so that β = β + (B + (X X) 1 X )ϵ, and the covariance matrix of β is Cov( β) = σ 2 (B + (X X) 1 X )(B + (X X) 1 X ) = σ 2 BB +σ 2 [BX(X X) 1 + (X X) 1 X B ] +σ 2 (X X) 1 = σ 2 [BB + (X X) 1 ] Therefore, the covariance matrix of the competing estimator exceeds that of the OLS estimator by σ 2 BB, which is a positive semidefinite matrix. 28
29 Now consider any linear combination w β of the elements of the parameter vector. The sampling variance of w β is Var(w β) = w [σ 2 BB + σ 2 (X X) 1 ]w σ 2 w (X X) 1 w. This shows that the sampling variance of w β is less than or equal to that of the corresponding combination of any other linear unbiased estimator of β. By choosing w to be the ith unit vector in R k, an implication of this is Var( β i ) Var( β i ), i = 1,..., k. 29
30 2) Given our distributional assumption about the error term (iid normality), we can say even more, by relying on the Cramér Rao lower bound (CRLB). In fact, in this case, the OLS estimator has the smallest variance among all unbiased estimators (not only linear estimators), meaning that it is an UMVUE (Uniformly Minimum Variance Unbiased Estimator). 30
31 Prediction Suppose we want to predict y T +1 = x T +1 β + ϵ T +1, using our OLS fitted model. Our predictor is ŷ T +1 = x T +1 β = xt +1 (X X) 1 X y, which has prediction error x T +1 β yt +1 = x T +1 ( β β) ϵ T +1. This is unbiased in the sense that the prediction error has zero expectation, and the prediction error variance is E[(ŷ T +1 y T +1 ) 2 ] = σ 2 [x T +1 (X X) 1 x T ]. 31
32 This is the best linear unbiased predictor, because a linear unbiased predictor of the form ỹ T +1 = ay can be written as [x T +1 (X X) 1 X + b]y, with b = a x T +1 (X X) 1 X. This has prediction error ỹ T +1 y T +1 = [x T +1 (X X) 1 X + b](xβ + ϵ) x T +1 β ϵ T +1 = bxβ +[x T +1 (X X) 1 X + b]ϵ ϵ T +1, so that bx = 0 is required for unbiasedness. The forecast error variance is then E[(ỹ T +1 y T +1 ) 2 ] = σ 2 (x T +1 (X X) 1 x T +1 + bb + 1) σ 2 (x T +1 (X X) 1 x T ) = E[(ŷ T +1 y T +1 ) 2 ]. 32
33 Generalized Least Squares If ϵ N(0, Σ), where Σ is a covariance matrix, the OLS estimator need no longer be efficient according to the Gauss-Markov theorem. However, we can find nonsingular C such that CC = Σ, for example, by employing the Choleski decomposition, with C nonsingular. Then, with ϵ = C 1 ϵ, we have E( ϵ ϵ ) = C 1 Σ(C ) 1 = C 1 CC (C ) 1 = I. Thus, we may consider the transformed equation C 1 y = C 1 X + C 1 ϵ, with GLS (generalized least squares) estimator β = (X Σ 1 X) 1 X Σ 1 y. 33
34 We will use this when we discuss the Black Litterman approach. 34
Simple Linear Regression
Simple Linear Regression Christopher Ting Christopher Ting : christophert@smu.edu.sg : 688 0364 : LKCSB 5036 January 7, 017 Web Site: http://www.mysmu.edu/faculty/christophert/ Christopher Ting QF 30 Week
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationHypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima
Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s
More informationLECTURE 5 HYPOTHESIS TESTING
October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.
More informationRecent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data
Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data July 2012 Bangkok, Thailand Cosimo Beverelli (World Trade Organization) 1 Content a) Classical regression model b)
More informationApplied Econometrics (QEM)
Applied Econometrics (QEM) based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #3 1 / 42 Outline 1 2 3 t-test P-value Linear
More informationIntroduction to Estimation Methods for Time Series models. Lecture 1
Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation
More informationQuantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017
Summary of Part II Key Concepts & Formulas Christopher Ting November 11, 2017 christopherting@smu.edu.sg http://www.mysmu.edu/faculty/christophert/ Christopher Ting 1 of 16 Why Regression Analysis? Understand
More informationThe Statistical Property of Ordinary Least Squares
The Statistical Property of Ordinary Least Squares The linear equation, on which we apply the OLS is y t = X t β + u t Then, as we have derived, the OLS estimator is ˆβ = [ X T X] 1 X T y Then, substituting
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationLinear Model Under General Variance
Linear Model Under General Variance We have a sample of T random variables y 1, y 2,, y T, satisfying the linear model Y = X β + e, where Y = (y 1,, y T )' is a (T 1) vector of random variables, X = (T
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationTESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST
Econometrics Working Paper EWP0402 ISSN 1485-6441 Department of Economics TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Lauren Bin Dong & David E. A. Giles Department
More informationSpatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved
Spatial Regression 3. Review - OLS and 2SLS Luc Anselin http://spatial.uchicago.edu OLS estimation (recap) non-spatial regression diagnostics endogeneity - IV and 2SLS OLS Estimation (recap) Linear Regression
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More information1. The Multivariate Classical Linear Regression Model
Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The
More informationMultiple Linear Regression
Multiple Linear Regression Asymptotics Asymptotics Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationChapter 15 Panel Data Models. Pooling Time-Series and Cross-Section Data
Chapter 5 Panel Data Models Pooling Time-Series and Cross-Section Data Sets of Regression Equations The topic can be introduced wh an example. A data set has 0 years of time series data (from 935 to 954)
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationSummer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.
Summer School in Statistics for Astronomers V June 1 - June 6, 2009 Regression Mosuk Chow Statistics Department Penn State University. Adapted from notes prepared by RL Karandikar Mean and variance Recall
More informationMFE Financial Econometrics 2018 Final Exam Model Solutions
MFE Financial Econometrics 2018 Final Exam Model Solutions Tuesday 12 th March, 2019 1. If (X, ε) N (0, I 2 ) what is the distribution of Y = µ + β X + ε? Y N ( µ, β 2 + 1 ) 2. What is the Cramer-Rao lower
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationRegression: Ordinary Least Squares
Regression: Ordinary Least Squares Mark Hendricks Autumn 2017 FINM Intro: Regression Outline Regression OLS Mathematics Linear Projection Hendricks, Autumn 2017 FINM Intro: Regression: Lecture 2/32 Regression
More information3. Linear Regression With a Single Regressor
3. Linear Regression With a Single Regressor Econometrics: (I) Application of statistical methods in empirical research Testing economic theory with real-world data (data analysis) 56 Econometrics: (II)
More informationThe Simple Regression Model. Simple Regression Model 1
The Simple Regression Model Simple Regression Model 1 Simple regression model: Objectives Given the model: - where y is earnings and x years of education - Or y is sales and x is spending in advertising
More informationInference in Regression Analysis
Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value
More informationRegression Models - Introduction
Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationMath 423/533: The Main Theoretical Topics
Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationMultivariate Regression Analysis
Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x
More information11 Hypothesis Testing
28 11 Hypothesis Testing 111 Introduction Suppose we want to test the hypothesis: H : A q p β p 1 q 1 In terms of the rows of A this can be written as a 1 a q β, ie a i β for each row of A (here a i denotes
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7
MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationThe Linear Regression Model
The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general
More informationthe error term could vary over the observations, in ways that are related
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance Var(u i x i ) = σ 2 is common to all observations i = 1,..., n In many applications, we may
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationThe Multiple Regression Model Estimation
Lesson 5 The Multiple Regression Model Estimation Pilar González and Susan Orbe Dpt Applied Econometrics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 5 Regression model:
More informationSemester 2, 2015/2016
ECN 3202 APPLIED ECONOMETRICS 2. Simple linear regression B Mr. Sydney Armstrong Lecturer 1 The University of Guyana 1 Semester 2, 2015/2016 PREDICTION The true value of y when x takes some particular
More informationLecture 3: Multiple Regression
Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u
More informationHomoskedasticity. Var (u X) = σ 2. (23)
Homoskedasticity How big is the difference between the OLS estimator and the true parameter? To answer this question, we make an additional assumption called homoskedasticity: Var (u X) = σ 2. (23) This
More informationWISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A
WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2016-17 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationEmpirical Economic Research, Part II
Based on the text book by Ramanathan: Introductory Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 7, 2011 Outline Introduction
More informationAnalysis of Cross-Sectional Data
Analysis of Cross-Sectional Data Kevin Sheppard http://www.kevinsheppard.com Oxford MFE This version: November 8, 2017 November 13 14, 2017 Outline Econometric models Specification that can be analyzed
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationR = µ + Bf Arbitrage Pricing Model, APM
4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.
More informationFinancial Econometrics and Quantitative Risk Managenent Return Properties
Financial Econometrics and Quantitative Risk Managenent Return Properties Eric Zivot Updated: April 1, 2013 Lecture Outline Course introduction Return definitions Empirical properties of returns Reading
More informationSo far our focus has been on estimation of the parameter vector β in the. y = Xβ + u
Interval estimation and hypothesis tests So far our focus has been on estimation of the parameter vector β in the linear model y i = β 1 x 1i + β 2 x 2i +... + β K x Ki + u i = x iβ + u i for i = 1, 2,...,
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More informationSimple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationThe Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility
The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory
More informationMultiple Regression Analysis
Chapter 4 Multiple Regression Analysis The simple linear regression covered in Chapter 2 can be generalized to include more than one variable. Multiple regression analysis is an extension of the simple
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationEconometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012
Econometric Methods Prediction / Violation of A-Assumptions Burcu Erdogan Universität Trier WS 2011/2012 (Universität Trier) Econometric Methods 30.11.2011 1 / 42 Moving on to... 1 Prediction 2 Violation
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationLinear Regression with Time Series Data
u n i v e r s i t y o f c o p e n h a g e n d e p a r t m e n t o f e c o n o m i c s Econometrics II Linear Regression with Time Series Data Morten Nyboe Tabor u n i v e r s i t y o f c o p e n h a g
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationInterpreting Regression Results
Interpreting Regression Results Carlo Favero Favero () Interpreting Regression Results 1 / 42 Interpreting Regression Results Interpreting regression results is not a simple exercise. We propose to split
More informationCHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model
CHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model Prof. Alan Wan 1 / 57 Table of contents 1. Assumptions in the Linear Regression Model 2 / 57
More informationA Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models
Journal of Finance and Investment Analysis, vol.1, no.1, 2012, 55-67 ISSN: 2241-0988 (print version), 2241-0996 (online) International Scientific Press, 2012 A Non-Parametric Approach of Heteroskedasticity
More informationThe regression model with one stochastic regressor.
The regression model with one stochastic regressor. 3150/4150 Lecture 6 Ragnar Nymoen 30 January 2012 We are now on Lecture topic 4 The main goal in this lecture is to extend the results of the regression
More informationECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests
ECON4150 - Introductory Econometrics Lecture 5: OLS with One Regressor: Hypothesis Tests Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 5 Lecture outline 2 Testing Hypotheses about one
More informationReview: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.
1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately
More information5.1 Model Specification and Data 5.2 Estimating the Parameters of the Multiple Regression Model 5.3 Sampling Properties of the Least Squares
5.1 Model Specification and Data 5. Estimating the Parameters of the Multiple Regression Model 5.3 Sampling Properties of the Least Squares Estimator 5.4 Interval Estimation 5.5 Hypothesis Testing for
More informationOSU Economics 444: Elementary Econometrics. Ch.10 Heteroskedasticity
OSU Economics 444: Elementary Econometrics Ch.0 Heteroskedasticity (Pure) heteroskedasticity is caused by the error term of a correctly speciþed equation: Var(² i )=σ 2 i, i =, 2,,n, i.e., the variance
More informationModel Mis-specification
Model Mis-specification Carlo Favero Favero () Model Mis-specification 1 / 28 Model Mis-specification Each specification can be interpreted of the result of a reduction process, what happens if the reduction
More informationSchool of Education, Culture and Communication Division of Applied Mathematics
School of Education, Culture and Communication Division of Applied Mathematics MASTER THESIS IN MATHEMATICS / APPLIED MATHEMATICS Estimation and Testing the Quotient of Two Models by Marko Dimitrov Masterarbete
More informationPANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1
PANEL DATA RANDOM AND FIXED EFFECTS MODEL Professor Menelaos Karanasos December 2011 PANEL DATA Notation y it is the value of the dependent variable for cross-section unit i at time t where i = 1,...,
More informationINTRODUCTORY ECONOMETRICS
INTRODUCTORY ECONOMETRICS Lesson 2b Dr Javier Fernández etpfemaj@ehu.es Dpt. of Econometrics & Statistics UPV EHU c J Fernández (EA3-UPV/EHU), February 21, 2009 Introductory Econometrics - p. 1/192 GLRM:
More informationWISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A
WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2016-17 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This
More informationy ˆ i = ˆ " T u i ( i th fitted value or i th fit)
1 2 INFERENCE FOR MULTIPLE LINEAR REGRESSION Recall Terminology: p predictors x 1, x 2,, x p Some might be indicator variables for categorical variables) k-1 non-constant terms u 1, u 2,, u k-1 Each u
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More information8. Instrumental variables regression
8. Instrumental variables regression Recall: In Section 5 we analyzed five sources of estimation bias arising because the regressor is correlated with the error term Violation of the first OLS assumption
More informationSimultaneous Equation Models Learning Objectives Introduction Introduction (2) Introduction (3) Solving the Model structural equations
Simultaneous Equation Models. Introduction: basic definitions 2. Consequences of ignoring simultaneity 3. The identification problem 4. Estimation of simultaneous equation models 5. Example: IS LM model
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More informationHeteroskedasticity and Autocorrelation
Lesson 7 Heteroskedasticity and Autocorrelation Pilar González and Susan Orbe Dpt. Applied Economics III (Econometrics and Statistics) Pilar González and Susan Orbe OCW 2014 Lesson 7. Heteroskedasticity
More informationEconometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague
Econometrics Week 4 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 23 Recommended Reading For the today Serial correlation and heteroskedasticity in
More informationF3: Classical normal linear rgression model distribution, interval estimation and hypothesis testing
F3: Classical normal linear rgression model distribution, interval estimation and hypothesis testing Feng Li Department of Statistics, Stockholm University What we have learned last time... 1 Estimating
More informationReview of probability and statistics 1 / 31
Review of probability and statistics 1 / 31 2 / 31 Why? This chapter follows Stock and Watson (all graphs are from Stock and Watson). You may as well refer to the appendix in Wooldridge or any other introduction
More informationFinite-sample quantiles of the Jarque-Bera test
Finite-sample quantiles of the Jarque-Bera test Steve Lawford Department of Economics and Finance, Brunel University First draft: February 2004. Abstract The nite-sample null distribution of the Jarque-Bera
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline
More informationLecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2
Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationInstead of using all the sample observations for estimation, the suggested procedure is to divide the data set
Chow forecast test: Instead of using all the sample observations for estimation, the suggested procedure is to divide the data set of N sample observations into N 1 observations to be used for estimation
More information2. A Review of Some Key Linear Models Results. Copyright c 2018 Dan Nettleton (Iowa State University) 2. Statistics / 28
2. A Review of Some Key Linear Models Results Copyright c 2018 Dan Nettleton (Iowa State University) 2. Statistics 510 1 / 28 A General Linear Model (GLM) Suppose y = Xβ + ɛ, where y R n is the response
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More information+ Specify 1 tail / 2 tail
Week 2: Null hypothesis Aeroplane seat designer wonders how wide to make the plane seats. He assumes population average hip size μ = 43.2cm Sample size n = 50 Question : Is the assumption μ = 43.2cm reasonable?
More informationIntroduction to Econometrics Midterm Examination Fall 2005 Answer Key
Introduction to Econometrics Midterm Examination Fall 2005 Answer Key Please answer all of the questions and show your work Clearly indicate your final answer to each question If you think a question is
More informationRoss (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.
4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.
More informationEcon 583 Final Exam Fall 2008
Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random
More informationVector autoregressions, VAR
1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,
More information