Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Similar documents
Simple Linear Regression

Simple Linear Regression: The Model

INTRODUCTORY ECONOMETRICS

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11

Introduction to Estimation Methods for Time Series models. Lecture 1

Review of Econometrics

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

The Simple Regression Model. Part II. The Simple Regression Model

STAT5044: Regression and Anova. Inyoung Kim

Multivariate Regression Analysis

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

Two-Variable Regression Model: The Problem of Estimation

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

ECON The Simple Regression Model

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Intermediate Econometrics

Introductory Econometrics

Heteroskedasticity. Part VII. Heteroskedasticity

Regression Models - Introduction

2 Regression Analysis

P1.T2. Stock & Watson Chapters 4 & 5. Bionic Turtle FRM Video Tutorials. By: David Harper CFA, FRM, CIPM

Basic Econometrics - rewiev

Regression Models - Introduction

Ordinary Least Squares Regression

Lecture 14 Simple Linear Regression

Homoskedasticity. Var (u X) = σ 2. (23)

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.

L2: Two-variable regression model

Lecture 3: Multiple Regression

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X =

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

Simple Linear Regression Model & Introduction to. OLS Estimation

4 Multiple Linear Regression

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

ECON3150/4150 Spring 2015

the error term could vary over the observations, in ways that are related

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional

. a m1 a mn. a 1 a 2 a = a n

Chapter 14. Linear least squares

Lecture 19 Multiple (Linear) Regression

Statistical View of Least Squares

3 Multiple Linear Regression

Econometrics I Lecture 3: The Simple Linear Regression Model

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Regression #2. Econ 671. Purdue University. Justin L. Tobias (Purdue) Regression #2 1 / 24

ECON 4230 Intermediate Econometric Theory Exam

Ch 2: Simple Linear Regression

FENG CHIA UNIVERSITY ECONOMETRICS I: HOMEWORK 4. Prof. Mei-Yuan Chen Spring 2008

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

HOW IS GENERALIZED LEAST SQUARES RELATED TO WITHIN AND BETWEEN ESTIMATORS IN UNBALANCED PANEL DATA?

Microeconometria Day # 5 L. Cembalo. Regressione con due variabili e ipotesi dell OLS

Weighted Least Squares

Model Mis-specification

Simple Linear Regression

Lecture notes to Stock and Watson chapter 4

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

The Multiple Regression Model Estimation

Heteroskedasticity and Autocorrelation

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Econometrics of Panel Data

Empirical Market Microstructure Analysis (EMMA)

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Econometrics of Panel Data

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

Analisi Statistica per le Imprese

Essential of Simple regression

Heteroscedasticity. Jamie Monogan. Intermediate Political Methodology. University of Georgia. Jamie Monogan (UGA) Heteroscedasticity POLS / 11

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares

Scatter plot of data from the study. Linear Regression

Intermediate Econometrics

Multiple Regression Analysis

Reference: Davidson and MacKinnon Ch 2. In particular page

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Statistics II Exercises Chapter 5

Chapter 2: simple regression model

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Linear Regression Spring 2014

Econometrics Summary Algebraic and Statistical Preliminaries

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Regression and Statistical Inference

BANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1

MIT Spring 2015

The Statistical Property of Ordinary Least Squares

Heteroscedasticity and Autocorrelation

Scatter plot of data from the study. Linear Regression

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

Correlation in Linear Regression

Measuring the fit of the model - SSR

Weighted Least Squares

Simple Linear Regression

Estimation of the Response Mean. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 27

Lecture 24: Weighted and Generalized Least Squares

We can relax the assumption that observations are independent over i = firms or plants which operate in the same industries/sectors

MAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )

Regression Analysis: Basic Concepts

Transcription:

Summary of Part II Key Concepts & Formulas Christopher Ting November 11, 2017 christopherting@smu.edu.sg http://www.mysmu.edu/faculty/christophert/ Christopher Ting 1 of 16

Why Regression Analysis? Understand the relationship between two variables (x i, y i ) of interest: y = Xβ + u (1) Is the relationship captured by β statistically significant? Is the statistical significance robust against heteroskedasticity and correlations? Once β estimate is obtained and given any new information x, what would be the point forecast of y? Also, what is the range of forecast? Christopher Ting 2 of 16

Matrix-Vector Framework Linear regression model of any number of explanatory variables is most general when it is written as y = Xβ + u. (2) The column vector y contains n observations that are to be explained by n K matrix X of observations inclusive of a column vector of ones (K := k + 1). The parameter vector β has K parameters. The noise vector u has rows. Christopher Ting 3 of 16

Regression by Sum of Least Squares of Noise Find β such that the sum of squared noise is as small as possible: β = min β u u = min β ( y Xβ ) ( y Xβ ) (3) Perform vector differentiation with respect to β to obtain the FOC: 2 ( X y X X β ) = 0. The resulting vector of parameter estimates is given by β = ( X X ) 1 X y. (4) Christopher Ting 4 of 16

Variance-Covariance Matrix of β Under the assumptions of homoscedasticity, i.e., V ( u ) = σ 2 ui, and E ( u X ) = 0, the variance-covariance matrix of β is V ( β ) = E ( ( β β )( β β ) ) = σ 2 u( X X ) 1. (5) The variance of β i is the i-th diagonal element of the variancecovariance matrix σ 2 u( X X ) 1. Christopher Ting 5 of 16

Steps to Compute OLS Standard Errors The variance of noise σ 2 u is unknown and it can be estimated as follows: 1. Compute the fitted values ŷy: 2. Compute the residuals or surprise 3. Compute the residual sum of squares (RSS) ŷy = Xβ (6) ûu = y ŷy (7) RSS = ûu ûu = n û i 2 (8) Christopher Ting 6 of 16

4. The variance of the residuals is 5. Let Ω := ( X X ) 1. The variance of βi is σ 2 u = 1 n Kûu ûu (9) V( β i ) = σ 2 uω ii. (10) Christopher Ting 7 of 16

Insight! The first-order conditions can be written as ( ) X y X β = 0, which is The OLS residuals are orthogonal to X. X ûu = 0. (11) Consequently, if X has a column vector of ones, then the average of the residuals is zero. This is because there is one row of ones in X and hence n 1 ûu = û i = 0. Christopher Ting 8 of 16

Ordinary Least Squares (OLS) s Goodness of Fit Explained sum of squares (ESS) is ESS := Total sum of squares (TSS) is n (ŷi y ) 2 TSS := y y = n ( yi y ) 2 = n (ŷi y ) 2 + n û 2 i (12) = ESS + RSS. (13) n The cross term (ŷi y )( ) y i ŷ i is zero because of the orthogonality: X ûu = 0. Christopher Ting 9 of 16

Proof n (ŷi y )( y i ŷ i ) = n (ŷi y ) û i = n ŷ i û i (since y n û i = 0 ) = = n ( ) X β û i i ( ûu X β) = β X ûu = 0. Christopher Ting 10 of 16

Properties of OLS Regression ) The estimates are unbiased, i.e., E ( β = β. The variance of the residuals is unbiased E ( σ 2 u) = σ 2 u. Efficiency: According to the Gauss-Markov theorem, among the classical linear regression models, the OLS estimator is the linear unbiased estimator of β with the minimum variance. Conditional normality β X N ( β, σ 2 u(x X) 1) (14) Christopher Ting 11 of 16

Statistical Inference For all j = 1, 2,..., K, the t test statistic for β j is β j β j σ u Ωjj t n K (15) Here, Ω := ( X X ) 1, and Ωjj is the j-th diagonal element. The α% significance level for β j is β j q σ u Ωjj β j β j + q σ u Ωjj, where q is the ( 1 α/2 ) -th quantile of the t n K distribution. Christopher Ting 12 of 16

Confidence Interval for Mean Response For a given observation x, which is a K 1 vector, the mean response is β x in sample. Given the unbiased estimate β, the variance of β x is V( β ) x ( = V x β ) = x V ( β ) x = σ 2 u x ( X X ) 1 x. (16) Hence, a 100 (1 α)% confidence interval for the in-sample mean response β x is β x ± qσ u x ( X X ) 1 x (17) Christopher Ting 13 of 16

Prediction Interval for a New Observation Suppose a future observation of x is obtained. assumption of u N (0, σu), 2 we have ( V y β ) x = V(u) + V( β ) x Then, by (18) Hence a 100 (1 α)% prediction interval for y is β x ± qσ u 1 + x ( X X ) 1 x (19) Christopher Ting 14 of 16

R 2 and Adjusted R 2 The coefficient of determination R 2 := ESS TSS = 1 RSS TSS (20) Denoted by R 2, the adjusted R 2 is based on the unbiased variances: RSS R 2 = 1 n K (21) TSS n 1 Christopher Ting 15 of 16

Simple Linear Regression: Special Case When K = 2 Slope and intercept estimators b = S xy S xx =: n (x i x)(y i y) n (x i x) 2 ; â = y b x. (22) OLS distribution is â d N a b b, σ 2 u ( ) 1 n + x2 S xx x σu 2 S xx x σu 2 S xx σu 2 S xx (23) Christopher Ting 16 of 16