Week 5 Quantitative Analysis of Financial Markets Modeling and Forecasting Trend

Similar documents
Francis X. Diebold, Elements of Forecasting, 4th Edition

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Introduction to Statistical modeling: handout for Math 489/583

Simple Linear Regression

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

How the mean changes depends on the other variable. Plots can show what s happening...

Machine Learning for OR & FE

Topic 4: Model Specifications

Model Selection. Frank Wood. December 10, 2009

Föreläsning /31

1/34 3/ Omission of a relevant variable(s) Y i = α 1 + α 2 X 1i + α 3 X 2i + u 2i

2.2 Classical Regression in the Time Series Context

Measures of Fit from AR(p)

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Econ 582 Nonparametric Regression

ST430 Exam 2 Solutions

Data Mining Stat 588

Econometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012

Automatic Forecasting

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

Modeling and Forecasting Currency in Circulation in Sri Lanka

Applied Econometrics (QEM)

Univariate ARIMA Models

The Multiple Regression Model

A NOTE ON THE SELECTION OF TIME SERIES MODELS. June 2001

Week 2 Quantitative Analysis of Financial Markets Bayesian Analysis

7. Forecasting with ARIMA models

BNAD 276 Lecture 10 Simple Linear Regression Model

STA121: Applied Regression Analysis

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Chapter 3: Regression Methods for Trends

Problem Set 2: Box-Jenkins methodology

Ch 2: Simple Linear Regression

MLR Model Selection. Author: Nicholas G Reich, Jeff Goldsmith. This material is part of the statsteachr project

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Probabilities & Statistics Revision

Inference for the Regression Coefficient

REVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b)

Chapter 9: Forecasting

ARIMA Modelling and Forecasting

MS&E 226: Small Data

INTRODUCTION TO TIME SERIES ANALYSIS. The Simple Moving Average Model

Machine Learning for OR & FE

Regression, Ridge Regression, Lasso

Chapter 14 Student Lecture Notes 14-1

ISyE 691 Data mining and analytics

Forecast comparison of principal component regression and principal covariate regression

The Multiple Regression Model Estimation

Inferences for Regression

Modelling using ARMA processes

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Multiple Regression Analysis

:Effects of Data Scaling We ve already looked at the effects of data scaling on the OLS statistics, 2, and R 2. What about test statistics?

Week 1 Quantitative Analysis of Financial Markets Distributions A

Machine Learning. Part 1. Linear Regression. Machine Learning: Regression Case. .. Dennis Sun DATA 401 Data Science Alex Dekhtyar..

Chapter 14 Simple Linear Regression (A)

10. Time series regression and forecasting

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

Measuring the fit of the model - SSR

Least Squares Model Averaging. Bruce E. Hansen University of Wisconsin. January 2006 Revised: August 2006

CS 195-5: Machine Learning Problem Set 1

Sparse Linear Models (10/7/13)

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

Generalized Linear Models

Covariance function estimation in Gaussian process regression

Econometría 2: Análisis de series de Tiempo

The regression model with one fixed regressor cont d

Transformations The bias-variance tradeoff Model selection criteria Remarks. Model selection I. Patrick Breheny. February 17

Chapter 4. Regression Models. Learning Objectives

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Introductory Econometrics

Chapter 4: Regression Models

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation

Estimating prediction error in mixed models

Model selection using penalty function criteria

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

Part 1. Multiple Choice (50 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 5 points each)

AR, MA and ARMA models

CHAPTER 8 MODEL DIAGNOSTICS. 8.1 Residual Analysis

Financial Econometrics

Linear model selection and regularization

A Short Introduction to the Lasso Methodology

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

On Autoregressive Order Selection Criteria

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

STAT Financial Time Series

Lab: Box-Jenkins Methodology - US Wholesale Price Indicator

Coefficient of Determination

= observed volume on day l for bin j = base volume in jth bin, and = residual error, assumed independent with mean zero.

Univariate linear models

Econ 423 Lecture Notes: Additional Topics in Time Series 1

1 The Classic Bivariate Least Squares Model

10. Alternative case influence statistics

Unit Root and Cointegration

Statistics 262: Intermediate Biostatistics Model selection

Chapter 7 Student Lecture Notes 7-1

Transcription:

Week 5 Quantitative Analysis of Financial Markets Modeling and Forecasting Trend Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 November 1, 2017 Christopher Ting QF 603 November 1, 2017 1/21

Table of Contents 1 Introduction 2 Modeling Trend 3 Estimating Trend Models 4 Forecasting Models 5 Akaike and Schwarz Criteria 6 Takeaways Christopher Ting QF 603 November 1, 2017 2/21

An Example of Trend a Is it a linear, quadratic, or generally nonlinear trend? a Can the trend be modeled? How? a How can we use the model to forecast future value? Christopher Ting QF 603 November 1, 2017 3/21

QA-10 Modeling and Forecasting Trend Chapter 5. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). a Describe linear and nonlinear trends. a Describe trend models to estimate and forecast trends. a Compare and evaluate model selection criteria, including mean squared error (MSE), s 2 (unbiased variance of residuals), the Akaike information criterion (AIC), and the Schwarz information criterion (SIC). a Explain the necessary conditions for a model selection criterion to demonstrate consistency. Christopher Ting QF 603 November 1, 2017 4/21

Modeling with TIME b Model 1: Trend t is a simple linear function of time Trend t = β 0 + β 1 TIME t, b The (deterministic) variable TIME is constructed artificially and is called a time trend or time dummy. b TIME t = t, where t = 1, 2,..., T. b β 0 is the regression intercept; it s the value of the trend at t = 0. b β 1 is the regression slope positive if the trend is increasing negative if the trend is decreasing. Christopher Ting QF 603 November 1, 2017 5/21

Quadratic b Model 1: Trend t is a quadratic function of time Trend t = β 0 + β 1 TIME t + β 2 TIME 2 t. b β 1 > 0 and β 2 > 0: the trend is monotonically increasing. b β 1 < 0 and β 2 < 0: the trend is monotonically decreasing. b β 1 < 0 and β 2 > 0: the trend has a U shape. b β 1 > 0 and β 2 < 0: the trend has an inverted U shape. Christopher Ting QF 603 November 1, 2017 6/21

Trend of Constant Growth b If trend is characterized by constant growth at rate, then we can write Trend t = β 0 e β 1TIME t. b The trend is a nonlinear (exponential) function of time in levels, but in logarithms we have ln ( Trend t ) = ln(β0 ) + β 1 TIME t. Thus, ln ( Trend t ) is a linear function of time. Christopher Ting QF 603 November 1, 2017 7/21

Estimation g Least squares proceeds by finding the argument (in this case, the value of θ) that minimizes the sum of squared residuals. g Thus, the least squares estimator is the argmin of the sum of squared residuals function. θ = arg min θ T ( yt Trend t (θ) ) 2, where θ denotes the set of parameters to be estimated. Christopher Ting QF 603 November 1, 2017 8/21

Estimation of Linear Trends g Linear trend ( β0, β 1 ) = arg min β 0, β 1 T ( ) 2. yt β 0 β 1 TIME t g Quadratic trend ( β0, β 1, β 2 ) = arg min β 0, β 1, β 2 T ( yt β 0 β 1 TIME t β 2 TIME 2 ) 2. t g Log linear trend ( β0, β 1 ) = arg min β 0, β 1 T ( ) 2. ln yt ln β 0 β 1 TIME t Christopher Ting QF 603 November 1, 2017 9/21

Forecast d The linear trend model, which holds for any time t, is y t = β 0 + β 1 TIME t + ɛ t. d At time T + h, the future time of interest, the forecast made at T is y T +h = β 0 + β 1 TIME T +h + ɛ T +h. d Key idea: TIME T +h is known at time T, because the artificially-constructed time variable is perfectly predictable; specifically, TIME T +h = T + h. d The point forecast is for time T + h and is made at time T. y T +h,t = β 0 + β 1 TIME T +h. Christopher Ting QF 603 November 1, 2017 10/21

Forecast in Practice d Replace unknown parameters with their least squares estimates, yielding the point estimate: ŷ T +h,t = β 0 + β 1 TIME T +h. d To form an interval forecast with 95% confidence, we use ŷ T +h,t ± 1.96 σ, where σ is the standard error of the trend regression. d To form a density forecast, we again assume that the trend regression disturbance is normally distributed N ( ŷ T +h,t, σ 2). Christopher Ting QF 603 November 1, 2017 11/21

Motivating Questions and Problems E How do we select among them when fitting a trend to a specific series? E What are the consequences, for example, of fitting a number of trend models and selecting the model with highest R 2? E It turns out that model-selection strategies such as selecting the model with highest R 2 do not produce good out-of-sample forecasting models. E In-sample overfitting and data mining including more variables in a forecasting model won t necessarily improve its out-of-sample forecasting performance Christopher Ting QF 603 November 1, 2017 12/21

Mean Squared Error (MSE) E Fitted Model 1 E Error ŷ t := β 0 + β 1 TIME t. e t := y t ŷ t, E Definition of MSE where T is the sample size. MSE := T e 2 t T, Christopher Ting QF 603 November 1, 2017 13/21

R 2 and Mean Squared Error E MSE s connection with R 2 T R 2 = 1 e 2 t T ( yt y ). 2 E Mean squared error corrected for degrees of freedom T e 2 t s 2 = T k, where k is the number of degrees of freedom used in model fitting. Christopher Ting QF 603 November 1, 2017 14/21

R 2 and R 2 E s 2 is just the usual unbiased estimate of the regression disturbance variance. That is, it is the square of the usual standard error of the regression. E Connection with adjusted R 2 R 2 = 1 T e 2 t T k T ( yt y ) = 1 2 s 2 T ( yt y ) 2 T 1 T 1 E Note that the denominator depends only on the data want to fit, not the particular model fit. Christopher Ting QF 603 November 1, 2017 15/21

Penalizing Degrees of Freedom E We need to correct somehow for degrees of freedom k when estimating out-of-sample MSE on the basis of in-sample MSE. E To highlight the degree-of-freedom penalty, let s rewrite s 2 as a penalty factor times the MSE, s 2 = ( T ) T k T T e 2 t = 1 1 k T T T e 2 t = 1 1 k MSE. T E Two very important such criteria are the Akaike information criterion (AIC) and the Schwarz information criterion (SIC). Their formulas are AIC = e 2k T MSE and SIC = T k T MSE. Christopher Ting QF 603 November 1, 2017 16/21

Comparison of Information Criteria Christopher Ting QF 603 November 1, 2017 17/21

Definition of Consistency E A model selection criterion is consistent if the following conditions are met: A when the true model that is, the data-generating process (DGP) is among the models considered, the probability of selecting the true DGP approaches 1 as the sample size gets large B when the true model is not among those considered, so that it s impossible to select the true DGP, the probability of selecting the best approximation to the true DGP approaches 1 as the sample size gets large. Christopher Ting QF 603 November 1, 2017 18/21

R 2 Which Criterion Is Consistent? E MSE is inconsistent, because it doesn t penalize for degrees of freedom. E As T increases, s 2 becomes MSE, thus it is not consistent model selection procedure. E The AIC penalizes degrees of freedom more heavily than s 2, but it too remains inconsistent. E Even as the sample size gets large. The AIC selects models that are too large ( overparameterized ). E The SIC, which penalizes degrees of freedom most heavily, is consistent. Christopher Ting QF 603 November 1, 2017 19/21

Asymptotic Efficiency E What if both the true DGP and the best approximation to it are much more complicated than any model we fit? E Definition An asymptotically efficient model selection criterion chooses a sequence of models, as the sample size gets large, whose 1-step-ahead forecast error variances approach the one that would be obtained using the true model with known parameters at a rate at least as fast as that of any other model selection criterion. E The AIC, although inconsistent, is asymptotically efficient, whereas the SIC is not. Christopher Ting QF 603 November 1, 2017 20/21

Takeaways R Trend as a function of time R Forecast: point estimate, interval, density R Model of in-sample fit: MSE, R 2 R Model selection criteria: R 2, s 2, AIC, SIC R Main desired property of a criterion: Consistency Christopher Ting QF 603 November 1, 2017 21/21