Univariate ARIMA Forecasts (Theory)
|
|
- Blaise Higgins
- 5 years ago
- Views:
Transcription
1 Univariate ARIMA Forecasts (Theory) Al Nosedal University of Toronto April 10, 2016 Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
2 All univariate forecasting methods (including ARIMA methods) are based on the same logic. First, the expected value of the time series process is calculated and, second, the expected value is extrapolated into the future. If the current time series observation is Y t, then we are interested in predicting the values of Y t+1, Y t+2,..., Y t+n. We will denote our ARIMA forecast of Y t+n by Y t (n). We call Y t (n) the origin-t forecast of Y with a lead time of n observations. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
3 As a first step in generating an estimate of Y t (n), we calculate the expected value of the Y t process. Our calculations will be simplified considerably if we work in terms of the deviate process, y t. Noting that the Y t and y t processes are related by y t = Y t µ, (where µ = E(Y t )) we can translate our calculations back into the Y t process simply by adding a constant to our result. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
4 Now there are actually two expected values of a time series process which can be used for univariate forecasts: the unconditional and the conditional process expectations. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
5 Example: ARIMA(1,0,0) To illustrate the differences between these two expectations, consider the ARIMA(1, 0, 0) process y t = φ 1 y t 1 + w t (where φ 1 < 1), this process can be expressed as a sum of past shocks (Remember?) y t = w t + φ 1 w t 1 + φ 2 1w t 2 + φ 3 1w t Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
6 Taking the expected value of this expression, E[y t ] = E[w t ] + φ 1 E[w t 1 ] + φ 2 1E[w t 2 ] + φ 3 1E[w t 3 ] +... and thus E[y t ] = 0 E[Y t ] = E[y t ] + µ = µ. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
7 Extrapolating this term into the future, y t (1) = E[y t+1 ] = 0 y t (2) = E[y t+2 ] = 0... y t (n) = E[y t+n ] = 0 When the unconditional expectation of the process is used as a univariate forecast, the process mean is the forecast regardless of lead time. (The problem with forecasts based on the unconditional expectation of a process is that much valuable information is ignored. ) Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
8 The conditional expectation of y t+1 is: E[y t+1 y t, y t 1,..., y 2, y 1 ]. The conditional expectation of y t+1 is conditional upon the t preceding observations of the time series process. (Recalling that y t+1 = w t+1 + φ 1 y t and y t+1 = w t+1 + φ 1 w t + φ 2 1 w t 1 + φ 3 1 w t 2 + φ 4 1 w t ) E[y t+1 y t, y t 1,..., y 2, y 1 ] = E[w t+1 ] + φ 1 y t = φ 1 y t. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
9 Conditional expectation forecasts of the ARIMA(1,0,0) process are, then, y t (1) = E[w t+1 + φ 1 y t ] = φ 1 y t... y t (n) = E[w t+n + φ 1 w t+n φ n 1 1 w t+1 + φ n 1 y t] = φ 1 y t = φ n 1 y t. It should be intuitively plausible that the best forecast of a time series process is the conditional expectation of the process. What we mean by best in this context is that the conditional expectation forecast has the lowest possible mean-square forecast error (MSFE) of any expectation-based forecast. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
10 Using this conditional expectation as a forecast of y t+1, the error in forecasting is: e t+1 = y t+1 y t (1) = w t+1. This error will always be equal to the random shock, w t+1, and the forecast variance is thus VAR(1) = E[w 2 t+1] = σ 2 w, which is the variance of the white noise process. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
11 A 95% interval forecast of y t+1 is thus y t (1) ± 1.96σ w. (We expect y t+1 to lie in this interval 95% of the time). Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
12 If we now wish to forecast the next value of the process, we begin with the expression for y t+2. y t+2 = w t+2 + φ 1 y t+1 y t+2 = w t+2 + φ 1 w t+1 + φ 2 1y t = w t+2 + φ 1 w t+1 + y t (2). Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
13 The error in forecasting is: e t+2 = y t+2 y t (2) = w t+2 + φ 1 w t+1. The forecast variance is thus VAR(2) = E[(w t+2 + φ 1 w t+1 ) 2 ] = σw 2 (1 + φ 2 1). Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
14 A 95% interval forecast of y t+2 is thus y t (2) ± 1.96 σw 2 (1 + φ 2 1 ). (We expect y t+2 to lie in this interval 95% of the time). Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
15 It can be shown that VAR(1) = σw 2 VAR(2) = (1 + φ 2 1 )σ2 w... VAR(n) = (1 + φ φ2n 2 1 )σw 2. Noting that the expression for VAR(n) is a geometric progression, forecast variance approaches a limit of lim n VAR(n) = 1 φ 2 1 which is the variance of the y t autoregressive process. σ2 w Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
16 Example: Moving averages An ARIMA(0, 0, 1) process written as y t = w t + θ 1 w t 1 Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
17 Point forecasts for the ARIMA(0, 0, 1) are: y t (1) = θ 1 w t y t (2) = 0... y t (n) = 0 Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
18 Forecast variance is given by: VAR(1) = σ 2 w VAR(2) = (1 + θ 2 1 )σ2 w... VAR(n) = (1 + θ 2 1 )σ2 w Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
19 After the second step into the future, forecast variance is constant. The limit of VAR(n) is thus lim n VAR(n) = (1 + θ 2 1)σ 2 w, which is the variance of the ARIMA(0, 0, 1) process. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
20 Example: Integrated Processes An ARIMA(0, 1, 0) process, or random walk, written as y t = y t 1 + w t Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
21 Point forecasts for the ARIMA(0, 1, 0) are: y t (1) = y t y t (2) = y t... y t (n) = y t Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
22 Forecast variance is given by: VAR(1) = σ 2 w VAR(2) = 2σ 2 w... VAR(n) = nσ 2 w Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
23 After two or three steps into the future, the confidence intervals become so large as to render the interval forecast meaningless. Al Nosedal University of Toronto Univariate ARIMA Forecasts (Theory) April 10, / 23
Univariate ARIMA Models
Univariate ARIMA Models ARIMA Model Building Steps: Identification: Using graphs, statistics, ACFs and PACFs, transformations, etc. to achieve stationary and tentatively identify patterns and model components.
More informationLecture 2: Univariate Time Series
Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:
More informationChapter 9: Forecasting
Chapter 9: Forecasting One of the critical goals of time series analysis is to forecast (predict) the values of the time series at times in the future. When forecasting, we ideally should evaluate the
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More informationat least 50 and preferably 100 observations should be available to build a proper model
III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or
More informationScenario 5: Internet Usage Solution. θ j
Scenario : Internet Usage Solution Some more information would be interesting about the study in order to know if we can generalize possible findings. For example: Does each data point consist of the total
More informationForecasting. This optimal forecast is referred to as the Minimum Mean Square Error Forecast. This optimal forecast is unbiased because
Forecasting 1. Optimal Forecast Criterion - Minimum Mean Square Error Forecast We have now considered how to determine which ARIMA model we should fit to our data, we have also examined how to estimate
More informationCh 9. FORECASTING. Time Series Analysis
In this chapter, we assume the model is known exactly, and consider the calculation of forecasts and their properties for both deterministic trend models and ARIMA models. 9.1 Minimum Mean Square Error
More information3 Theory of stationary random processes
3 Theory of stationary random processes 3.1 Linear filters and the General linear process A filter is a transformation of one random sequence {U t } into another, {Y t }. A linear filter is a transformation
More informationSpectral Analysis (Theory)
Spectral Analysis (Theory) Al Nosedal University of Toronto Winter 2016 Al Nosedal University of Toronto Spectral Analysis (Theory) Winter 2016 1 / 28 Idea: Decompose a stationary time series {X t } into
More informationCh. 14 Stationary ARMA Process
Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable
More informationLecture 8: ARIMA Forecasting Please read Chapters 7 and 8 of MWH Book
Lecture 8: ARIMA Forecasting Please read Chapters 7 and 8 of MWH Book 1 Predicting Error 1. y denotes a random variable (stock price, weather, etc) 2. Sometimes we want to do prediction (guessing). Let
More informationEcon 424 Time Series Concepts
Econ 424 Time Series Concepts Eric Zivot January 20 2015 Time Series Processes Stochastic (Random) Process { 1 2 +1 } = { } = sequence of random variables indexed by time Observed time series of length
More informationStat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)
Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting) (overshort example) White noise H 0 : Let Z t be the stationary
More informationEE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16
EE539R: Problem Set 3 Assigned: 24/08/6, Due: 3/08/6. Cover and Thomas: Problem 2.30 (Maimum Entropy): Solution: We are required to maimize H(P X ) over all distributions P X on the non-negative integers
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationChapter 5: Models for Nonstationary Time Series
Chapter 5: Models for Nonstationary Time Series Recall that any time series that is a stationary process has a constant mean function. So a process that has a mean function that varies over time must be
More informationResiduals in Time Series Models
Residuals in Time Series Models José Alberto Mauricio Universidad Complutense de Madrid, Facultad de Económicas, Campus de Somosaguas, 83 Madrid, Spain. (E-mail: jamauri@ccee.ucm.es.) Summary: Three types
More informationVector Auto-Regressive Models
Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationVAR Models and Applications
VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationLesson 2: What is a time series Model
Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Time series forecasts Having observed a time series {x 1, x 2,..., x T } over the
More informationECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests
ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Unit Root Tests ECON/FIN
More informationLecture 8 Sampling Theory
Lecture 8 Sampling Theory Thais Paiva STA 111 - Summer 2013 Term II July 11, 2013 1 / 25 Thais Paiva STA 111 - Summer 2013 Term II Lecture 8, 07/11/2013 Lecture Plan 1 Sampling Distributions 2 Law of Large
More informationTrend-Cycle Decompositions
Trend-Cycle Decompositions Eric Zivot April 22, 2005 1 Introduction A convenient way of representing an economic time series y t is through the so-called trend-cycle decomposition y t = TD t + Z t (1)
More informationStatistical Methods for Forecasting
Statistical Methods for Forecasting BOVAS ABRAHAM University of Waterloo JOHANNES LEDOLTER University of Iowa John Wiley & Sons New York Chichester Brisbane Toronto Singapore Contents 1 INTRODUCTION AND
More informationTheoretical and Simulation-guided Exploration of the AR(1) Model
Theoretical and Simulation-guided Exploration of the AR() Model Overview: Section : Motivation Section : Expectation A: Theory B: Simulation Section : Variance A: Theory B: Simulation Section : ACF A:
More information5 Transfer function modelling
MSc Further Time Series Analysis 5 Transfer function modelling 5.1 The model Consider the construction of a model for a time series (Y t ) whose values are influenced by the earlier values of a series
More informationA primer on Structural VARs
A primer on Structural VARs Claudia Foroni Norges Bank 10 November 2014 Structural VARs 1/ 26 Refresh: what is a VAR? VAR (p) : where y t K 1 y t = ν + B 1 y t 1 +... + B p y t p + u t, (1) = ( y 1t...
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition
More informationModeling and forecasting global mean temperature time series
Modeling and forecasting global mean temperature time series April 22, 2018 Abstract: An ARIMA time series model was developed to analyze the yearly records of the change in global annual mean surface
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49
State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationCh 5. Models for Nonstationary Time Series. Time Series Analysis
We have studied some deterministic and some stationary trend models. However, many time series data cannot be modeled in either way. Ex. The data set oil.price displays an increasing variation from the
More information2.1 Linear regression with matrices
21 Linear regression with matrices The values of the independent variables are united into the matrix X (design matrix), the values of the outcome and the coefficient are represented by the vectors Y and
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Fall 24 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing
More informationClassic Time Series Analysis
Classic Time Series Analysis Concepts and Definitions Let Y be a random number with PDF f Y t ~f,t Define t =E[Y t ] m(t) is known as the trend Define the autocovariance t, s =COV [Y t,y s ] =E[ Y t t
More informationMoment Generating Functions
MATH 382 Moment Generating Functions Dr. Neal, WKU Definition. Let X be a random variable. The moment generating function (mgf) of X is the function M X : R R given by M X (t ) = E[e X t ], defined for
More information10.0 REPLICATED FULL FACTORIAL DESIGN
10.0 REPLICATED FULL FACTORIAL DESIGN (Updated Spring, 001) Pilot Plant Example ( 3 ), resp - Chemical Yield% Lo(-1) Hi(+1) Temperature 160 o 180 o C Concentration 10% 40% Catalyst A B Test# Temp Conc
More informationUnit Roots in White Noise?!
Unit Roots in White Noise?! A.Onatski and H. Uhlig September 26, 2008 Abstract We show that the empirical distribution of the roots of the vector auto-regression of order n fitted to T observations of
More information7. Forecasting with ARIMA models
7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability
More informationA Beginner s Introduction. Box-Jenkins Models
A Beginner s Introduction To Box-Jenkins Models I. Introduction In their seminal work, Time Series Analysis: Forecasting and Control (1970, Holden Day), Professors Box and Jenkins introduced a set of time
More informationCh. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations
Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53
State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State
More informationReview: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.
1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately
More informationMarket Risk. MFM Practitioner Module: Quantitiative Risk Management. John Dodson. February 8, Market Risk. John Dodson.
MFM Practitioner Module: Quantitiative Risk Management February 8, 2017 This week s material ties together our discussion going back to the beginning of the fall term about risk measures based on the (single-period)
More informationCovariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )
Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y
More informationSOME BASICS OF TIME-SERIES ANALYSIS
SOME BASICS OF TIME-SERIES ANALYSIS John E. Floyd University of Toronto December 8, 26 An excellent place to learn about time series analysis is from Walter Enders textbook. For a basic understanding of
More informationInterval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean
Interval estimation October 3, 2018 STAT 151 Class 7 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0
More informationARIMA Modelling and Forecasting
ARIMA Modelling and Forecasting Economic time series often appear nonstationary, because of trends, seasonal patterns, cycles, etc. However, the differences may appear stationary. Δx t x t x t 1 (first
More informationUnivariate Nonstationary Time Series 1
Univariate Nonstationary Time Series 1 Sebastian Fossati University of Alberta 1 These slides are based on Eric Zivot s time series notes available at: http://faculty.washington.edu/ezivot Introduction
More informationHeteroskedasticity in Time Series
Heteroskedasticity in Time Series Figure: Time Series of Daily NYSE Returns. 206 / 285 Key Fact 1: Stock Returns are Approximately Serially Uncorrelated Figure: Correlogram of Daily Stock Market Returns.
More informationTime Series of Proportions: A Compositional Approach
Time Series of Proportions: A Compositional Approach C. Barceló-Vidal 1 and L. Aguilar 2 1 Dept. Informàtica i Matemàtica Aplicada, Campus de Montilivi, Univ. de Girona, E-17071 Girona, Spain carles.barcelo@udg.edu
More information1 Class Organization. 2 Introduction
Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat
More informationLecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages
Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages ELEC206 Probability and Random Processes, Fall 2014 Gil-Jin Jang gjang@knu.ac.kr School of EE, KNU page 1 / 15 Chapter 7. Sums of Random
More informationTIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.
TIME SERIES ANALYSIS Forecasting and Control Fifth Edition GEORGE E. P. BOX GWILYM M. JENKINS GREGORY C. REINSEL GRETA M. LJUNG Wiley CONTENTS PREFACE TO THE FIFTH EDITION PREFACE TO THE FOURTH EDITION
More informationESSE Mid-Term Test 2017 Tuesday 17 October :30-09:45
ESSE 4020 3.0 - Mid-Term Test 207 Tuesday 7 October 207. 08:30-09:45 Symbols have their usual meanings. All questions are worth 0 marks, although some are more difficult than others. Answer as many questions
More informationChapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes
Chapter 2 Some basic tools 2.1 Time series: Theory 2.1.1 Stochastic processes A stochastic process is a sequence of random variables..., x 0, x 1, x 2,.... In this class, the subscript always means time.
More informationBasics: Definitions and Notation. Stationarity. A More Formal Definition
Basics: Definitions and Notation A Univariate is a sequence of measurements of the same variable collected over (usually regular intervals of) time. Usual assumption in many time series techniques is that
More informationForecasting Global Recessions in a GVAR Model of Actual and Expected Output in the G7
Forecasting Global Recessions in a GVAR Model of Actual and Expected Output in the G Tony Garratt, Kevin Lee and Kalvinder Shields Ron Smith Conference London, June Lee (Ron Smith Conference) G Recessions
More informationWhite Noise Processes (Section 6.2)
White Noise Processes (Section 6.) Recall that covariance stationary processes are time series, y t, such. E(y t ) = µ for all t. Var(y t ) = σ for all t, σ < 3. Cov(y t,y t-τ ) = γ(τ) for all t and τ
More informationForecasting. Al Nosedal University of Toronto. March 8, Al Nosedal University of Toronto Forecasting March 8, / 80
Forecasting Al Nosedal University of Toronto March 8, 2016 Al Nosedal University of Toronto Forecasting March 8, 2016 1 / 80 Forecasting Methods: An Overview There are many forecasting methods available,
More informationRegression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.
Regression Bivariate i linear regression: Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables and. Generally describe as a
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationMarcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design
Marcel Dettling Institute for Data Analysis and Process Design Zurich University of Applied Sciences marcel.dettling@zhaw.ch http://stat.ethz.ch/~dettling ETH Zürich, March 18, 2013 1 Basics of Modeling
More informationTIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA
CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis
More informationGraduate Econometrics I: Asymptotic Theory
Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory
More informationINTRODUCTION TO TIME SERIES ANALYSIS. The Simple Moving Average Model
INTRODUCTION TO TIME SERIES ANALYSIS The Simple Moving Average Model The Simple Moving Average Model The simple moving average (MA) model: More formally: where t is mean zero white noise (WN). Three parameters:
More informationSTA 2201/442 Assignment 2
STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution
More informationDEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS
ISSN 1440-771X ISBN 0 7326 1085 0 Unmasking the Theta Method Rob J. Hyndman and Baki Billah Working Paper 5/2001 2001 DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS AUSTRALIA Unmasking the Theta method
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationSeasonal Models and Seasonal Adjustment
LECTURE 10 Seasonal Models and Seasonal Adjustment So far, we have relied upon the method of trigonometrical regression for building models which can be used for forecasting seasonal economic time series.
More informationAutoregressive and Moving-Average Models
Chapter 3 Autoregressive and Moving-Average Models 3.1 Introduction Let y be a random variable. We consider the elements of an observed time series {y 0,y 1,y2,...,y t } as being realizations of this randoms
More informationThe hitch in all of this is figuring out the two principal angles and which principal stress goes with which principal angle.
Mohr s Circle The stress basic transformation equations that we developed allowed us to determine the stresses acting on an element regardless of its orientation as long as we know the basic stresses σx,
More information7 Estimation. 7.1 Population and Sample (P.91-92)
7 Estimation MATH1015 Biostatistics Week 7 7.1 Population and Sample (P.91-92) Suppose that we wish to study a particular health problem in Australia, for example, the average serum cholesterol level for
More informationFE570 Financial Markets and Trading. Stevens Institute of Technology
FE570 Financial Markets and Trading Lecture 5. Linear Time Series Analysis and Its Applications (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 9/25/2012
More informationPsychology 282 Lecture #3 Outline
Psychology 8 Lecture #3 Outline Simple Linear Regression (SLR) Given variables,. Sample of n observations. In study and use of correlation coefficients, and are interchangeable. In regression analysis,
More informationEconometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11
Econometrics A Keio University, Faculty of Economics Simple linear model (2) Simon Clinet (Keio University) Econometrics A October 16, 2018 1 / 11 Estimation of the noise variance σ 2 In practice σ 2 too
More informationStatistics 349(02) Review Questions
Statistics 349(0) Review Questions I. Suppose that for N = 80 observations on the time series { : t T} the following statistics were calculated: _ x = 10.54 C(0) = 4.99 In addition the sample autocorrelation
More informationUNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014
UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014 STAD57H3 Time Series Analysis Duration: One hour and fifty minutes Last Name: First Name: Student
More informationCharacteristics of Time Series
Characteristics of Time Series Al Nosedal University of Toronto January 12, 2016 Al Nosedal University of Toronto Characteristics of Time Series January 12, 2016 1 / 37 Signal and Noise In general, most
More informationinterval forecasting
Interval Forecasting Based on Chapter 7 of the Time Series Forecasting by Chatfield Econometric Forecasting, January 2008 Outline 1 2 3 4 5 Terminology Interval Forecasts Density Forecast Fan Chart Most
More informationECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models
ECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Standard Univariate Models ECON/FIN
More informationEconometrics II Heij et al. Chapter 7.1
Chapter 7.1 p. 1/2 Econometrics II Heij et al. Chapter 7.1 Linear Time Series Models for Stationary data Marius Ooms Tinbergen Institute Amsterdam Chapter 7.1 p. 2/2 Program Introduction Modelling philosophy
More informationLecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2
Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y
More informationSTA 6857 VAR, VARMA, VARMAX ( 5.7)
STA 6857 VAR, VARMA, VARMAX ( 5.7) Outline 1 Multivariate Time Series Modeling 2 VAR 3 VARIMA/VARMAX Arthur Berg STA 6857 VAR, VARMA, VARMAX ( 5.7) 2/ 16 Outline 1 Multivariate Time Series Modeling 2 VAR
More informationThe Geometric Random Walk: More Applications to Gambling
MATH 540 The Geometric Random Walk: More Applications to Gambling Dr. Neal, Spring 2008 We now shall study properties of a random walk process with only upward or downward steps that is stopped after the
More informationFinancial Econometrics
Financial Econometrics Multivariate Time Series Analysis: VAR Gerald P. Dwyer Trinity College, Dublin January 2013 GPD (TCD) VAR 01/13 1 / 25 Structural equations Suppose have simultaneous system for supply
More informationECON 616: Lecture 1: Time Series Basics
ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationClass 1: Stationary Time Series Analysis
Class 1: Stationary Time Series Analysis Macroeconometrics - Fall 2009 Jacek Suda, BdF and PSE February 28, 2011 Outline Outline: 1 Covariance-Stationary Processes 2 Wold Decomposition Theorem 3 ARMA Models
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More informationMathematics 1a, Section 4.3 Solutions
Mathematics 1a, Section 4.3 Solutions Alexander Ellis November 30, 2004 1. f(8) f(0) 8 0 = 6 4 8 = 1 4 The values of c which satisfy f (c) = 1/4 seem to be about c = 0.8, 3.2, 4.4, and 6.1. 2. a. g is
More informationTime Series Solutions HT 2009
Time Series Solutions HT 2009 1. Let {X t } be the ARMA(1, 1) process, X t φx t 1 = ɛ t + θɛ t 1, {ɛ t } WN(0, σ 2 ), where φ < 1 and θ < 1. Show that the autocorrelation function of {X t } is given by
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationMathematical statistics
November 15 th, 2018 Lecture 21: The two-sample t-test Overview Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationSimple Linear Regression (Part 3)
Chapter 1 Simple Linear Regression (Part 3) 1 Write an Estimated model Statisticians/Econometricians usually write an estimated model together with some inference statistics, the following are some formats
More informationEcon 4120 Applied Forecasting Methods L10: Forecasting with Regression Models. Sung Y. Park CUHK
Econ 4120 Applied Forecasting Methods L10: Forecasting with Regression Models Sung Y. Park CUHK Conditional forecasting model Forecast a variable conditional on assumptions about other variables. (scenario
More informationIntroduction to the Analysis of Variance (ANOVA)
Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more
More informationSampling Distributions
Sampling Distributions Mathematics 47: Lecture 9 Dan Sloughter Furman University March 16, 2006 Dan Sloughter (Furman University) Sampling Distributions March 16, 2006 1 / 10 Definition We call the probability
More information