MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS)
|
|
- Magnus Clark
- 6 years ago
- Views:
Transcription
1 MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes. Only non programmable calculators are permitted. There are FIFTEEN questions. Cellular phones, unauthorized electronic devices or course notes (unless an open-book exam) are not allowed during this exam. Phones and devices must be turned off and put away in your bag. Do not keep them in your possession, such as in your pockets. If caught with such a device or document, the following may occur: you will be asked to leave immediately the exam, academic fraud allegations will be filed which may result in you obtaining a 0 (zero) for the exam.by signing below, you acknowledge that you have ensured that you are complying with the above statement. Your signature: Question 1 Question 2 Question 3 Question 4 Question 5 Question 6 Question 7 Question 8 Question 9 Question 10 Question 11 Question 12 Question 13 Question 14 Question 15 Total GOOD LUCK!!! Maximal no. of points Your Score 1
2 THIS PAGE IS EMPTY
3 Q1. (4 points) Consider the sequence X t = Z t Z t 1 +Z t 3, where Z t are i.i.d random variables with mean 0 and variance σ 2 Z. Compute E[X t X t+h ] for h = 0,1,3. (3 points) Is the sequence X t stationary? Why? (1 point) Solution to Q1: E[X 2 t ] = σ4 Z +σ2 Z, E[X t X t+1 ] = 0, E[X t X t+3 ] = 0, We also have E[X t ] = 0. Hence the expected value and the covariance do not depend on t. The sequence is stationary. Marking scheme for Q1: Each correct calculation in part a) - 1 point; correct conclusion in part b) - 1 point. Q2. (2 points) Consider ARMA(2, 1) model given by Is this process causal? Is this process stationary? Solution to Q2: The autoroegressive polynomial is The autoregressive polynomial is The roots are X t +0.75X t X t 2 = Z t +2.25Z t 1. φ(z) = z z , One of the solutions lies inside the unit circle. Hence, there is a stationary and a non-casual solution. Marking scheme for Q2: 1 point for correct calculation of the roots, 0.5 point for concluding stationarity, 0.5 point for concluding no-causality. Total number of points: 2. Q3. (2 points) Consider the linear process given by X t = (φ j +φ 2j )Z t j, where φ < 1 and Z t are i.i.d. random variables with mean 0 and variance σ 2 Z. Write the formula for ρ X (10), that is the correlation at lag 10. Note: You can use the formula for the autocovariance of the linear process Solution to Q3:
4 We have for all h 0, Hence, γ X (h) = σ 2 Z = σ 2 Z (φ j +φ 2j )(φ j+h +φ 2(j+h) ) { } φ h φ 2 +φ2h 1 φ 3 +φh 1 φ 3 +φ2h 1 φ 4 ρ X (10) = γ X(10) γ X (0), where γ X (0) and γ X (10) are obtained by substituting h = 0 and h = 10 in the general formula above. Marking scheme for Q3: 1 point for evaluation of γ X (10), one point for γ X (0). If you give just a general formula, you receive one point only. Total number of points: 2.. Q4. (3 points) Consider a stationary process X t = φ 1 X t 1 +φ 2 X t 2 +φ 3 X t 3 +Z t, where Z t are i.i.d random variables with mean 0 and variance σ 2 Z. For this model: Apply the Yule-Walker procedure to obtain P n X n+1. Compute the corresponding MSPE n (1). Solution to Q4: Guess the solution P n X n+3 = φ 1 X n +φ 2 X n 1 +φ 3 X n 3 and verify the Yule-Walker matrix equation. MSPE n (1) = γ X (0) φ 1 γ X (1) φ 2 γ X (2) φ 3 γ X (3). Marking scheme for Q4: Part a) - 1 point for the formula, 1 point for verification. Part b) - 1 point for the correct answer. Q5. (3 points) Consider the MA(1) model X t = Z t +θz t 1, θ R, where Z t are i.i.d. random variables with mean 0 and variance σ 2 Z. For this model: Apply thedurbin-levinsonalgorithmtogetthe coefficientsin thelinearpredictionp 2 X 3 = φ 21 X 2 +φ 22 X 1. (2 points) What are PACF at lags 1 and 2? (1 point) Solution to Q5: φ 21 = θ(1+θ2 ) 1+θ 2 +θ 4, φ 22 = 1+θ 2 +θ 4. φ 11 = θ 1+θ 2, φ 22 = 1+θ 2 +θ 4. θ 2 θ 2
5 Q6. (4 points) Consider a stationary AR(2) model X t = φ 1 X t 1 +φ 2 X t 2 +Z t, where Z t are i.i.d. normal random variables with mean zero and variance σ 2 Z. Assume that φ 1 is known. (2 points) Derive MLE for φ 2. (2 points) Derive MLE for σ 2 Z. Solution to Q6: ˆφ 2 = n t=1 X tx t 2 φ 1 n t=1 X t 1X t 2 n t=1 X2 t 2 Marking scheme for Q6: ˆσ Z 2 = 1 n (X t φ 1 X t 1 n ˆφ ) 2 X t 2 t=1. Part a) - 2 points for the correct answer, part b) - 2 points for the correct answer. Q7. (5 points) Assume that a data set X is well-described by an AR(2) model X t = φ 1 X t 1 +φ 2 X t 1 +Z t, where Z t are i.i.d. random variables with mean 0 and variance σ 2 Z. Below you can find the following output for its ACF: Autocorrelations of series TimeSeries, by lag Also, the sample variance is 1. For this model: Estimate the parameters φ 1, φ 2 and σ 2 Z. (3 points) Is the estimated model stationary? Is the estimated model causal? (2 points) Solution to Q7: Use Yule-Walker equation (φ 1,φ 2 ) T = Γ 1 2 (γ X(1),γ X (2)) T to obtain ˆφ 1 = , ˆφ2 = 0.104, ˆσ 2 Z = 0.4. Typical mistake: You used Γ 2 instead of Γ 1 2. The roots of the autoregressive polynomial are 7.87 and The model is causal and stationary. Marking scheme for Q7: Part a) - 1 point for each correct answer; part b) 1 point the roots, 1 point for correct conclusion on stationarity and causality. Q8. (6 points) We have AR(1) time series with the following characteristics: ACF:. Autocorrelations of series Data, by lag > length(data) [1] 250 > mean(data); var(data) [1] , > Data[250] [1]
6 Find the Yule-Walker estimates of φ and σ 2 Z. Compute 95% confidence interval for φ. Note: z = (c) Predict the 251st value. (d) What is the mean square error of the prediction in part (c)? (e) What is the estimated value of covariance at lag 7, i.e. ˆγ X (7)? Solution to Q8: ˆφ±z (c) If X is my time series (non-centered) then (d) (e) φ = 0.645, σ 2 Z = ˆσ Z ˆγ X (0) = (0.55,0.73) P n X n+1 = X(1 ˆφ)+ ˆφX 250 = ˆγ X (0) ˆφ ˆγ X (1) = ˆγ X (7) = Marking scheme for Q8: Part a) - 1 point for each correct answer; all other pars - 1 point. Q9. (5 points) We consider Yule-Walker estimation for AR(p) models. Consider AR(2) model X t = φ 1 X t 1 + φ 2 X t 2 + Z t, where Z t are i.i.d random variables with mean 0 and variance σ 2 Z. Write the theoretical formula for 95% confidence intervals for ˆφ 1 and ˆφ 2. Note that the appropriate normal quantile is (You do not need to derive the confidence intervals). Assume that a data set X is well-described by an AR(2) model. Below you can find the following output for its ACF: Autocorrelations of series MyTimeSeries, by lag Also, the estimated variance of the time series is Use your results from part to calculate the 95% confidence intervals for ˆφ 1 and ˆφ 2. (c) Calculate ˆσ 2 Z. Solution to Q9: ˆφ 1 ± n 1 ˆφ 2, ˆφ 2 ± n 1 ˆφ 2 (c) ˆφ 1 = 0.407, ˆφ2 = ˆσ 2 Z =
7 Marking scheme for Q9: Part a) - 1 point for each correct answer; part b) - 1 point for each correct answer; part c) - 1 point. Q10. (1 point) For a given time series X t we fitted AR(1) model. We estimated φ using the Yule-Walker estimator and computed residuals as R t = X t+1 ˆφX t. The following graph shows ACF and PACF of residuals. Is the fit appropriate? Series fit$resid[2:100] Series fit$resid[2:100] ACF Partial ACF Lag Lag Solution to Q10: Since there is no dependence left in residuals, the fit is appropriate.
8 Q11. (2 points) Consider the following sequence Y 2 t = 10 φ j Z 2 t Z2 t 1 Z2 t j, t = 1,2,..., where Z t, t = 1,2,..., are independent standard normal random variables. Compute E[Y 2 t ]. Solution to Q11: 10 E[Yt 2 ] = φ j = 1 φ11 1 φ. Q12. (6 points) Consider the following ARCH(1) process X t = σ t Z t, σ 2 t = X2 t 1, where Z t are i.i.d. random variables with mean 0 and variance σ 2 = 1. (1 point) Is the model stationary? Why? (1 points) Calculate E(X t ). (c) (2 points) Calculate Var(X t ). (d) (2 points) Assume furthermore that Z t are normal. Calculate E[X t 1 X 4 t ]. Solution to Q12: Stationary since α 1 < 1; E[X t ] = E[σ t ]E[Z t ] = 0. (c) We can use the linear representation: Xt 2 = α j 1 Z2 tzt 1 Z 2 t j 2, t = 1,2,..., (d) to get E[X t 1 X 4 t ] = E[X t 1σ 4 t ]E[Z4 t ] = 3E[X t 1σ 4 t ] Var[X t ] = E[X 2 t ] = 1 1 α 1 = = 10. = 3E[X t 1 ( X 2 t 1 )2 ] = E[X t 1 ] E[X 3 t 1 ] E[X 5 t 1 ] = 0 since for an odd q > 0 we have E[X q t 1 ] = E[σq t 1 ]E[Zq t 1 ] = 0. Q13. (2 points) To a data set we fitted an ARCH(1) model. The estimated parameters are as follows: a0 a Is the model stationary? Why? (1 point) Predict the next value of the squared volatility. Note that the last observation from the data sequence is , whereas the last fitted value of volatility is (1 point) Solution to Q13: Yes since α 1 < 1; ˆσ 2 n+1 = ( )2 =
9 Q14. (3 points) To a data set we fitted an GARCH(1,1) model. The estimated parameters are as follows: a0 a1 b e Is the model stationary? Why? (1 point) The last observation from the data sequence X is X 1000 = We also know that ˆσ 1000 = Predict the next value of the squared volatility σ (2 points) Solution to Q14: The model stationary because α 1 +β 1 < ˆσ = e Q15. (3 points) Assume that Z t = (Z t1,z t2 ) is an i.i.d. sequence of random vectors with the mean vector 0 and the covariance matrix [ ] Σ = Define the bivariate linear process by X t = Z t +ΨZ t 1, where [ ] 1 0 Ψ =. 1 1 Find Γ(0), Γ(1) and Γ( 1), where Γ(h) is the covariance matrix function of the bivariate time series X t. Solution to Q15: We have [ X t = (X t1,x t2 ) T = ] Z t1 +Z t 1,1. Z t2 +Z t 1,1 +Z t 1,2 and [ ] [ ] Var(X Γ(0) = t1 ) Cov(X t1,x t2 ) 2 2 =. Cov(X t2,x t1 ) Var(X t2 ) 2 6 [ ] [ ] Cov(Xt1,X Γ(1) = t+1,1 ) Cov(X t1,x t+1,2 ) =. Cov(X t2,x t+1,1 ) Cov(X t2,x t+1,2 ) [ ] [ ] Cov(Xt1,X Γ( 1) = t 1,1 ) Cov(X t1,x t 1,2 ) =. Cov(X t2,x t 1,1 ) Cov(X t2,x t 1,2 ) For example, how to get 6 above: Var(Z t2 +Z t 1,1 +Z t 1,2 ) = Var(Z t2 )+Var(Z t 1,1 )+Var(Z t 1,2 )+2Cov(Z t 1,1,Z t 1,2 ) = = 6. THIS IS THE LAST QUESTION
MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)
MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes. Only
More informationMAT3379 (Winter 2016)
MAT3379 (Winter 2016) Assignment 4 - SOLUTIONS The following questions will be marked: 1a), 2, 4, 6, 7a Total number of points for Assignment 4: 20 Q1. (Theoretical Question, 2 points). Yule-Walker estimation
More informationSTAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong
STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X
More information3. ARMA Modeling. Now: Important class of stationary processes
3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average
More informationMAT 2377C FINAL EXAM PRACTICE
Department of Mathematics and Statistics University of Ottawa MAT 2377C FINAL EXAM PRACTICE 10 December 2015 Professor: Rafal Kulik Time: 180 minutes Student Number: Family Name: First Name: This is a
More informationModule 3. Descriptive Time Series Statistics and Introduction to Time Series Models
Module 3 Descriptive Time Series Statistics and Introduction to Time Series Models Class notes for Statistics 451: Applied Time Series Iowa State University Copyright 2015 W Q Meeker November 11, 2015
More informationUniversity of Ottawa
University of Ottawa Department of Mathematics and Statistics MAT 1302B : Mathematical Methods II Professor: Hadi Salmasian First Midterm Exam Version A February 3, 2017 Surname First Name Student # DGD
More informationParameter estimation: ACVF of AR processes
Parameter estimation: ACVF of AR processes Yule-Walker s for AR processes: a method of moments, i.e. µ = x and choose parameters so that γ(h) = ˆγ(h) (for h small ). 12 novembre 2013 1 / 8 Parameter estimation:
More informationProblem Set 1 Solution Sketches Time Series Analysis Spring 2010
Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically
More informationPart III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to
TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let
More informationECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models
ECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Standard Univariate Models ECON/FIN
More informationMATH 5075: Time Series Analysis
NAME: MATH 5075: Time Series Analysis Final For the entire test {Z t } WN(0, 1)!!! 1 1) Let {Y t, t Z} be a stationary time series with EY t = 0 and autocovariance function γ Y (h). Assume that a) Show
More informationReview Session: Econometrics - CLEFIN (20192)
Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =
More informationApplied time-series analysis
Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,
More informationEASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION
ARIMA MODELS: IDENTIFICATION A. Autocorrelations and Partial Autocorrelations 1. Summary of What We Know So Far: a) Series y t is to be modeled by Box-Jenkins methods. The first step was to convert y t
More informationContents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8
A N D R E W T U L L O C H T I M E S E R I E S A N D M O N T E C A R L O I N F E R E N C E T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Time Series Analysis 5
More informationCovariances of ARMA Processes
Statistics 910, #10 1 Overview Covariances of ARMA Processes 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation
More informationMarcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design
Marcel Dettling Institute for Data Analysis and Process Design Zurich University of Applied Sciences marcel.dettling@zhaw.ch http://stat.ethz.ch/~dettling ETH Zürich, March 18, 2013 1 Basics of Modeling
More informationModule 4. Stationary Time Series Models Part 1 MA Models and Their Properties
Module 4 Stationary Time Series Models Part 1 MA Models and Their Properties Class notes for Statistics 451: Applied Time Series Iowa State University Copyright 2015 W. Q. Meeker. February 14, 2016 20h
More informationCircle the single best answer for each multiple choice question. Your choice should be made clearly.
TEST #1 STA 4853 March 6, 2017 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. There are 32 multiple choice
More informationNonlinear time series
Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of
More informationIntroduction to Time Series Analysis. Lecture 11.
Introduction to Time Series Analysis. Lecture 11. Peter Bartlett 1. Review: Time series modelling and forecasting 2. Parameter estimation 3. Maximum likelihood estimator 4. Yule-Walker estimation 5. Yule-Walker
More informationIntroduction to ARMA and GARCH processes
Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,
More informationLecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications
Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive
More informationEconometric Forecasting
Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend
More informationCircle a single answer for each multiple choice question. Your choice should be made clearly.
TEST #1 STA 4853 March 4, 215 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. There are 31 questions. Circle
More informationUNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014
UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014 STAD57H3 Time Series Analysis Duration: One hour and fifty minutes Last Name: First Name: Student
More information18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013
18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013 1. Covariance Stationary AR(2) Processes Suppose the discrete-time stochastic process {X t } follows a secondorder auto-regressive process
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 III. Stationary models 1 Purely random process 2 Random walk (non-stationary)
More informationThe autocorrelation and autocovariance functions - helpful tools in the modelling problem
The autocorrelation and autocovariance functions - helpful tools in the modelling problem J. Nowicka-Zagrajek A. Wy lomańska Institute of Mathematics and Computer Science Wroc law University of Technology,
More informationChapter 4: Models for Stationary Time Series
Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t
More informationSTAT Financial Time Series
STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR
More informationAn introduction to volatility models with indices
Applied Mathematics Letters 20 (2007) 177 182 www.elsevier.com/locate/aml An introduction to volatility models with indices S. Peiris,A.Thavaneswaran 1 School of Mathematics and Statistics, The University
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Fall 24 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing
More informationITSM-R Reference Manual
ITSM-R Reference Manual George Weigt February 11, 2018 1 Contents 1 Introduction 3 1.1 Time series analysis in a nutshell............................... 3 1.2 White Noise Variance.....................................
More informationEmpirical Market Microstructure Analysis (EMMA)
Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg
More informationUniversity of Ottawa
University of Ottawa Department of Mathematics and Statistics MAT 1302A: Mathematical Methods II Instructor: Hadi Salmasian Final Exam April 2016 Surname First Name Seat # Instructions: (a) You have 3
More informationSTAD57 Time Series Analysis. Lecture 8
STAD57 Time Series Analysis Lecture 8 1 ARMA Model Will be using ARMA models to describe times series dynamics: ( B) X ( B) W X X X W W W t 1 t1 p t p t 1 t1 q tq Model must be causal (i.e. stationary)
More informationChapter 6: Model Specification for Time Series
Chapter 6: Model Specification for Time Series The ARIMA(p, d, q) class of models as a broad class can describe many real time series. Model specification for ARIMA(p, d, q) models involves 1. Choosing
More informationUniversity of Oxford. Statistical Methods Autocorrelation. Identification and Estimation
University of Oxford Statistical Methods Autocorrelation Identification and Estimation Dr. Órlaith Burke Michaelmas Term, 2011 Department of Statistics, 1 South Parks Road, Oxford OX1 3TG Contents 1 Model
More informationSTAT 443 (Winter ) Forecasting
Winter 2014 TABLE OF CONTENTS STAT 443 (Winter 2014-1141) Forecasting Prof R Ramezan University of Waterloo L A TEXer: W KONG http://wwkonggithubio Last Revision: September 3, 2014 Table of Contents 1
More informationMidterm Suggested Solutions
CUHK Dept. of Economics Spring 2011 ECON 4120 Sung Y. Park Midterm Suggested Solutions Q1 (a) In time series, autocorrelation measures the correlation between y t and its lag y t τ. It is defined as. ρ(τ)
More informationLesson 9: Autoregressive-Moving Average (ARMA) models
Lesson 9: Autoregressive-Moving Average (ARMA) models Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Introduction We have seen
More informationSupplemental Exam CHM1321-B. Professor Sandro Gambarotta. Date:_6-July-2017 Length: 3 hrs Last Name:
1 Supplemental Exam CHM1321-B Professor Sandro Gambarotta Date:_6-July-2017 Length: 3 hrs Last Name: First Name: Student # Seat # - Instructions: - Calculator permitted (Faculty approved or any other non-programmable
More informationStationary Stochastic Time Series Models
Stationary Stochastic Time Series Models When modeling time series it is useful to regard an observed time series, (x 1,x,..., x n ), as the realisation of a stochastic process. In general a stochastic
More informationAutoregressive Moving Average (ARMA) Models and their Practical Applications
Autoregressive Moving Average (ARMA) Models and their Practical Applications Massimo Guidolin February 2018 1 Essential Concepts in Time Series Analysis 1.1 Time Series and Their Properties Time series:
More informationExamination paper for Solution: TMA4285 Time series models
Department of Mathematical Sciences Examination paper for Solution: TMA4285 Time series models Academic contact during examination: Håkon Tjelmeland Phone: 4822 1896 Examination date: December 7th 2013
More informationNANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS
NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION 2012-2013 MAS451/MTH451 Time Series Analysis May 2013 TIME ALLOWED: 2 HOURS INSTRUCTIONS TO CANDIDATES 1. This examination paper contains FOUR (4)
More informationWe will only present the general ideas on how to obtain. follow closely the AR(1) and AR(2) cases presented before.
ACF and PACF of an AR(p) We will only present the general ideas on how to obtain the ACF and PACF of an AR(p) model since the details follow closely the AR(1) and AR(2) cases presented before. Recall that
More informationUniversity of Ottawa
University of Ottawa Department of Mathematics and Statistics MAT 0A: Mathematical Methods II Instructor: Hadi Salmasian Final Exam Solutions April 08 Surname First Name Student # Seat # Instructions:
More informationCh 4. Models For Stationary Time Series. Time Series Analysis
This chapter discusses the basic concept of a broad class of stationary parametric time series models the autoregressive moving average (ARMA) models. Let {Y t } denote the observed time series, and {e
More information6.3 Forecasting ARMA processes
6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss
More informationLecture 1: Fundamental concepts in Time Series Analysis (part 2)
Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC)
More information{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }
Stochastic processes Time series are an example of a stochastic or random process Models for time series A stochastic process is 'a statistical phenomenon that evolves in time according to probabilistic
More informationCovariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )
Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 II. Basic definitions A time series is a set of observations X t, each
More informationγ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1
4.2 Autoregressive (AR) Moving average models are causal linear processes by definition. There is another class of models, based on a recursive formulation similar to the exponentially weighted moving
More informationA time series is called strictly stationary if the joint distribution of every collection (Y t
5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a
More informationFinal Examination 7/6/2011
The Islamic University of Gaza Faculty of Commerce Department of Economics & Applied Statistics Time Series Analysis - Dr. Samir Safi Spring Semester 211 Final Examination 7/6/211 Name: ID: INSTRUCTIONS:
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationSTOR 356: Summary Course Notes
STOR 356: Summary Course Notes Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC 7599-360 rls@email.unc.edu February 19, 008 Course text: Introduction
More informationTime Series Analysis -- An Introduction -- AMS 586
Time Series Analysis -- An Introduction -- AMS 586 1 Objectives of time series analysis Data description Data interpretation Modeling Control Prediction & Forecasting 2 Time-Series Data Numerical data
More information2. An Introduction to Moving Average Models and ARMA Models
. An Introduction to Moving Average Models and ARMA Models.1 White Noise. The MA(1) model.3 The MA(q) model..4 Estimation and forecasting of MA models..5 ARMA(p,q) models. The Moving Average (MA) models
More informationLast name: First name: Signature: Student number:
MAT 2141 The final exam Instructor: K. Zaynullin Last name: First name: Signature: Student number: Do not detach the pages of this examination. You may use the back of the pages as scrap paper for calculations,
More informationNext tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2
Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column
More informationTMA4285 December 2015 Time series models, solution.
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 5 TMA4285 December 205 Time series models, solution. Problem a) (i) The slow decay of the ACF of z t suggest that
More informationProblem Set 2 Solution Sketches Time Series Analysis Spring 2010
Problem Set 2 Solution Sketches Time Series Analysis Spring 2010 Forecasting 1. Let X and Y be two random variables such that E(X 2 ) < and E(Y 2 )
More informationDiscrete time processes
Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following
More informationTime Series 2. Robert Almgren. Sept. 21, 2009
Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models
More informationSTA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)
STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) Outline 1 Announcements 2 Autocorrelation and Cross-Correlation 3 Stationary Time Series 4 Homework 1c Arthur Berg STA
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Chapter 9 Multivariate time series 2 Transfer function
More informationUniversity of Ottawa
University of Ottawa Department of Mathematics and Statistics MAT B: Mathematical Methods II Instructor: Hadi Salmasian Final Exam Solutions April 7 Surname First Name Student # Seat # Instructions: (a)
More informationPart II. Time Series
Part II Time Series 12 Introduction This Part is mainly a summary of the book of Brockwell and Davis (2002). Additionally the textbook Shumway and Stoffer (2010) can be recommended. 1 Our purpose is to
More informationStochastic Processes: I. consider bowl of worms model for oscilloscope experiment:
Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing
More informationLecture 2: ARMA(p,q) models (part 2)
Lecture 2: ARMA(p,q) models (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.
More informationAkaike criterion: Kullback-Leibler discrepancy
Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ψ), ψ Ψ}, Kullback-Leibler s index of f ( ; ψ) relative to f ( ; θ) is (ψ
More informationTime Series Solutions HT 2009
Time Series Solutions HT 2009 1. Let {X t } be the ARMA(1, 1) process, X t φx t 1 = ɛ t + θɛ t 1, {ɛ t } WN(0, σ 2 ), where φ < 1 and θ < 1. Show that the autocorrelation function of {X t } is given by
More informationLECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.
LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation
More informationWe use the centered realization z t z in the computation. Also used in computing sample autocovariances and autocorrelations.
Stationary Time Series Models Part 1 MA Models and Their Properties Class notes for Statistics 41: Applied Time Series Ioa State University Copyright 1 W. Q. Meeker. Segment 1 ARMA Notation, Conventions,
More informationTime Series Analysis
Time Series Analysis Christopher Ting http://mysmu.edu.sg/faculty/christophert/ christopherting@smu.edu.sg Quantitative Finance Singapore Management University March 3, 2017 Christopher Ting Week 9 March
More informationBooth School of Business, University of Chicago Business 41914, Spring Quarter 2017, Mr. Ruey S. Tsay Midterm
Booth School of Business, University of Chicago Business 41914, Spring Quarter 2017, Mr. Ruey S. Tsay Midterm Chicago Booth Honor Code: I pledge my honor that I have not violated the Honor Code during
More information3 Theory of stationary random processes
3 Theory of stationary random processes 3.1 Linear filters and the General linear process A filter is a transformation of one random sequence {U t } into another, {Y t }. A linear filter is a transformation
More informationTime Series I Time Domain Methods
Astrostatistics Summer School Penn State University University Park, PA 16802 May 21, 2007 Overview Filtering and the Likelihood Function Time series is the study of data consisting of a sequence of DEPENDENT
More informationExercises - Time series analysis
Descriptive analysis of a time series (1) Estimate the trend of the series of gasoline consumption in Spain using a straight line in the period from 1945 to 1995 and generate forecasts for 24 months. Compare
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationNon-Stationary Time Series and Unit Root Testing
Econometrics II Non-Stationary Time Series and Unit Root Testing Morten Nyboe Tabor Course Outline: Non-Stationary Time Series and Unit Root Testing 1 Stationarity and Deviation from Stationarity Trend-Stationarity
More informationEconometrics II Heij et al. Chapter 7.1
Chapter 7.1 p. 1/2 Econometrics II Heij et al. Chapter 7.1 Linear Time Series Models for Stationary data Marius Ooms Tinbergen Institute Amsterdam Chapter 7.1 p. 2/2 Program Introduction Modelling philosophy
More informationStat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)
Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting) (overshort example) White noise H 0 : Let Z t be the stationary
More informationTIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets
TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:
More informationCOMPUTER SESSION 3: ESTIMATION AND FORECASTING.
UPPSALA UNIVERSITY Department of Mathematics JR Analysis of Time Series 1MS014 Spring 2010 COMPUTER SESSION 3: ESTIMATION AND FORECASTING. 1 Introduction The purpose of this exercise is two-fold: (i) By
More informationLINEAR STOCHASTIC MODELS
LINEAR STOCHASTIC MODELS Let {x τ+1,x τ+2,...,x τ+n } denote n consecutive elements from a stochastic process. If their joint distribution does not depend on τ, regardless of the size of n, then the process
More informationSome Time-Series Models
Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random
More informationARMA models with time-varying coefficients. Periodic case.
ARMA models with time-varying coefficients. Periodic case. Agnieszka Wy lomańska Hugo Steinhaus Center Wroc law University of Technology ARMA models with time-varying coefficients. Periodic case. 1 Some
More informationRoss Bettinger, Analytical Consultant, Seattle, WA
ABSTRACT DYNAMIC REGRESSION IN ARIMA MODELING Ross Bettinger, Analytical Consultant, Seattle, WA Box-Jenkins time series models that contain exogenous predictor variables are called dynamic regression
More informationModel selection using penalty function criteria
Model selection using penalty function criteria Laimonis Kavalieris University of Otago Dunedin, New Zealand Econometrics, Time Series Analysis, and Systems Theory Wien, June 18 20 Outline Classes of models.
More informationMODELING INFLATION RATES IN NIGERIA: BOX-JENKINS APPROACH. I. U. Moffat and A. E. David Department of Mathematics & Statistics, University of Uyo, Uyo
Vol.4, No.2, pp.2-27, April 216 MODELING INFLATION RATES IN NIGERIA: BOX-JENKINS APPROACH I. U. Moffat and A. E. David Department of Mathematics & Statistics, University of Uyo, Uyo ABSTRACT: This study
More informationFinal Exam CHM1321-B. Date: April 19 Length: 3 hrs Last Name: Professor Sandro Gambarotta Name:
Faculté des sciences Faculty of Science Département de chimie et science biomoléculaire Department of Chemistry and Biomolecular Sciences Pavillon d Iorio Hall 10 Marie-Curie Ottawa ON Canada K1N 6N5 '
More informationIntroduction to Time Series Analysis. Lecture 7.
Last lecture: Introduction to Time Series Analysis. Lecture 7. Peter Bartlett 1. ARMA(p,q) models: stationarity, causality, invertibility 2. The linear process representation of ARMA processes: ψ. 3. Autocovariance
More informationIntroduction to Time Series Analysis. Lecture 12.
Last lecture: Introduction to Time Series Analysis. Lecture 12. Peter Bartlett 1. Parameter estimation 2. Maximum likelihood estimator 3. Yule-Walker estimation 1 Introduction to Time Series Analysis.
More informationExtremogram and Ex-Periodogram for heavy-tailed time series
Extremogram and Ex-Periodogram for heavy-tailed time series 1 Thomas Mikosch University of Copenhagen Joint work with Richard A. Davis (Columbia) and Yuwei Zhao (Ulm) 1 Jussieu, April 9, 2014 1 2 Extremal
More information