You must continuously work on this project over the course of four weeks.
|
|
- Darlene Ramsey
- 5 years ago
- Views:
Transcription
1 The project Five project topics are described below. You should choose one the projects. Maximum of two people per project is allowed. If two people are working on a topic they are expected to do double the work as compared to one person on a project. You must continuously work on this project over the course of four weeks. Each week on Monday you should post on a wikipage (that I will create) the work you have done on it so far. Everyone in the class can view your work. The wiki link (you need to use VPN to access this page from outside the university) By the end of the semester you are expected to submit a final report on the topic and present a talk on the topic. The literature I give is only a starting point. You are expected to find more papers and or books related to the subject and use them in your report and talk. Provisionally the talks will be presented on the reading data Thursday 6th December (if there are no objections). 1
2 Topic 1: Integer Valued Autoregressive Processes INAR(p) Literature An introduction to Discrete valued time series, Christian Weiss. Here is the online book: (Weiss) You may have to be within the university domain to access it. Review paper by Eddie McKenzie (2000), Discrete Variate Time Series (Mckenzie: Here) Typically we assume that time series is continuous. But in many cases the time series is will be positive and integer valued. If the highest integer value is large this can often be approximated with as a continuous value time series. However, this may not always be hold. The plot in Figure 1 is an example from Weiss (2018). Figure 1: Real data. Several authors including E. McKenzie and C. Weiss have proposed methods for modelling such data. We describe a GLM type approach in Topic 5. But here we discuss an approach which appears to be a close generalisation of the classical autoregressive process. To do this, 2
3 we first define the thinning operator is known as the thinning operator and a Z is defined Z t a Z t = B t,j Bin(a, Z t ) j=1 where {B t,j } Bernoulli random variables with P (B t,j = 1) = a. Clearly 0 a 1 and 0 Z t = 0 and 1 Z t = Z t, thus in some respects a Z t resembles the usual az t. The Integer Valued AR process (INAR) of order 1 INAR(1) is defined as X t = a X t 1 + ε t, where ε t are iid discrete random variable, independent of X t 1 and Thus We observe that X t = X t 1 j=1 B t,j + ε t. E (X t X t 1 ) = ax }{{ t 1 +E[ε } t ], mean of Bin(a,X t 1 ) which resembles the conditional expectation of the regular AR(1) which is E[X t X t 1 ] = ax t 1. Note the marginal distribution of X t will not binomial. The moments and covariances between X t have interesting properties, which are worth investigating. Objectives Properties of the INAR model. Estimation methods of the INAR model. 3
4 Summarize sampling properties of the INAR model. Simulations; are the sampling properties reliable. Extensions of the model. 4
5 Topic 2: Granger Causality and Vector AR processes Literature Graphical interaction models for multivariate time series: Metrika (2000) by Rainer Dahlhaus. Application to Neurosciences (in particular looking at particular wave bands): Here I know that Hernando Ombao also worked on this, I just cannot find his papers! A paper set in high dimensions; High dimension Consider the stationary multivariate time series X t = (X 1,t,..., X d,t ). The notion of partial correlation (defined in Section 4.3 of the class notes) is generalized to the multivariate set-up. Let X a,t and X b,t correspond to two elements in the vector X t. Let Y t, (a,b) = (X i,t ; i a or b) (after removing the elements a and b). The best linear predictor of X a,t and X b,t given {Y τ, (a,b) } τ= is X a,t = X b,t = d a (τ)y t τ, (a,b) τ= τ= d b (τ)y t τ, (a,b). The partial correlation between X a,t and X b,τ after removing this dependence is cov (a,b) (X a,t, X b,τ ) = cov (X a,t X a,t, X b,τ X ) b,τ. The above can be considered as a measure of dependence between X a,t and X b,τ after removing the information from the other components in the time series. If this is zero for all t and τ, then it means there is not dependence (if Gaussian) between the two time series after removing the other time series. However, calculating this can be difficult. However, Dahlhaus reformulates the problem in terms of spectral density matrices (recall for univariate time series these are like the 5
6 eigenvalues we discussed in class). We recall that for the univariate ARMA the spectral density is σ 2 1+θ exp(iω) 2 1 φ exp(iω) 2 = τ= c(τ) exp(iτω). This can be generalized to multivariate time series. Define the spectral matrix Σ(ω) = τ= cov(x 0, X τ ) exp(iτω) ω [0, 2π] The following result holds. Let Σ(ω) (a,b) denote the (a, b)th element of Σ(ω) 1 (a form of precision matrix), then if for all ω [0, 2ω] Σ(ω) (a,b) = 0, then cov (a,b) (X a,t, X b,τ ) = 0 for all t and τ. This result is analogous to the result for partial correlation given in Section Applications Suppose we are modelling X t with a vector AR(p) X t = p Φ j X t j + ε t, j=1 where {ε t } t are iid random vectors with variance σ 2 I d. The non-zero components in Φ j can be interpretated as understanding which time series directly impact other time series. Suppose Φ j,a,b, then it says that X b,t j has a direct impact on X a,t. Therefore, the dependence between X a,t and X b,t j remains even after removing the dependence on other variables. Thus Σ(ω) (a,b) will be non-zero for at least some frequency ω. Conversely, if for all ω Σ(ω) (a,b) = 0 then Φ j (a, b) = Φ j (b, a) = 0 for 1 j p. This means there is only a indirect dependence between the time series. This can be represented in terms of an undirected graph. If Σ(ω) (a,b) 0 and the VAR(p) is the correct model, then it means that for some j either Φ j (a, b) 0 or Φ j (b, a) 0. In other words, one variable has an actual influence on other (not indirectly, through the other variables). 1 To understand why it is true note that Σ(ω k ) is approximately the variance of the DFT of the vector time series J n (ω k ) = n 1/2 n t=1 X t exp(itω k). Under certain conditions J n (ω k ) is approximately (of order O(n 1 )) uncorrelated at different frequencies. Thus if Σ(ω k ) a,b it tells us that J a (ω k ) and J b (ω k ) (ath and bth element of J n (ω k ) are uncorrelated after the linear dependence on the other time series). This together with the near uncorrelatedness of J n (ω k ) at different frequencies gives the result. 6
7 Looking at regions of [0, π] (α, β waves etc discussed in Section 1.4.1) where Σ(ω) (a,b) = 0 is considered in the neuroscience paper described above. Directed graphs are also important. This when we are only interested in how a time series in the past may influence a time series in the future. For example, if X τ,b for τ t has a direct influence on X t,a this direct influence is modelled with a VAR X t = p Φ j X t j + ε t. j=1 But this time there is a direct influence if Φ j (a, b) 0 for some 1 j p (notice the is no parity here we do not look for Φ j (b, a) 0). This is an example of a directed graph. Objectives Understand the above through directed and undirected graphs. How does one test for partial correlation in a time series? Extensions to high dimensions? 7
8 Topic 3: Estimation of burst type signals with application to modelling ECG data Literature Statistical Signal Processing: Kundu and Nandi: Here Book We focus on the burst type model Y t = p A j exp (b j (1 cos(αt + c j ))) cos (θ j t + φ j ) + ε t j=1 Two plots of the burst model with different parameters are given in Figure 2 (no noise was added). signal Time signal Time Figure 2: Plots in the case p = 1 and (a) A = 0.05, b = 3.5, α = 0.132, θ = (b) A = 0.05, b = 3.5, α = 0.25, θ = Observe the only change in the models is the slightly change in the α parameter. In Figure 3 we plot the same burst signal but white and coloured noise added. The parameters of the model can be estimated by minimising the non-linear least squares 8
9 signal1n Time signal1nn Time Figure 3: Plots in the case p = 1 and A = 0.05, b = 3.5, α = (a) with white noise (b) with coloured noise. criterion; L(b, α, c, θ, φ) = ( 2 n p Y t A j exp (b j (1 cos(αt + c j ))) cos (θ j t + φ j )). t=1 j=1 Here are some suggestions in the case the errors are dependent. We assume that the errors follow an AR(1) process, ε t = φε t 1 + η t where {η t } are iid random variables and the model has order p = 1. In other words Y t = A exp (b (1 cos(αt + c))) cos (θt + φ) + ε t and ε t = φε t 1 + η t. To simplify notation, we define the function G(t, θ) = A exp (b (1 cos(αt + c))) cos (θt + φ), where θ = (A, b, α, c, φ), thus Y t = G(t, θ) + ε t Below we describe are two possible approaches for estimating the parameters. 9
10 Approach 1 Since η t = ε t φε t 1, then ε t = Y t G(t, θ) ε t 1 = Y t 1 G(t 1, θ). Therefore the white noise η t can be written as η t = ε t φε t 1 = (Y t G(t, θ)) φ (Y t 1 G(t 1, θ)). (1) Since η t are iid random variables we can simultaneously estimate φ and θ using the least squares criterion L(θ, φ) = n (Y t G(t, θ) φy t 1 + φg(t 1, θ)) 2 t=2 Then find the parameters θ, φ which minimise the above. Approach 2 This is a three stage approach, which may be easier to implement, especially when the order of the AR(p) process is higher than 1. The idea is that we estimate θ using regular least squares. Use this to estimate the residuals ε t, which is used to estimate the autoregressive parameters. Which we then use again to reestimate θ, in the hope one obtains and more efficient estimator (estimator with a smaller variance) the second time round. Step 1 Find θ which minimises L 1 (θ) = n (Y t G(t, θ)) 2. t=1 Let θ 1 = arg min L 1 (θ). Evaluate the residuals ε t = Y t G(t, θ 1 ) and make an ACF plot of the estimated residuals. If there appears to be correlation go on to the next step. 10
11 Step 2 Find the φ which minimises L 2 (φ) = n ( ε t φ ε t 1 ) 2. t=1 Let φ = arg min L 2 (φ). Step 3 We substitute φ into (1) to give L 3 (θ) = n t=2 n t=2 ( Y t G(t, θ) φy t 1 + φg(t 1, θ)) 2. Observe that the only unknown in the above criterion is θ. Now use θ where θ 2 = arg min L 3 (θ). Step 4 Our final estimator of θ is θ 3. Evaluate the residuals η t = ( Y t G(t, θ ) 3 ) φ ( Y t 1 G(t 1, θ ) 3 ) and make an ACF plot of these residuals. 11
12 Objectives: Investigate the parameter estimators using least squares and the methods described above for both white and correlated noise. Run many simulations (say a thousand) for the parameter estimators. Evaluate the average squared error of the estimating and the plott the histogram/qqplot of the centralized estimators. Do the plots look close to normal? What does the periodogram of the model look like. What information does it convey. How does peaks in the periodogram correspond to the different parameters? any information about the period in the ECG? Fit the model to real ECG data. 12
13 Topic 4: Multivariate GARCH Literature A simple review. Here; paper The review paper, Multivariate GARCH models by Silvennoinen and Terasvirta (2009). Let y t be the price of a stock at time t. Let r t = log y t log y t 1. The univariate GARCH(p, q) model models y t as r t = σ t Z t where {Z t } are iid random variables with E[Z t ] = 0 and var(z t ) = 1 and var(r t r t 1, r t 2,...) = σ 2 t = a 0 + p q a i rt i 2 + b j σt j. 2 i=1 j=1 But this is just how one stock is modelled. Suppose we observe the price of d stocks. The log returns r t = (r 1,t,..., r d,t ) can be modelled as r t = Σ t Z t where Z t are iid d-dimension random vector with variance/covariance matrix I d. But how to model the conditional variance matrix Σ t. The simplest model is Σ t = A 0 + p q A i r t i r t i + B j Σ t j i=1 j=1 where crucially all A 0, A i and B j are diagonal matrices. By expanding this out, you then this the is exactly the same as modelling each element of r t as a univarate GARCH(p, q) process and does not allow for dependencies between the different returns. Clearly in many situations this is a highly unrealistic modelling assumption. It does not allow for the influence one return on the other. 13
14 To model all the possible dependencies we need to remember that the conditonal variance Σ t is a non-negative definite matrix and not a vector. So we cannot simply generalize to the vector set-up as was done for the VAR(p) and VARMA(p, q) model. Since Σ t is a d d matrix, which is a variance/covariance matrix so is symmetric it has a maximum of d(d+1)/2 different entries. Thus we rewrite Σ t as a d(d + 1)/2 vector called the vech-transform vech(sigma t ). Thus the most general model is the VEC-GARCH(p, q) model vech(σ t ) = a 0 + p q A i vech(y t y ) + B t j vech(σ t j ). i=1 j=1 This is a very general and flexible model. However, there are two major considerations. The first is that vech(σ t ) should correspond to the entries of a non-negative definite matrix. This means finding conditons on the parameters which will ensure this is true (and then maximising over the appropriate parameters space to ensure that it is true). The second is the large number of parameters in this model. A 0, A i and B i are d(d + 1)/2 d(d + 1)/2 matrices. This means there are (p + q)(d(d + 1)/2) 2 unknown parameters plus those in A 0. Clearly, it is almost infeasible to estimate so many parameters when d is even a few dimensions. The corresponding likelihood is also extremely difficult to maximise. Therefore several simpler models have been proposed. Probably the post popular is the BEKK-GARCH model, which is a subset of VEC-GARCH. Objectives Review several different Multivariate GARCH models. and disadvantages. Discussing their advantages Code the estimators and make an extensive simulation studies. Validate the models by assessing the residuals (see Chapter 7). Fit the models to real data sets and compare the different predictors. 14
15 Topic 5: Categorical time series within a GLM framework Literature The book Weiss (2018), Chapter 7, gives a nice introduction. Here Weiss Book Parameter driven models: Time series analysis of non-gaussian Time series Models based on state space models from both a classical and Bayesian perspective (JRSSB). Durbin and Koopman (2000). Here Observation driven models: Here The review paper: Discrete time series: A GLM perspective covers both parameter driven and observation driven models. Here Here Suppose that Y i is a categorical or discrete random variable with probability distribution ( ) P (Y = y; θ) = exp yθ κ(θ) + c(y). The above distribution describes and a general class of distributions called the exponential family. Thus the log-probability distribution is log P (Y i = y; θ) = yθ κ(θ) + c(y). Famous examples, which belong to the above class are binary, binomial, multinomial (for 15
16 categorical data) and Poisson random variables; Binary: P (Y = 1; p) = p ( ) n Binomial: P (Y n = k; p) = p k (1 p) n k 0 k n k ( ) d n Multinomial: P (Y n = k 1,..., k d ; p 1,..., p d 1 ) = k 1, k 2,..., k k Poisson: P (Y n = k) = exp( λ)λk k! i=1 p i k k d = n Often observed with Y i are explanatory variables x i = (x i,1,..., x i,p ). The GLM-approach to model the dependence of x i on Y i through a linear transform. Let η x i = p j=1 η jx i,j. Often this is done in the following way; Binary: P (Y i = 1; x i ) = exp(η x i ) 1 + exp(η x i ) ( ) n Binomial: P (Y i = k; x i ) = p k (1 p) n k ; p = exp(η x i ) k 1 + exp(η x i ) Multinomial: P (Y i = k 1,..., k d ; x i ); p j = exp(η j x i) p s=1 exp(η s x i) Poisson: P (Y i = k; x i ) = exp( λ)λk ; λ = exp(η x k! i ). Let us return to the case that {Y t } is observed over time and Y t are either categorical or discrete valued random variables. Since {Y t } is observed over time it is highly likely that there is dependence. There are various methods for inducing the dependence. We summarize two of the approaches below. Parameter driven state-space models This can considered as a state-space model for non-gaussian data. We model the dependence through an unobserved latent variable. And by conditioning on the latent variables we obtain 16
17 a classical GLM. For example Binary: P (Y t = 1 α t ) exp(α t) 1 + exp(α t ) Multinomial: P (Y t = k 1,..., k d α t ); p j = exp(β j α t) p s=1 exp(β j α t) Poisson: P (Y t = k α t ) = exp( λ)λk ; log λ = α t. k! We model the dependence of Y t over time through the latent variable using, say, the classical autoregressive process α t = Aα t 1 + ε t where {ε t } t are iid vector random variables. Since α t is dependent over time, this induces the dependence in the time series {Y t }. Observation driven models The observation driven models use a similar idea, but avoid the need of introducing a latent random variable to model the time dependence. We give two such approaches below for numerical discrete time series. Model 1; Poisson: ( ) P (Y t = k Y t 1 ) = exp( λ)λk Yt 1 λ t 1 ; log λ t = β 0 + β 1 k! λ η t 1 for η > 0. Essentially, since Y t 1 is discrete, by defining the residual Y t 1 λ t 1 one is making λ η t 1 a transformation of the data to something closer to Gaussian. This approach is described in this paper. An alternative model is the ARMA type model Poisson: P (Y t = k Y t 1, Y t 2,...) = exp( λ)λk ; log λ t = β 0 + k! p φ j Y t j + j=1 q θ i λ t i. i=1 17
18 This model is described in this paper. Objectives Review the different models that are available, describing the similarities and differences. Describe and implement the different estimation methods. Look for methods which check for correlation in the residuals. 18
Chapter 3 - Temporal processes
STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect
More informationStatistics & Data Sciences: First Year Prelim Exam May 2018
Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationBooth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay. Midterm
Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay Midterm Chicago Booth Honor Code: I pledge my honor that I have not violated the Honor Code during
More informationTIME SERIES DATA ANALYSIS USING EVIEWS
TIME SERIES DATA ANALYSIS USING EVIEWS I Gusti Ngurah Agung Graduate School Of Management Faculty Of Economics University Of Indonesia Ph.D. in Biostatistics and MSc. in Mathematical Statistics from University
More information2. Multivariate ARMA
2. Multivariate ARMA JEM 140: Quantitative Multivariate Finance IES, Charles University, Prague Summer 2018 JEM 140 () 2. Multivariate ARMA Summer 2018 1 / 19 Multivariate AR I Let r t = (r 1t,..., r kt
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationNew Introduction to Multiple Time Series Analysis
Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2
More informationIntro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH
ntro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH JEM 140: Quantitative Multivariate Finance ES, Charles University, Prague Summer 2018 JEM 140 () 4. MGARCH Summer 2018
More informationIntroduction to ARMA and GARCH processes
Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,
More informationChapter 4: Models for Stationary Time Series
Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t
More informationGeneralized Linear Models and Exponential Families
Generalized Linear Models and Exponential Families David M. Blei COS424 Princeton University April 12, 2012 Generalized Linear Models x n y n β Linear regression and logistic regression are both linear
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationDiscrete time processes
Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More information2. An Introduction to Moving Average Models and ARMA Models
. An Introduction to Moving Average Models and ARMA Models.1 White Noise. The MA(1) model.3 The MA(q) model..4 Estimation and forecasting of MA models..5 ARMA(p,q) models. The Moving Average (MA) models
More informationECON 4160, Lecture 11 and 12
ECON 4160, 2016. Lecture 11 and 12 Co-integration Ragnar Nymoen Department of Economics 9 November 2017 1 / 43 Introduction I So far we have considered: Stationary VAR ( no unit roots ) Standard inference
More informationECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications
ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationStochastic Processes
Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationØkonomisk Kandidateksamen 2005(I) Econometrics 2 January 20, 2005
Økonomisk Kandidateksamen 2005(I) Econometrics 2 January 20, 2005 This is a four hours closed-book exam (uden hjælpemidler). Answer all questions! The questions 1 to 4 have equal weight. Within each question,
More informationLECTURE 11: EXPONENTIAL FAMILY AND GENERALIZED LINEAR MODELS
LECTURE : EXPONENTIAL FAMILY AND GENERALIZED LINEAR MODELS HANI GOODARZI AND SINA JAFARPOUR. EXPONENTIAL FAMILY. Exponential family comprises a set of flexible distribution ranging both continuous and
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationLecture 2: Univariate Time Series
Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:
More informationCovariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )
Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y
More informationMore Spectral Clustering and an Introduction to Conjugacy
CS8B/Stat4B: Advanced Topics in Learning & Decision Making More Spectral Clustering and an Introduction to Conjugacy Lecturer: Michael I. Jordan Scribe: Marco Barreno Monday, April 5, 004. Back to spectral
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationARIMA Modelling and Forecasting
ARIMA Modelling and Forecasting Economic time series often appear nonstationary, because of trends, seasonal patterns, cycles, etc. However, the differences may appear stationary. Δx t x t x t 1 (first
More informationStatistics 910, #15 1. Kalman Filter
Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations
More informationSGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection
SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector
More informationProf. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis
Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation
More informationDefine y t+h t as the forecast of y t+h based on I t known parameters. The forecast error is. Forecasting
Forecasting Let {y t } be a covariance stationary are ergodic process, eg an ARMA(p, q) process with Wold representation y t = X μ + ψ j ε t j, ε t ~WN(0,σ 2 ) j=0 = μ + ε t + ψ 1 ε t 1 + ψ 2 ε t 2 + Let
More informationSTAT 518 Intro Student Presentation
STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible
More informationMachine Learning. Lecture 3: Logistic Regression. Feng Li.
Machine Learning Lecture 3: Logistic Regression Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2016 Logistic Regression Classification
More informationMAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)
MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes. Only
More informationRegression. ECO 312 Fall 2013 Chris Sims. January 12, 2014
ECO 312 Fall 2013 Chris Sims Regression January 12, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License What
More informationFinancial Times Series. Lecture 12
Financial Times Series Lecture 12 Multivariate Volatility Models Here our aim is to generalize the previously presented univariate volatility models to their multivariate counterparts We assume that returns
More informationSimultaneity in the Multivariate Count Data Autoregressive Model
Simultaneity in the Multivariate Count Data Autoregressive Model Kurt Brännäs Department of Economics Umeå School of Business and Economics, Umeå University email: kurt.brannas@econ.umu.se Umeå Economic
More information5: MULTIVARATE STATIONARY PROCESSES
5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationHeteroskedasticity in Time Series
Heteroskedasticity in Time Series Figure: Time Series of Daily NYSE Returns. 206 / 285 Key Fact 1: Stock Returns are Approximately Serially Uncorrelated Figure: Correlogram of Daily Stock Market Returns.
More informationLatent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent
Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary
More informationVector Auto-Regressive Models
Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationGARCH processes probabilistic properties (Part 1)
GARCH processes probabilistic properties (Part 1) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/
More informationTime Series 2. Robert Almgren. Sept. 21, 2009
Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models
More informationVAR Models and Applications
VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Chapter 9 Multivariate time series 2 Transfer function
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More information9 Graphical modelling of dynamic relationships in multivariate time series
9 Graphical modelling of dynamic relationships in multivariate time series Michael Eichler Institut für Angewandte Mathematik Universität Heidelberg Germany SUMMARY The identification and analysis of interactions
More informationA time series is called strictly stationary if the joint distribution of every collection (Y t
5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a
More informationIntroduction to Probabilistic Graphical Models
Introduction to Probabilistic Graphical Models Sargur Srihari srihari@cedar.buffalo.edu 1 Topics 1. What are probabilistic graphical models (PGMs) 2. Use of PGMs Engineering and AI 3. Directionality in
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationMAT3379 (Winter 2016)
MAT3379 (Winter 2016) Assignment 4 - SOLUTIONS The following questions will be marked: 1a), 2, 4, 6, 7a Total number of points for Assignment 4: 20 Q1. (Theoretical Question, 2 points). Yule-Walker estimation
More informationMultivariate GARCH models.
Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains
More informationECON 4160, Spring term Lecture 12
ECON 4160, Spring term 2013. Lecture 12 Non-stationarity and co-integration 2/2 Ragnar Nymoen Department of Economics 13 Nov 2013 1 / 53 Introduction I So far we have considered: Stationary VAR, with deterministic
More informationVector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.
Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationUC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes
UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 6: Probability and Random Processes Problem Set 3 Spring 9 Self-Graded Scores Due: February 8, 9 Submit your self-graded scores
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationCS Lecture 19. Exponential Families & Expectation Propagation
CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces
More informationRegression of Time Series
Mahlerʼs Guide to Regression of Time Series CAS Exam S prepared by Howard C. Mahler, FCAS Copyright 2016 by Howard C. Mahler. Study Aid 2016F-S-9Supplement Howard Mahler hmahler@mac.com www.howardmahler.com/teaching
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R
More informationProblem Set 1 Solution Sketches Time Series Analysis Spring 2010
Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically
More informationIf we want to analyze experimental or simulated data we might encounter the following tasks:
Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationStochastic Processes: I. consider bowl of worms model for oscilloscope experiment:
Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing
More informationCh 6. Model Specification. Time Series Analysis
We start to build ARIMA(p,d,q) models. The subjects include: 1 how to determine p, d, q for a given series (Chapter 6); 2 how to estimate the parameters (φ s and θ s) of a specific ARIMA(p,d,q) model (Chapter
More informationNonlinear Time Series Modeling
Nonlinear Time Series Modeling Part II: Time Series Models in Finance Richard A. Davis Colorado State University (http://www.stat.colostate.edu/~rdavis/lectures) MaPhySto Workshop Copenhagen September
More informationDifference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m.
Difference equations Definitions: A difference equation takes the general form x t fx t 1,x t 2, defining the current value of a variable x as a function of previously generated values. A finite order
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationPart III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to
TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let
More informationVector Autoregression
Vector Autoregression Jamie Monogan University of Georgia February 27, 2018 Jamie Monogan (UGA) Vector Autoregression February 27, 2018 1 / 17 Objectives By the end of these meetings, participants should
More information7. MULTIVARATE STATIONARY PROCESSES
7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationReview Session: Econometrics - CLEFIN (20192)
Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =
More informationLecture 4 - Spectral Estimation
Lecture 4 - Spectral Estimation The Discrete Fourier Transform The Discrete Fourier Transform (DFT) is the equivalent of the continuous Fourier Transform for signals known only at N instants separated
More informationNonlinear time series
Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More informationComment on Article by Scutari
Bayesian Analysis (2013) 8, Number 3, pp. 543 548 Comment on Article by Scutari Hao Wang Scutari s paper studies properties of the distribution of graphs ppgq. This is an interesting angle because it differs
More informationCh. 14 Stationary ARMA Process
Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable
More informationNon-Gaussian Maximum Entropy Processes
Non-Gaussian Maximum Entropy Processes Georgi N. Boshnakov & Bisher Iqelan First version: 3 April 2007 Research Report No. 3, 2007, Probability and Statistics Group School of Mathematics, The University
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationSymmetric btw positive & negative prior returns. where c is referred to as risk premium, which is expected to be positive.
Advantages of GARCH model Simplicity Generates volatility clustering Heavy tails (high kurtosis) Weaknesses of GARCH model Symmetric btw positive & negative prior returns Restrictive Provides no explanation
More informationProblem Set 2: Box-Jenkins methodology
Problem Set : Box-Jenkins methodology 1) For an AR1) process we have: γ0) = σ ε 1 φ σ ε γ0) = 1 φ Hence, For a MA1) process, p lim R = φ γ0) = 1 + θ )σ ε σ ε 1 = γ0) 1 + θ Therefore, p lim R = 1 1 1 +
More informationTAMS39 Lecture 2 Multivariate normal distribution
TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationTime-Varying Vector Autoregressive Models with Structural Dynamic Factors
Time-Varying Vector Autoregressive Models with Structural Dynamic Factors Paolo Gorgi, Siem Jan Koopman, Julia Schaumburg http://sjkoopman.net Vrije Universiteit Amsterdam School of Business and Economics
More informationSTAT Financial Time Series
STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR
More informationTime Series Modeling of Financial Data. Prof. Daniel P. Palomar
Time Series Modeling of Financial Data Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19,
More informationGeneralised AR and MA Models and Applications
Chapter 3 Generalised AR and MA Models and Applications 3.1 Generalised Autoregressive Processes Consider an AR1) process given by 1 αb)x t = Z t ; α < 1. In this case, the acf is, ρ k = α k for k 0 and
More informationStep-Stress Models and Associated Inference
Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More information