You must continuously work on this project over the course of four weeks.

Size: px
Start display at page:

Download "You must continuously work on this project over the course of four weeks."

Transcription

1 The project Five project topics are described below. You should choose one the projects. Maximum of two people per project is allowed. If two people are working on a topic they are expected to do double the work as compared to one person on a project. You must continuously work on this project over the course of four weeks. Each week on Monday you should post on a wikipage (that I will create) the work you have done on it so far. Everyone in the class can view your work. The wiki link (you need to use VPN to access this page from outside the university) By the end of the semester you are expected to submit a final report on the topic and present a talk on the topic. The literature I give is only a starting point. You are expected to find more papers and or books related to the subject and use them in your report and talk. Provisionally the talks will be presented on the reading data Thursday 6th December (if there are no objections). 1

2 Topic 1: Integer Valued Autoregressive Processes INAR(p) Literature An introduction to Discrete valued time series, Christian Weiss. Here is the online book: (Weiss) You may have to be within the university domain to access it. Review paper by Eddie McKenzie (2000), Discrete Variate Time Series (Mckenzie: Here) Typically we assume that time series is continuous. But in many cases the time series is will be positive and integer valued. If the highest integer value is large this can often be approximated with as a continuous value time series. However, this may not always be hold. The plot in Figure 1 is an example from Weiss (2018). Figure 1: Real data. Several authors including E. McKenzie and C. Weiss have proposed methods for modelling such data. We describe a GLM type approach in Topic 5. But here we discuss an approach which appears to be a close generalisation of the classical autoregressive process. To do this, 2

3 we first define the thinning operator is known as the thinning operator and a Z is defined Z t a Z t = B t,j Bin(a, Z t ) j=1 where {B t,j } Bernoulli random variables with P (B t,j = 1) = a. Clearly 0 a 1 and 0 Z t = 0 and 1 Z t = Z t, thus in some respects a Z t resembles the usual az t. The Integer Valued AR process (INAR) of order 1 INAR(1) is defined as X t = a X t 1 + ε t, where ε t are iid discrete random variable, independent of X t 1 and Thus We observe that X t = X t 1 j=1 B t,j + ε t. E (X t X t 1 ) = ax }{{ t 1 +E[ε } t ], mean of Bin(a,X t 1 ) which resembles the conditional expectation of the regular AR(1) which is E[X t X t 1 ] = ax t 1. Note the marginal distribution of X t will not binomial. The moments and covariances between X t have interesting properties, which are worth investigating. Objectives Properties of the INAR model. Estimation methods of the INAR model. 3

4 Summarize sampling properties of the INAR model. Simulations; are the sampling properties reliable. Extensions of the model. 4

5 Topic 2: Granger Causality and Vector AR processes Literature Graphical interaction models for multivariate time series: Metrika (2000) by Rainer Dahlhaus. Application to Neurosciences (in particular looking at particular wave bands): Here I know that Hernando Ombao also worked on this, I just cannot find his papers! A paper set in high dimensions; High dimension Consider the stationary multivariate time series X t = (X 1,t,..., X d,t ). The notion of partial correlation (defined in Section 4.3 of the class notes) is generalized to the multivariate set-up. Let X a,t and X b,t correspond to two elements in the vector X t. Let Y t, (a,b) = (X i,t ; i a or b) (after removing the elements a and b). The best linear predictor of X a,t and X b,t given {Y τ, (a,b) } τ= is X a,t = X b,t = d a (τ)y t τ, (a,b) τ= τ= d b (τ)y t τ, (a,b). The partial correlation between X a,t and X b,τ after removing this dependence is cov (a,b) (X a,t, X b,τ ) = cov (X a,t X a,t, X b,τ X ) b,τ. The above can be considered as a measure of dependence between X a,t and X b,τ after removing the information from the other components in the time series. If this is zero for all t and τ, then it means there is not dependence (if Gaussian) between the two time series after removing the other time series. However, calculating this can be difficult. However, Dahlhaus reformulates the problem in terms of spectral density matrices (recall for univariate time series these are like the 5

6 eigenvalues we discussed in class). We recall that for the univariate ARMA the spectral density is σ 2 1+θ exp(iω) 2 1 φ exp(iω) 2 = τ= c(τ) exp(iτω). This can be generalized to multivariate time series. Define the spectral matrix Σ(ω) = τ= cov(x 0, X τ ) exp(iτω) ω [0, 2π] The following result holds. Let Σ(ω) (a,b) denote the (a, b)th element of Σ(ω) 1 (a form of precision matrix), then if for all ω [0, 2ω] Σ(ω) (a,b) = 0, then cov (a,b) (X a,t, X b,τ ) = 0 for all t and τ. This result is analogous to the result for partial correlation given in Section Applications Suppose we are modelling X t with a vector AR(p) X t = p Φ j X t j + ε t, j=1 where {ε t } t are iid random vectors with variance σ 2 I d. The non-zero components in Φ j can be interpretated as understanding which time series directly impact other time series. Suppose Φ j,a,b, then it says that X b,t j has a direct impact on X a,t. Therefore, the dependence between X a,t and X b,t j remains even after removing the dependence on other variables. Thus Σ(ω) (a,b) will be non-zero for at least some frequency ω. Conversely, if for all ω Σ(ω) (a,b) = 0 then Φ j (a, b) = Φ j (b, a) = 0 for 1 j p. This means there is only a indirect dependence between the time series. This can be represented in terms of an undirected graph. If Σ(ω) (a,b) 0 and the VAR(p) is the correct model, then it means that for some j either Φ j (a, b) 0 or Φ j (b, a) 0. In other words, one variable has an actual influence on other (not indirectly, through the other variables). 1 To understand why it is true note that Σ(ω k ) is approximately the variance of the DFT of the vector time series J n (ω k ) = n 1/2 n t=1 X t exp(itω k). Under certain conditions J n (ω k ) is approximately (of order O(n 1 )) uncorrelated at different frequencies. Thus if Σ(ω k ) a,b it tells us that J a (ω k ) and J b (ω k ) (ath and bth element of J n (ω k ) are uncorrelated after the linear dependence on the other time series). This together with the near uncorrelatedness of J n (ω k ) at different frequencies gives the result. 6

7 Looking at regions of [0, π] (α, β waves etc discussed in Section 1.4.1) where Σ(ω) (a,b) = 0 is considered in the neuroscience paper described above. Directed graphs are also important. This when we are only interested in how a time series in the past may influence a time series in the future. For example, if X τ,b for τ t has a direct influence on X t,a this direct influence is modelled with a VAR X t = p Φ j X t j + ε t. j=1 But this time there is a direct influence if Φ j (a, b) 0 for some 1 j p (notice the is no parity here we do not look for Φ j (b, a) 0). This is an example of a directed graph. Objectives Understand the above through directed and undirected graphs. How does one test for partial correlation in a time series? Extensions to high dimensions? 7

8 Topic 3: Estimation of burst type signals with application to modelling ECG data Literature Statistical Signal Processing: Kundu and Nandi: Here Book We focus on the burst type model Y t = p A j exp (b j (1 cos(αt + c j ))) cos (θ j t + φ j ) + ε t j=1 Two plots of the burst model with different parameters are given in Figure 2 (no noise was added). signal Time signal Time Figure 2: Plots in the case p = 1 and (a) A = 0.05, b = 3.5, α = 0.132, θ = (b) A = 0.05, b = 3.5, α = 0.25, θ = Observe the only change in the models is the slightly change in the α parameter. In Figure 3 we plot the same burst signal but white and coloured noise added. The parameters of the model can be estimated by minimising the non-linear least squares 8

9 signal1n Time signal1nn Time Figure 3: Plots in the case p = 1 and A = 0.05, b = 3.5, α = (a) with white noise (b) with coloured noise. criterion; L(b, α, c, θ, φ) = ( 2 n p Y t A j exp (b j (1 cos(αt + c j ))) cos (θ j t + φ j )). t=1 j=1 Here are some suggestions in the case the errors are dependent. We assume that the errors follow an AR(1) process, ε t = φε t 1 + η t where {η t } are iid random variables and the model has order p = 1. In other words Y t = A exp (b (1 cos(αt + c))) cos (θt + φ) + ε t and ε t = φε t 1 + η t. To simplify notation, we define the function G(t, θ) = A exp (b (1 cos(αt + c))) cos (θt + φ), where θ = (A, b, α, c, φ), thus Y t = G(t, θ) + ε t Below we describe are two possible approaches for estimating the parameters. 9

10 Approach 1 Since η t = ε t φε t 1, then ε t = Y t G(t, θ) ε t 1 = Y t 1 G(t 1, θ). Therefore the white noise η t can be written as η t = ε t φε t 1 = (Y t G(t, θ)) φ (Y t 1 G(t 1, θ)). (1) Since η t are iid random variables we can simultaneously estimate φ and θ using the least squares criterion L(θ, φ) = n (Y t G(t, θ) φy t 1 + φg(t 1, θ)) 2 t=2 Then find the parameters θ, φ which minimise the above. Approach 2 This is a three stage approach, which may be easier to implement, especially when the order of the AR(p) process is higher than 1. The idea is that we estimate θ using regular least squares. Use this to estimate the residuals ε t, which is used to estimate the autoregressive parameters. Which we then use again to reestimate θ, in the hope one obtains and more efficient estimator (estimator with a smaller variance) the second time round. Step 1 Find θ which minimises L 1 (θ) = n (Y t G(t, θ)) 2. t=1 Let θ 1 = arg min L 1 (θ). Evaluate the residuals ε t = Y t G(t, θ 1 ) and make an ACF plot of the estimated residuals. If there appears to be correlation go on to the next step. 10

11 Step 2 Find the φ which minimises L 2 (φ) = n ( ε t φ ε t 1 ) 2. t=1 Let φ = arg min L 2 (φ). Step 3 We substitute φ into (1) to give L 3 (θ) = n t=2 n t=2 ( Y t G(t, θ) φy t 1 + φg(t 1, θ)) 2. Observe that the only unknown in the above criterion is θ. Now use θ where θ 2 = arg min L 3 (θ). Step 4 Our final estimator of θ is θ 3. Evaluate the residuals η t = ( Y t G(t, θ ) 3 ) φ ( Y t 1 G(t 1, θ ) 3 ) and make an ACF plot of these residuals. 11

12 Objectives: Investigate the parameter estimators using least squares and the methods described above for both white and correlated noise. Run many simulations (say a thousand) for the parameter estimators. Evaluate the average squared error of the estimating and the plott the histogram/qqplot of the centralized estimators. Do the plots look close to normal? What does the periodogram of the model look like. What information does it convey. How does peaks in the periodogram correspond to the different parameters? any information about the period in the ECG? Fit the model to real ECG data. 12

13 Topic 4: Multivariate GARCH Literature A simple review. Here; paper The review paper, Multivariate GARCH models by Silvennoinen and Terasvirta (2009). Let y t be the price of a stock at time t. Let r t = log y t log y t 1. The univariate GARCH(p, q) model models y t as r t = σ t Z t where {Z t } are iid random variables with E[Z t ] = 0 and var(z t ) = 1 and var(r t r t 1, r t 2,...) = σ 2 t = a 0 + p q a i rt i 2 + b j σt j. 2 i=1 j=1 But this is just how one stock is modelled. Suppose we observe the price of d stocks. The log returns r t = (r 1,t,..., r d,t ) can be modelled as r t = Σ t Z t where Z t are iid d-dimension random vector with variance/covariance matrix I d. But how to model the conditional variance matrix Σ t. The simplest model is Σ t = A 0 + p q A i r t i r t i + B j Σ t j i=1 j=1 where crucially all A 0, A i and B j are diagonal matrices. By expanding this out, you then this the is exactly the same as modelling each element of r t as a univarate GARCH(p, q) process and does not allow for dependencies between the different returns. Clearly in many situations this is a highly unrealistic modelling assumption. It does not allow for the influence one return on the other. 13

14 To model all the possible dependencies we need to remember that the conditonal variance Σ t is a non-negative definite matrix and not a vector. So we cannot simply generalize to the vector set-up as was done for the VAR(p) and VARMA(p, q) model. Since Σ t is a d d matrix, which is a variance/covariance matrix so is symmetric it has a maximum of d(d+1)/2 different entries. Thus we rewrite Σ t as a d(d + 1)/2 vector called the vech-transform vech(sigma t ). Thus the most general model is the VEC-GARCH(p, q) model vech(σ t ) = a 0 + p q A i vech(y t y ) + B t j vech(σ t j ). i=1 j=1 This is a very general and flexible model. However, there are two major considerations. The first is that vech(σ t ) should correspond to the entries of a non-negative definite matrix. This means finding conditons on the parameters which will ensure this is true (and then maximising over the appropriate parameters space to ensure that it is true). The second is the large number of parameters in this model. A 0, A i and B i are d(d + 1)/2 d(d + 1)/2 matrices. This means there are (p + q)(d(d + 1)/2) 2 unknown parameters plus those in A 0. Clearly, it is almost infeasible to estimate so many parameters when d is even a few dimensions. The corresponding likelihood is also extremely difficult to maximise. Therefore several simpler models have been proposed. Probably the post popular is the BEKK-GARCH model, which is a subset of VEC-GARCH. Objectives Review several different Multivariate GARCH models. and disadvantages. Discussing their advantages Code the estimators and make an extensive simulation studies. Validate the models by assessing the residuals (see Chapter 7). Fit the models to real data sets and compare the different predictors. 14

15 Topic 5: Categorical time series within a GLM framework Literature The book Weiss (2018), Chapter 7, gives a nice introduction. Here Weiss Book Parameter driven models: Time series analysis of non-gaussian Time series Models based on state space models from both a classical and Bayesian perspective (JRSSB). Durbin and Koopman (2000). Here Observation driven models: Here The review paper: Discrete time series: A GLM perspective covers both parameter driven and observation driven models. Here Here Suppose that Y i is a categorical or discrete random variable with probability distribution ( ) P (Y = y; θ) = exp yθ κ(θ) + c(y). The above distribution describes and a general class of distributions called the exponential family. Thus the log-probability distribution is log P (Y i = y; θ) = yθ κ(θ) + c(y). Famous examples, which belong to the above class are binary, binomial, multinomial (for 15

16 categorical data) and Poisson random variables; Binary: P (Y = 1; p) = p ( ) n Binomial: P (Y n = k; p) = p k (1 p) n k 0 k n k ( ) d n Multinomial: P (Y n = k 1,..., k d ; p 1,..., p d 1 ) = k 1, k 2,..., k k Poisson: P (Y n = k) = exp( λ)λk k! i=1 p i k k d = n Often observed with Y i are explanatory variables x i = (x i,1,..., x i,p ). The GLM-approach to model the dependence of x i on Y i through a linear transform. Let η x i = p j=1 η jx i,j. Often this is done in the following way; Binary: P (Y i = 1; x i ) = exp(η x i ) 1 + exp(η x i ) ( ) n Binomial: P (Y i = k; x i ) = p k (1 p) n k ; p = exp(η x i ) k 1 + exp(η x i ) Multinomial: P (Y i = k 1,..., k d ; x i ); p j = exp(η j x i) p s=1 exp(η s x i) Poisson: P (Y i = k; x i ) = exp( λ)λk ; λ = exp(η x k! i ). Let us return to the case that {Y t } is observed over time and Y t are either categorical or discrete valued random variables. Since {Y t } is observed over time it is highly likely that there is dependence. There are various methods for inducing the dependence. We summarize two of the approaches below. Parameter driven state-space models This can considered as a state-space model for non-gaussian data. We model the dependence through an unobserved latent variable. And by conditioning on the latent variables we obtain 16

17 a classical GLM. For example Binary: P (Y t = 1 α t ) exp(α t) 1 + exp(α t ) Multinomial: P (Y t = k 1,..., k d α t ); p j = exp(β j α t) p s=1 exp(β j α t) Poisson: P (Y t = k α t ) = exp( λ)λk ; log λ = α t. k! We model the dependence of Y t over time through the latent variable using, say, the classical autoregressive process α t = Aα t 1 + ε t where {ε t } t are iid vector random variables. Since α t is dependent over time, this induces the dependence in the time series {Y t }. Observation driven models The observation driven models use a similar idea, but avoid the need of introducing a latent random variable to model the time dependence. We give two such approaches below for numerical discrete time series. Model 1; Poisson: ( ) P (Y t = k Y t 1 ) = exp( λ)λk Yt 1 λ t 1 ; log λ t = β 0 + β 1 k! λ η t 1 for η > 0. Essentially, since Y t 1 is discrete, by defining the residual Y t 1 λ t 1 one is making λ η t 1 a transformation of the data to something closer to Gaussian. This approach is described in this paper. An alternative model is the ARMA type model Poisson: P (Y t = k Y t 1, Y t 2,...) = exp( λ)λk ; log λ t = β 0 + k! p φ j Y t j + j=1 q θ i λ t i. i=1 17

18 This model is described in this paper. Objectives Review the different models that are available, describing the similarities and differences. Describe and implement the different estimation methods. Look for methods which check for correlation in the residuals. 18

Chapter 3 - Temporal processes

Chapter 3 - Temporal processes STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

X t = a t + r t, (7.1)

X t = a t + r t, (7.1) Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical

More information

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay. Midterm

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay. Midterm Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay Midterm Chicago Booth Honor Code: I pledge my honor that I have not violated the Honor Code during

More information

TIME SERIES DATA ANALYSIS USING EVIEWS

TIME SERIES DATA ANALYSIS USING EVIEWS TIME SERIES DATA ANALYSIS USING EVIEWS I Gusti Ngurah Agung Graduate School Of Management Faculty Of Economics University Of Indonesia Ph.D. in Biostatistics and MSc. in Mathematical Statistics from University

More information

2. Multivariate ARMA

2. Multivariate ARMA 2. Multivariate ARMA JEM 140: Quantitative Multivariate Finance IES, Charles University, Prague Summer 2018 JEM 140 () 2. Multivariate ARMA Summer 2018 1 / 19 Multivariate AR I Let r t = (r 1t,..., r kt

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

New Introduction to Multiple Time Series Analysis

New Introduction to Multiple Time Series Analysis Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2

More information

Intro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH

Intro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH ntro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH JEM 140: Quantitative Multivariate Finance ES, Charles University, Prague Summer 2018 JEM 140 () 4. MGARCH Summer 2018

More information

Introduction to ARMA and GARCH processes

Introduction to ARMA and GARCH processes Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,

More information

Chapter 4: Models for Stationary Time Series

Chapter 4: Models for Stationary Time Series Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t

More information

Generalized Linear Models and Exponential Families

Generalized Linear Models and Exponential Families Generalized Linear Models and Exponential Families David M. Blei COS424 Princeton University April 12, 2012 Generalized Linear Models x n y n β Linear regression and logistic regression are both linear

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

2. An Introduction to Moving Average Models and ARMA Models

2. An Introduction to Moving Average Models and ARMA Models . An Introduction to Moving Average Models and ARMA Models.1 White Noise. The MA(1) model.3 The MA(q) model..4 Estimation and forecasting of MA models..5 ARMA(p,q) models. The Moving Average (MA) models

More information

ECON 4160, Lecture 11 and 12

ECON 4160, Lecture 11 and 12 ECON 4160, 2016. Lecture 11 and 12 Co-integration Ragnar Nymoen Department of Economics 9 November 2017 1 / 43 Introduction I So far we have considered: Stationary VAR ( no unit roots ) Standard inference

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Økonomisk Kandidateksamen 2005(I) Econometrics 2 January 20, 2005

Økonomisk Kandidateksamen 2005(I) Econometrics 2 January 20, 2005 Økonomisk Kandidateksamen 2005(I) Econometrics 2 January 20, 2005 This is a four hours closed-book exam (uden hjælpemidler). Answer all questions! The questions 1 to 4 have equal weight. Within each question,

More information

LECTURE 11: EXPONENTIAL FAMILY AND GENERALIZED LINEAR MODELS

LECTURE 11: EXPONENTIAL FAMILY AND GENERALIZED LINEAR MODELS LECTURE : EXPONENTIAL FAMILY AND GENERALIZED LINEAR MODELS HANI GOODARZI AND SINA JAFARPOUR. EXPONENTIAL FAMILY. Exponential family comprises a set of flexible distribution ranging both continuous and

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 ) Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y

More information

More Spectral Clustering and an Introduction to Conjugacy

More Spectral Clustering and an Introduction to Conjugacy CS8B/Stat4B: Advanced Topics in Learning & Decision Making More Spectral Clustering and an Introduction to Conjugacy Lecturer: Michael I. Jordan Scribe: Marco Barreno Monday, April 5, 004. Back to spectral

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

ARIMA Modelling and Forecasting

ARIMA Modelling and Forecasting ARIMA Modelling and Forecasting Economic time series often appear nonstationary, because of trends, seasonal patterns, cycles, etc. However, the differences may appear stationary. Δx t x t x t 1 (first

More information

Statistics 910, #15 1. Kalman Filter

Statistics 910, #15 1. Kalman Filter Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations

More information

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

Define y t+h t as the forecast of y t+h based on I t known parameters. The forecast error is. Forecasting

Define y t+h t as the forecast of y t+h based on I t known parameters. The forecast error is. Forecasting Forecasting Let {y t } be a covariance stationary are ergodic process, eg an ARMA(p, q) process with Wold representation y t = X μ + ψ j ε t j, ε t ~WN(0,σ 2 ) j=0 = μ + ε t + ψ 1 ε t 1 + ψ 2 ε t 2 + Let

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Machine Learning. Lecture 3: Logistic Regression. Feng Li.

Machine Learning. Lecture 3: Logistic Regression. Feng Li. Machine Learning Lecture 3: Logistic Regression Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2016 Logistic Regression Classification

More information

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE) MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes. Only

More information

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014 ECO 312 Fall 2013 Chris Sims Regression January 12, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License What

More information

Financial Times Series. Lecture 12

Financial Times Series. Lecture 12 Financial Times Series Lecture 12 Multivariate Volatility Models Here our aim is to generalize the previously presented univariate volatility models to their multivariate counterparts We assume that returns

More information

Simultaneity in the Multivariate Count Data Autoregressive Model

Simultaneity in the Multivariate Count Data Autoregressive Model Simultaneity in the Multivariate Count Data Autoregressive Model Kurt Brännäs Department of Economics Umeå School of Business and Economics, Umeå University email: kurt.brannas@econ.umu.se Umeå Economic

More information

5: MULTIVARATE STATIONARY PROCESSES

5: MULTIVARATE STATIONARY PROCESSES 5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

Heteroskedasticity in Time Series

Heteroskedasticity in Time Series Heteroskedasticity in Time Series Figure: Time Series of Daily NYSE Returns. 206 / 285 Key Fact 1: Stock Returns are Approximately Serially Uncorrelated Figure: Correlogram of Daily Stock Market Returns.

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

GARCH processes probabilistic properties (Part 1)

GARCH processes probabilistic properties (Part 1) GARCH processes probabilistic properties (Part 1) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/

More information

Time Series 2. Robert Almgren. Sept. 21, 2009

Time Series 2. Robert Almgren. Sept. 21, 2009 Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Chapter 9 Multivariate time series 2 Transfer function

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

9 Graphical modelling of dynamic relationships in multivariate time series

9 Graphical modelling of dynamic relationships in multivariate time series 9 Graphical modelling of dynamic relationships in multivariate time series Michael Eichler Institut für Angewandte Mathematik Universität Heidelberg Germany SUMMARY The identification and analysis of interactions

More information

A time series is called strictly stationary if the joint distribution of every collection (Y t

A time series is called strictly stationary if the joint distribution of every collection (Y t 5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Sargur Srihari srihari@cedar.buffalo.edu 1 Topics 1. What are probabilistic graphical models (PGMs) 2. Use of PGMs Engineering and AI 3. Directionality in

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

MAT3379 (Winter 2016)

MAT3379 (Winter 2016) MAT3379 (Winter 2016) Assignment 4 - SOLUTIONS The following questions will be marked: 1a), 2, 4, 6, 7a Total number of points for Assignment 4: 20 Q1. (Theoretical Question, 2 points). Yule-Walker estimation

More information

Multivariate GARCH models.

Multivariate GARCH models. Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains

More information

ECON 4160, Spring term Lecture 12

ECON 4160, Spring term Lecture 12 ECON 4160, Spring term 2013. Lecture 12 Non-stationarity and co-integration 2/2 Ragnar Nymoen Department of Economics 13 Nov 2013 1 / 53 Introduction I So far we have considered: Stationary VAR, with deterministic

More information

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I. Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 6: Probability and Random Processes Problem Set 3 Spring 9 Self-Graded Scores Due: February 8, 9 Submit your self-graded scores

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

CS Lecture 19. Exponential Families & Expectation Propagation

CS Lecture 19. Exponential Families & Expectation Propagation CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces

More information

Regression of Time Series

Regression of Time Series Mahlerʼs Guide to Regression of Time Series CAS Exam S prepared by Howard C. Mahler, FCAS Copyright 2016 by Howard C. Mahler. Study Aid 2016F-S-9Supplement Howard Mahler hmahler@mac.com www.howardmahler.com/teaching

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment: Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing

More information

Ch 6. Model Specification. Time Series Analysis

Ch 6. Model Specification. Time Series Analysis We start to build ARIMA(p,d,q) models. The subjects include: 1 how to determine p, d, q for a given series (Chapter 6); 2 how to estimate the parameters (φ s and θ s) of a specific ARIMA(p,d,q) model (Chapter

More information

Nonlinear Time Series Modeling

Nonlinear Time Series Modeling Nonlinear Time Series Modeling Part II: Time Series Models in Finance Richard A. Davis Colorado State University (http://www.stat.colostate.edu/~rdavis/lectures) MaPhySto Workshop Copenhagen September

More information

Difference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m.

Difference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m. Difference equations Definitions: A difference equation takes the general form x t fx t 1,x t 2, defining the current value of a variable x as a function of previously generated values. A finite order

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be  ed to TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let

More information

Vector Autoregression

Vector Autoregression Vector Autoregression Jamie Monogan University of Georgia February 27, 2018 Jamie Monogan (UGA) Vector Autoregression February 27, 2018 1 / 17 Objectives By the end of these meetings, participants should

More information

7. MULTIVARATE STATIONARY PROCESSES

7. MULTIVARATE STATIONARY PROCESSES 7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

Review Session: Econometrics - CLEFIN (20192)

Review Session: Econometrics - CLEFIN (20192) Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =

More information

Lecture 4 - Spectral Estimation

Lecture 4 - Spectral Estimation Lecture 4 - Spectral Estimation The Discrete Fourier Transform The Discrete Fourier Transform (DFT) is the equivalent of the continuous Fourier Transform for signals known only at N instants separated

More information

Nonlinear time series

Nonlinear time series Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

Comment on Article by Scutari

Comment on Article by Scutari Bayesian Analysis (2013) 8, Number 3, pp. 543 548 Comment on Article by Scutari Hao Wang Scutari s paper studies properties of the distribution of graphs ppgq. This is an interesting angle because it differs

More information

Ch. 14 Stationary ARMA Process

Ch. 14 Stationary ARMA Process Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable

More information

Non-Gaussian Maximum Entropy Processes

Non-Gaussian Maximum Entropy Processes Non-Gaussian Maximum Entropy Processes Georgi N. Boshnakov & Bisher Iqelan First version: 3 April 2007 Research Report No. 3, 2007, Probability and Statistics Group School of Mathematics, The University

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Symmetric btw positive & negative prior returns. where c is referred to as risk premium, which is expected to be positive.

Symmetric btw positive & negative prior returns. where c is referred to as risk premium, which is expected to be positive. Advantages of GARCH model Simplicity Generates volatility clustering Heavy tails (high kurtosis) Weaknesses of GARCH model Symmetric btw positive & negative prior returns Restrictive Provides no explanation

More information

Problem Set 2: Box-Jenkins methodology

Problem Set 2: Box-Jenkins methodology Problem Set : Box-Jenkins methodology 1) For an AR1) process we have: γ0) = σ ε 1 φ σ ε γ0) = 1 φ Hence, For a MA1) process, p lim R = φ γ0) = 1 + θ )σ ε σ ε 1 = γ0) 1 + θ Therefore, p lim R = 1 1 1 +

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

Time-Varying Vector Autoregressive Models with Structural Dynamic Factors

Time-Varying Vector Autoregressive Models with Structural Dynamic Factors Time-Varying Vector Autoregressive Models with Structural Dynamic Factors Paolo Gorgi, Siem Jan Koopman, Julia Schaumburg http://sjkoopman.net Vrije Universiteit Amsterdam School of Business and Economics

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

Time Series Modeling of Financial Data. Prof. Daniel P. Palomar

Time Series Modeling of Financial Data. Prof. Daniel P. Palomar Time Series Modeling of Financial Data Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19,

More information

Generalised AR and MA Models and Applications

Generalised AR and MA Models and Applications Chapter 3 Generalised AR and MA Models and Applications 3.1 Generalised Autoregressive Processes Consider an AR1) process given by 1 αb)x t = Z t ; α < 1. In this case, the acf is, ρ k = α k for k 0 and

More information

Step-Stress Models and Associated Inference

Step-Stress Models and Associated Inference Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information