ECONOMETRIC METHODS II TA session 1 MATLAB Intro: Simulation of VAR(p) processes
|
|
- Sophie Rose
- 6 years ago
- Views:
Transcription
1 ECONOMETRIC METHODS II TA session 1 MATLAB Intro: Simulation of VAR(p) processes Fernando Pérez Forero April 19th, Introduction In this rst session we will cover the simulation of Vector Autoregressive (VAR) processes of order p. Besides, we will cover how to compute Impulse Response Functions (IRF) and the Forecast Error Variance Decomposition (FEVD). It is assumed that students are familiar with general concepts of Time Series Econometrics and Matrix Algebra, i.e. we will not put many de nitions explicitly, since we assume that they were learnt by heart in the past. This session will be particularly useful for those students who are not familiar with MATLAB 1. It is important to emphasize that when we produce codes, they need to be as general as possible, so that it is easy to consider di erent parametrizations in the future. However there is no free lunch, i.e. this comes at the cost of investing a large amount of time in debugging the code. On the other hand, MATLAB is easy to use relative to other languages such as C++ or Phyton. In particular, since we will not perform object-oriented programming, it will not be necessary to declare variables before using them. Most of the material covered in this session can be found in Lütkepohl (2005) (L) and Hamilton (1994) (H). In order to provide a better understanding of each step, I will indicate the number of equation of each reference as follows: (Book, equation). PhD student. Department of Economics, Universitat Pompeu Fabra, Ramon Trias Fargas 25 27, Barcelona, Spain ( fernandojose.perez@upf.edu) 1 A very nice MATLAB primer by Winistorfer and Canova (2008) can be also found in 1
2 2 Preliminaries 2.1 VAR(p) process Consider a VAR(p) process (L,2.1.1), y t = v + A 1 y t A p y t p + u t ; t = 0; 1; 2; : : : ; (2.1) where y t ; u t and v are (K 1) vectors and A i are (K K) matrices for each i = 1; : : : ; p. In addition, the error term u t is a white noise random vector such that E (u t ) = 0, E (u t u 0 t) = u and E (u t u 0 s) = 0 for s 6= t, where u is a (K K) positive de nite matrix. It is useful to re-write (2:1) in its companion form (L,2.1.8), Y t = v + AY t 1 + U t (2.2) where Y t = y 0 t; y 0 t 1; : : : ; y 0 t p+1 0, v = [v 0 ; 0 0 ; : : : ; 0 0 ] 0 and U t = [u 0 t; 0 0 ; : : : ; 0 0 ] 0 are (Kp 1) vectors and 2 A = 6 4 A 1 A 2 A p 1 A p I K I K I K 0 is a (Kp Kp) matrix. The model (2:2) is said to be stable i max (j A j) < 1, where A denotes the vector of eigenvalues of A. The stability condition implies that the model can be re-written as a Moving-Average (MA) representation (L,2.1.13): X y t = JY t = J + J A i U t i (2.3) h where =E (Y t ) = (I Kp A) 1 v is the unconditional expectation and J = I K 0 0 is a (K Kp) selection matrix. Notice also that U t = J 0 JU t and u t = JU t, thus i=0 i y t = JY t = J+ 1X i=0 JA i J 0 JU t i where i = JA i J 0 denotes the i th matrix lag on the MA representation (L,2.1.17) y t = JY t = J+ 1X i=0 JA i J 0 u t i (2.4) Fernando Pérez Forero 2
3 2.2 Impulse Response Function (IRF) In multivariate systems one could be interested in explore the dynamic propagation of innovations across variables. For that purpose, there exists the impulse response function. Basically, one could study the e ect of a one-time innovation in a particular variable n on another variable m over time. The duration of that typoe of innovation will dep exclusively on the structure of the system. In this section we compute the Impulse Response function for the model (2:1). We proceed as follows: If the matrix u is positive de nite, it can be decomposed such that u = P P 0, where P is a lower triangular nonsingular matrix with positive diagonal elements. Moreover, we can re-write (2:4) as (L,2.3.33): y t = + 1X i! t i (2.5) i=0 where i = JA i J 0 P is a (K K) matrix and! t = P 1 u t is a white noise process with covariance matrix equal to the identity matrix, i.e.! = I K. Thus, the impulse response function t = i ; i = 0; 1; 2; : : : (2.6) where i = i jk is such that i jk denotes the response of variable j to a shock in variable k after i periods. 2.3 Forecast Error Variance Decomposition (FEVD) We have shown how to compute the responses of variables after innovations in the system (2:5). It is also possible to perform a forecasting exercise using the latter system. Denote the forecast error h periods ahead as (L,2.3.34): Xh 1 e t+h = y t+h y t (h) = i! t+h i (2.7) Denote the contribution of innovation on variable k to the forecast error h periods ahead of variable j (L,2.3.37): where! jk;h = MSE [y j;t (h)] = i=0 P h 1 i=0 e 0 j i e k 2 MSE [y j;t (h)] Xh 1 i=0 KX k=1 (2.8) i jk 2 (2.9) Fernando Pérez Forero 3
4 3 Introducing the model in MATLAB In this section we simulate data using the system (2:1). We set matrices A i, i = 1; : : : ; p such that the VAR(p) is stationary. 3.1 Data simulation We set K = 4, p = 2 and the number of observations T = 100. p=4; % Lag-order T=100; % Observations K=4; % Variables; Then, we set the matrix u such that it is positive de nite: scale=2*rand(k,1)+1e-30; Sigma_U=0.5*rand(K,K)+diag(scale); Sigma_U=Sigma_U *Sigma_U; if all(eig((sigma_u+sigma_u )/2)) >= 0 disp( Sigma_U is positive definite ) else error( Sigma_U must be positive definite ) And we set lag matrices A i (K K) for each i = 1; : : : ; p in a 3-dimensional vector: Alag=zeros(K,K,p); lambda=2.5; for k=1:p Alag(:,:,k)=(lambda^(-k))*rand(K,K)-((2*lambda)^(-k))*ones(K,K); With all these ingredients, we procced to simulate T = 100 observations of the process (2:1) assuming u t N (0; u ) and p initial conditions Y p+1 ; : : : ; Y 1 ; Y 0 that come from the same distribution. Besides, the intercept v is a vector of ones: Ylag=zeros(K,p); for k=1:p Ylag(:,k)=mvnrnd(zeros(K,1),Sigma_U) ; Y=Ylag; v=ones(k,1); for i=1:t Fernando Pérez Forero 4
5 temp=v+mvnrnd(zeros(k,1),sigma_u) ; for k=1:p temp=temp+alag(:,:,k)*ylag(:,k); Y=[Y,temp]; Ylag(:,p)=temp; for k=1:p-1 Ylag(:,k)=Ylag(:,k+1); clear Ylag temp A_coeff=v; for k=1:p A_coeff=[A_coeff,Alag(:,:,k)]; Y=Y(:,p+1:); We then plot the simulated data in Figure 3.1: FS=15; LW=2; gr_size2=ceil(k/3); figure(1) set(0, DefaultAxesColorOrder,[0 0 1],... DefaultAxesLineStyleOrder, -j-j- ) set(gcf, Color,[1 1 1]) set(gcf, defaultaxesfontsize,fs) for k=1:k subplot(gr_size2,2,k) plot(y(k,:), b, Linewidth,LW) title(sprintf( Y_%d,k)) Fernando Pérez Forero 5
6 Figure 3.1: Simulated Data Next step is to generate the companion form (2:2) and check for stationarity given the parameter values: F1=[A_coeff(:,2:)]; Q1=Sigma_U; if p>1 % Matrix F - [H, ] F2=cell(1,p-1); [F2{:}]=deal(sparse(eye(K))); F2=blkdiag(F2{:}); F2=full(F2); F3=zeros(K*(p-1),K); F=[F1;F2,F3]; % Var_cov Q - [H, ] Q2=cell(1,p-1); [Q2{:}]=deal(sparse(zeros(K,K))); Q2=blkdiag(Q2{:}); Fernando Pérez Forero 6
7 Q2=full(Q2); Q=blkdiag(Q1,Q2); clear F1 F2 F3 else F=F1; Q=Q1; clear F1 Q1 if max(abs(eig(f)))<1 % [H, ] disp( VAR is stationary ) else error( VAR is not stationary! ) The companion form vector Y t has a covariance matrix Y. Using the fact that the model is stationary, we take variances on (2:2), (H, ) Y = A Y A 0 + Q (3.1) where Q =E (U t U 0 t). (3:1) is a discrete Lyapunov equation that can be solved in two ways: i) using a while loop and iterating equation this equation or ii) using a built-in MATLAB function dlyap.m. We use both methods and compare our results: % Companion form s Variance-Covariance matrix % [H, ] Cap_Sigma0=eye(size(F)); Cap_Sigma1=F*Cap_Sigma0*F +Q; diff=max(max(abs(cap_sigma1-cap_sigma0))); diff1=[]; epsilon=1e-10; iter=0; figure(1) set(gcf, Color,[1 1 1]) tic disp( Iterating covariance matrix... ) while diff>epsilon iter=iter+1; Cap_Sigma0=Cap_Sigma1; Cap_Sigma1=F*Cap_Sigma0*F +Q; Fernando Pérez Forero 7
8 diff=max(max(abs(cap_sigma1-cap_sigma0))); diff1=[diff1,diff]; figure(2) plot(diff1) title( Evolution of convergence, FontSize,14) sprintf( Convergence achieved after %d iterations,iter) % Alternatively, solve the discrete Lyapunov equation Cap_Sigma1l=dlyap(F,Q); % Compare results diff2=max(max(abs(cap_sigma1l-cap_sigma1))); toc Figure 3.2 shows that convergence is roughly achieved after iterating this equation 7 times. However, since we impose a convergence criteria of " = , the loop converges after 20 iterations. Moreover, in this application di 2 is less than , which shows that the loop produces a good approximation. We now proceed to the next section for a more interesting analysis. Figure 3.2: Evolution of convergence of Y Fernando Pérez Forero 8
9 3.2 Impulse Response Function (IRF) Once we have generated the companion form components in (2:2), computing Impulse Responses is extremely simple. First, we set a nite horizon h and we generate matrices P and J. h=24; % Horizon P=chol(Sigma_U); % Cholesky decomposition P=P ; % Following the notation of Lutkepohl - Section 3.7 : capj=zeros(k,k*p); capj(1:k,1:k)=eye(k); U=P*E Then we construct matrices i(kk) according to equation (2:6) and store all of them in a 3-dimensional object: Iresp=zeros(K,K,h); Iresp(:,:,1)=P; for j=1:h-1 temp=(f^j); Iresp(:,:,j+1)=capJ*temp*capJ *P; clear temp Finally, we plot the Impulse Responses in Figure 3.3: FS=15; LW=2; figure(3) set(0, DefaultAxesColorOrder,[0 0 1],... DefaultAxesLineStyleOrder, -j-j- ) set(gcf, Color,[1 1 1]) set(gcf, defaultaxesfontsize,fs-5) for i=1:k for j=1:k subplot(k,k,(j-1)*k+i) plot(squeeze(iresp(i,j,:)), Linewidth,LW) hold on plot(zeros(1,h), k ) title(sprintf( Y_{%d} to U_{%d},i,j)) xlim([1 h]) set(gca, XTick,0:ceil(h/4):h) Fernando Pérez Forero 9
10 Figure 3.3: Impulse responses 3.3 Forecast Error Variance Decomposition (FEVD) Once we have computed the Impulse Responses, we can now proceed to compute the contribution of each variable to the Forecast Error variance j = 1; : : : ; h periods ahead. We rst compute the numerator and denominator of equation (2:8) separately, i.e. the absolute contribution and the Mean-Squared-Error (MSE). MSE=zeros(K,h); CONTR=zeros(K,K,h); for j=1:h % Compute MSE temp2=eye(k); for i=1:k if j==1 Fernando Pérez Forero 10
11 MSE(i,j)=temp2(:,i) *Iresp(:,:,j)*Sigma_U*Iresp(:,:,j) *temp2(:,i); for k=1:k CONTR(i,k,j)=(temp2(:,i) *Iresp(:,:,j)*P*temp2(:,k))^2; else MSE(i,j)=MSE(i,j-1)+temp2(:,i) *Iresp(:,:,j)*Sigma_U*Iresp(:,:,j) *temp2(:,i); for k=1:k CONTR(i,k,j)=CONTR(i,k,j-1)+(temp2(:,i) *Iresp(:,:,j)*P*temp2(:,k))^2; % (L,2.3.36) Then we compute the fraction (2:8) and plot the results in Figure 3.4: VD=zeros(K,h,K); for k=1:k VD(:,:,k)=squeeze(CONTR(:,k,:))./MSE; % (L,2.3.37) clear temp2 CONTR MSE figure(4) set(0, DefaultAxesColorOrder,[0 0 1],... DefaultAxesLineStyleOrder, -j-j- ) set(gcf, Color,[1 1 1]) set(gcf, defaultaxesfontsize,fs-5) for i=1:k for j=1:k subplot(k,k,(j-1)*k+i) plot(squeeze(vd(i,:,j)), Linewidth,LW) hold on plot(zeros(1,h), k ) title(sprintf( U{%d} to Var(e_{%d,t+h}).,j,i)) xlim([1 h]) set(gca, XTick,0:ceil(h/4):h) ylim([0 1]) set(gca, YTick,0:0.25:1) Fernando Pérez Forero 11
12 Figure 3.4: Forecast Error Variance decomposition Fernando Pérez Forero 12
13 A List of new commands Type help followed by the command name for more details. clear: Clear variables and functions from memory clc: Clear command window rand(m,n): Produces a m by n matrix with uniformly distributed pseudorandom numbers. diag: Produces diagonal matrices and captures the main diagonal of a matrix. eig: Eigenvalues and eigenvectors all: True (=1) if all elements of a vector are nonzero. disp: displays the array, without printing the array name. error: Display message and abort function. zeros(m,n): Produces a m by n matrix of zeros. ones(m,n): Produces a m by n matrix of ones. eye(m): Produces an identity matrix of order m. for: Repeat statements a speci c number of times. while: Repeat statements an inde nite number of times. figure: Create gure window set: Set object properties. mvnrnd(mu,sigma): Random vectors from the multivariate normal distribution (use also randn(m,n)). ceil(x): rounds the elements of X to the nearest integers towards in nity. plot: Linear plot. See also subplot and options (e.g., title, xlim, etc.) cell(m,n): Create a m by n cell array of empty matrices. sparse(x): converts a sparse or full matrix to sparse form by squeezing out any zero elements. deal: matches up the input and output lists. Fernando Pérez Forero 13
14 blkdiag: Block diagonal concatenation of matrix input arguments. full: Convert sparse matrix to full matrix. abs(x): Produces the absolute value of X. The number X can be either real or complex. max(x) / min(x): Captures the largest / smallest component. chol(x): Cholesky factorization. Uses only the diagonal and upper triangle of X. squeeze(x): returns an array with the same elements as X but with all the singleton dimensions removed. sprintf: Write formatted data to string. B MATLAB tips B.1 Growing arrays Sometimes we do not know the nal size of a vector/matrix in advance. In such cases it is useful to create an empty vector and make it grow at each iteration. However, if you know the exact size of the vector in advance, you should generate (e.g., x=zeros(1,k)) to avoid out-of-memory outcomes. x=[]; A0=rand; A1=rand; diff=abs(a1-a0); while diff>1e-2 x=[x,diff]; A0=rand; A1=rand; diff=abs(a1-a0); B.2 Cumulative sums In MATLAB we can always assign a new value of an object using the previous one. This is useful for cumulative sums: a=0; Fernando Pérez Forero 14
15 for i=1:10; a=a+1; B.3 Matrix partitions It is possible to extract a partition of an existing matrix. As a matter of fact, it is also possible to select and reorder rows and columns: A=magic(4); B=A(2:3,1:2); C=A([2 4 1],[1 3]); References Hamilton, J. D. (1994). Time Series Analysis. Princeton University Press. Lütkepohl, H. (2005). New Introduction to Multiple Time Series Analysis. Springer. Fernando Pérez Forero 15
ECONOMETRIC METHODS II TA session 2 Estimating VAR(p) processes
ECONOMETRIC METHODS II TA session 2 Estimating VAR(p) processes Fernando Pérez Forero May 3rd, 2012 1 Introduction In this second session we will cover the estimation of Vector Autoregressive (VAR) processes
More informationTitle. Description. var intro Introduction to vector autoregressive models
Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More informationVector Auto-Regressive Models
Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationVAR Models and Applications
VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More informationChapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models
Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models Fall 22 Contents Introduction 2. An illustrative example........................... 2.2 Discussion...................................
More informationVector autoregressions, VAR
1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,
More informationTOPICS IN ADVANCED ECONOMETRICS: SVAR MODELS PART I LESSON 2. Luca Fanelli. University of Bologna.
TOPICS IN ADVANCED ECONOMETRICS: SVAR MODELS PART I LESSON 2 Luca Fanelli University of Bologna luca.fanelli@unibo.it http://www.rimini.unibo.it/fanelli The material in these slides is based on the following
More informationNew Introduction to Multiple Time Series Analysis
Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2
More information2. Multivariate ARMA
2. Multivariate ARMA JEM 140: Quantitative Multivariate Finance IES, Charles University, Prague Summer 2018 JEM 140 () 2. Multivariate ARMA Summer 2018 1 / 19 Multivariate AR I Let r t = (r 1t,..., r kt
More informationVector autoregressive Moving Average Process. Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem
Vector autoregressive Moving Average Process Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem Road Map 1. Introduction 2. Properties of MA Finite Process 3. Stationarity of MA Process 4. VARMA
More informationLecture 7a: Vector Autoregression (VAR)
Lecture 7a: Vector Autoregression (VAR) 1 Big Picture We are done with univariate time series analysis Now we switch to multivariate analysis, that is, studying several time series simultaneously. VAR
More informationLecture 7a: Vector Autoregression (VAR)
Lecture 7a: Vector Autoregression (VAR) 1 2 Big Picture We are done with univariate time series analysis Now we switch to multivariate analysis, that is, studying several time series simultaneously. VAR
More information2.5 Forecasting and Impulse Response Functions
2.5 Forecasting and Impulse Response Functions Principles of forecasting Forecast based on conditional expectations Suppose we are interested in forecasting the value of y t+1 based on a set of variables
More informationMultivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]
1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector
More informationEstimating Moving Average Processes with an improved version of Durbin s Method
Estimating Moving Average Processes with an improved version of Durbin s Method Maximilian Ludwig this version: June 7, 4, initial version: April, 3 arxiv:347956v [statme] 6 Jun 4 Abstract This paper provides
More informationNumerical Methods I: Numerical linear algebra
1/3 Numerical Methods I: Numerical linear algebra Georg Stadler Courant Institute, NYU stadler@cimsnyuedu September 1, 017 /3 We study the solution of linear systems of the form Ax = b with A R n n, x,
More informationForecasting Levels of log Variables in Vector Autoregressions
September 24, 200 Forecasting Levels of log Variables in Vector Autoregressions Gunnar Bårdsen Department of Economics, Dragvoll, NTNU, N-749 Trondheim, NORWAY email: gunnar.bardsen@svt.ntnu.no Helmut
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationUNIVERSITY OF CALIFORNIA, SAN DIEGO DEPARTMENT OF ECONOMICS
2-7 UNIVERSITY OF LIFORNI, SN DIEGO DEPRTMENT OF EONOMIS THE JOHNSEN-GRNGER REPRESENTTION THEOREM: N EXPLIIT EXPRESSION FOR I() PROESSES Y PETER REINHRD HNSEN DISUSSION PPER 2-7 JULY 2 The Johansen-Granger
More informationA primer on Structural VARs
A primer on Structural VARs Claudia Foroni Norges Bank 10 November 2014 Structural VARs 1/ 26 Refresh: what is a VAR? VAR (p) : where y t K 1 y t = ν + B 1 y t 1 +... + B p y t p + u t, (1) = ( y 1t...
More informationLECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT
MARCH 29, 26 LECTURE 2 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT (Davidson (2), Chapter 4; Phillips Lectures on Unit Roots, Cointegration and Nonstationarity; White (999), Chapter 7) Unit root processes
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationStructural VARs II. February 17, 2016
Structural VARs II February 17, 216 Structural VARs Today: Long-run restrictions Two critiques of SVARs Blanchard and Quah (1989), Rudebusch (1998), Gali (1999) and Chari, Kehoe McGrattan (28). Recap:
More informationMultivariate forecasting with VAR models
Multivariate forecasting with VAR models Franz Eigner University of Vienna UK Econometric Forecasting Prof. Robert Kunst 16th June 2009 Overview Vector autoregressive model univariate forecasting multivariate
More informationPrinciples of forecasting
2.5 Forecasting Principles of forecasting Forecast based on conditional expectations Suppose we are interested in forecasting the value of y t+1 based on a set of variables X t (m 1 vector). Let y t+1
More informationLinear Systems of n equations for n unknowns
Linear Systems of n equations for n unknowns In many application problems we want to find n unknowns, and we have n linear equations Example: Find x,x,x such that the following three equations hold: x
More informationFINANCIAL ECONOMETRICS AND EMPIRICAL FINANCE -MODULE2 Midterm Exam Solutions - March 2015
FINANCIAL ECONOMETRICS AND EMPIRICAL FINANCE -MODULE2 Midterm Exam Solutions - March 205 Time Allowed: 60 minutes Family Name (Surname) First Name Student Number (Matr.) Please answer all questions by
More informationMSc Time Series Econometrics
MSc Time Series Econometrics Module 2: multivariate time series topics Lecture 1: VARs, introduction, motivation, estimation, preliminaries Tony Yates Spring 2014, Bristol Topics we will cover Vector autoregressions:
More informationIdentifying Aggregate Liquidity Shocks with Monetary Policy Shocks: An Application using UK Data
Identifying Aggregate Liquidity Shocks with Monetary Policy Shocks: An Application using UK Data Michael Ellington and Costas Milas Financial Services, Liquidity and Economic Activity Bank of England May
More informationEUI Working Papers DEPARTMENT OF ECONOMICS ECO 2009/24 DEPARTMENT OF ECONOMICS FORECASTING LEVELS OF LOG VARIABLES IN VECTOR AUTOREGRESSIONS
DEPARTMENT OF ECONOMICS EUI Working Papers ECO 2009/24 DEPARTMENT OF ECONOMICS FORECASTING LEVELS OF LOG VARIABLES IN VECTOR AUTOREGRESSIONS Gunnar Bårdsen and Helmut Lütkepohl EUROPEAN UNIVERSITY INSTITUTE,
More information(Linear equations) Applied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB (Linear equations) Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots
More informationScientific Computing
Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting
More informationUsing MATLAB. Linear Algebra
Using MATLAB in Linear Algebra Edward Neuman Department of Mathematics Southern Illinois University at Carbondale One of the nice features of MATLAB is its ease of computations with vectors and matrices.
More information5.1 Banded Storage. u = temperature. The five-point difference operator. uh (x, y + h) 2u h (x, y)+u h (x, y h) uh (x + h, y) 2u h (x, y)+u h (x h, y)
5.1 Banded Storage u = temperature u= u h temperature at gridpoints u h = 1 u= Laplace s equation u= h u = u h = grid size u=1 The five-point difference operator 1 u h =1 uh (x + h, y) 2u h (x, y)+u h
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationNumerical Linear Algebra
Chapter 3 Numerical Linear Algebra We review some techniques used to solve Ax = b where A is an n n matrix, and x and b are n 1 vectors (column vectors). We then review eigenvalues and eigenvectors and
More informationLinear Algebra Solutions 1
Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7
More information1. Introduction. Hang Qian 1 Iowa State University
Users Guide to the VARDAS Package Hang Qian 1 Iowa State University 1. Introduction The Vector Autoregression (VAR) model is widely used in macroeconomics. However, macroeconomic data are not always observed
More information3. Array and Matrix Operations
3. Array and Matrix Operations Almost anything you learned about in your linear algebra classmatlab has a command to do. Here is a brief summary of the most useful ones for physics. In MATLAB matrices
More informationB y t = γ 0 + Γ 1 y t + ε t B(L) y t = γ 0 + ε t ε t iid (0, D) D is diagonal
Structural VAR Modeling for I(1) Data that is Not Cointegrated Assume y t =(y 1t,y 2t ) 0 be I(1) and not cointegrated. That is, y 1t and y 2t are both I(1) and there is no linear combination of y 1t and
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationLinear Algebra March 16, 2019
Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationLecture 9: Markov Switching Models
Lecture 9: Markov Switching Models Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Defining a Markov Switching VAR model Structure and mechanics of Markov Switching: from
More informationMatlab for Review. NDSU Matlab Review pg 1
NDSU Matlab Review pg 1 Becoming familiar with MATLAB The console The editor The graphics windows The help menu Saving your data (diary) General environment and the console Matlab for Review Simple numerical
More informationNumerical Linea Algebra Assignment 009
Numerical Linea Algebra Assignment 009 Nguyen Quan Ba Hong Students at Faculty of Math and Computer Science, Ho Chi Minh University of Science, Vietnam email. nguyenquanbahong@gmail.com blog. http://hongnguyenquanba.wordpress.com
More informationA. Recursively orthogonalized. VARs
Orthogonalized VARs A. Recursively orthogonalized VAR B. Variance decomposition C. Historical decomposition D. Structural interpretation E. Generalized IRFs 1 A. Recursively orthogonalized Nonorthogonal
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More information1 Teaching notes on structural VARs.
Bent E. Sørensen November 8, 2016 1 Teaching notes on structural VARs. 1.1 Vector MA models: 1.1.1 Probability theory The simplest to analyze, estimation is a different matter time series models are the
More informationModeling and Stability Analysis of a Communication Network System
Modeling and Stability Analysis of a Communication Network System Zvi Retchkiman Königsberg Instituto Politecnico Nacional e-mail: mzvi@cic.ipn.mx Abstract In this work, the modeling and stability problem
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationChap 3. Linear Algebra
Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions
More informationStructural Macroeconometrics. Chapter 4. Summarizing Time Series Behavior
Structural Macroeconometrics Chapter 4. Summarizing Time Series Behavior David N. DeJong Chetan Dave The sign of a truly educated man is to be deeply moved by statistics. George Bernard Shaw This chapter
More informationSECTION 2: VECTORS AND MATRICES. ENGR 112 Introduction to Engineering Computing
SECTION 2: VECTORS AND MATRICES ENGR 112 Introduction to Engineering Computing 2 Vectors and Matrices The MAT in MATLAB 3 MATLAB The MATrix (not MAThematics) LABoratory MATLAB assumes all numeric variables
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMATLAB crash course 1 / 27. MATLAB crash course. Cesar E. Tamayo Economics - Rutgers. September 27th, /27
1/27 MATLAB crash course 1 / 27 MATLAB crash course Cesar E. Tamayo Economics - Rutgers September 27th, 2013 2/27 MATLAB crash course 2 / 27 Program Program I Interface: layout, menus, help, etc.. I Vectors
More informationNumerical Methods Lecture 2 Simultaneous Equations
CGN 42 - Computer Methods Numerical Methods Lecture 2 Simultaneous Equations Topics: matrix operations solving systems of equations Matrix operations: Adding / subtracting Transpose Multiplication Adding
More informationWeb Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models
Web Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models Diaa Noureldin Department of Economics, University of Oxford, & Oxford-Man Institute, Eagle House, Walton Well Road, Oxford OX
More informationLecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation
Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition
More informationMatrix Factorizations
1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular
More informationLecture 8: Multivariate GARCH and Conditional Correlation Models
Lecture 8: Multivariate GARCH and Conditional Correlation Models Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Three issues in multivariate modelling of CH covariances
More informationVector Autogregression and Impulse Response Functions
Chapter 8 Vector Autogregression and Impulse Response Functions 8.1 Vector Autogregressions Consider two sequences {y t } and {z t }, where the time path of {y t } is affected by current and past realizations
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationRepeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models
Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models EPSY 905: Multivariate Analysis Spring 2016 Lecture #12 April 20, 2016 EPSY 905: RM ANOVA, MANOVA, and Mixed Models
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationInference in VARs with Conditional Heteroskedasticity of Unknown Form
Inference in VARs with Conditional Heteroskedasticity of Unknown Form Ralf Brüggemann a Carsten Jentsch b Carsten Trenkler c University of Konstanz University of Mannheim University of Mannheim IAB Nuremberg
More informationGMM estimation of spatial panels
MRA Munich ersonal ReEc Archive GMM estimation of spatial panels Francesco Moscone and Elisa Tosetti Brunel University 7. April 009 Online at http://mpra.ub.uni-muenchen.de/637/ MRA aper No. 637, posted
More informationLinear Algebra: A Constructive Approach
Chapter 2 Linear Algebra: A Constructive Approach In Section 14 we sketched a geometric interpretation of the simplex method In this chapter, we describe the basis of an algebraic interpretation that allows
More information1 Teaching notes on structural VARs.
Bent E. Sørensen February 22, 2007 1 Teaching notes on structural VARs. 1.1 Vector MA models: 1.1.1 Probability theory The simplest (to analyze, estimation is a different matter) time series models are
More informationComparing Forecast Accuracy of Different Models for Prices of Metal Commodities
Comparing Forecast Accuracy of Different Models for Prices of Metal Commodities João Victor Issler (FGV) and Claudia F. Rodrigues (VALE) August, 2012 J.V. Issler and C.F. Rodrigues () Forecast Models for
More informationClassification with Gaussians
Classification with Gaussians Andreas C. Kapourani (Credit: Hiroshi Shimodaira) 09 March 2018 1 Classification In the previous lab sessions, we applied Bayes theorem in order to perform statistical pattern
More informationChapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes
Chapter 2 Some basic tools 2.1 Time series: Theory 2.1.1 Stochastic processes A stochastic process is a sequence of random variables..., x 0, x 1, x 2,.... In this class, the subscript always means time.
More informationL3: Review of linear algebra and MATLAB
L3: Review of linear algebra and MATLAB Vector and matrix notation Vectors Matrices Vector spaces Linear transformations Eigenvalues and eigenvectors MATLAB primer CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna
More informationNumerical Linear Algebra
Numerical Linear Algebra Direct Methods Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) Linear Systems: Direct Solution Methods Fall 2017 1 / 14 Introduction The solution of linear systems is one
More informationLinear Algebra, part 3 QR and SVD
Linear Algebra, part 3 QR and SVD Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2012 Going back to least squares (Section 1.4 from Strang, now also see section 5.2). We
More information7. MULTIVARATE STATIONARY PROCESSES
7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability
More informationForecasting 1 to h steps ahead using partial least squares
Forecasting 1 to h steps ahead using partial least squares Philip Hans Franses Econometric Institute, Erasmus University Rotterdam November 10, 2006 Econometric Institute Report 2006-47 I thank Dick van
More informationANOVA: Analysis of Variance - Part I
ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.
More informationCointegrated VARIMA models: specification and. simulation
Cointegrated VARIMA models: specification and simulation José L. Gallego and Carlos Díaz Universidad de Cantabria. Abstract In this note we show how specify cointegrated vector autoregressive-moving average
More informationFactor models. March 13, 2017
Factor models March 13, 2017 Factor Models Macro economists have a peculiar data situation: Many data series, but usually short samples How can we utilize all this information without running into degrees
More informationQuiz ) Locate your 1 st order neighbors. 1) Simplify. Name Hometown. Name Hometown. Name Hometown.
Quiz 1) Simplify 9999 999 9999 998 9999 998 2) Locate your 1 st order neighbors Name Hometown Me Name Hometown Name Hometown Name Hometown Solving Linear Algebraic Equa3ons Basic Concepts Here only real
More informationGMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails
GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,
More informationParametric Inference on Strong Dependence
Parametric Inference on Strong Dependence Peter M. Robinson London School of Economics Based on joint work with Javier Hualde: Javier Hualde and Peter M. Robinson: Gaussian Pseudo-Maximum Likelihood Estimation
More informationMath 502 Fall 2005 Solutions to Homework 4
Math 502 Fall 2005 Solutions to Homework 4 (1) It is easy to see that A A is positive definite and hermitian, since A is non-singular. Thus, according to Theorem 23.1 (page 174 in Trefethen and Bau), A
More informationWILEY. Differential Equations with MATLAB (Third Edition) Brian R. Hunt Ronald L. Lipsman John E. Osborn Jonathan M. Rosenberg
Differential Equations with MATLAB (Third Edition) Updated for MATLAB 2011b (7.13), Simulink 7.8, and Symbolic Math Toolbox 5.7 Brian R. Hunt Ronald L. Lipsman John E. Osborn Jonathan M. Rosenberg All
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.
More informationSOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS We want to solve the linear system a, x + + a,n x n = b a n, x + + a n,n x n = b n This will be done by the method used in beginning algebra, by successively eliminating unknowns
More informationStochastic Trends & Economic Fluctuations
Stochastic Trends & Economic Fluctuations King, Plosser, Stock & Watson (AER, 1991) Cesar E. Tamayo Econ508 - Economics - Rutgers November 14, 2011 Cesar E. Tamayo Stochastic Trends & Economic Fluctuations
More informationNumerical Linear Algebra
Numerical Linear Algebra Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt2@illinois.edu Sep 19th, 2017 C. Hurtado (UIUC - Economics) Numerical Methods On the Agenda
More informationNumerical Linear Algebra
Numerical Linear Algebra Decompositions, numerical aspects Gerard Sleijpen and Martin van Gijzen September 27, 2017 1 Delft University of Technology Program Lecture 2 LU-decomposition Basic algorithm Cost
More informationProgram Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects
Numerical Linear Algebra Decompositions, numerical aspects Program Lecture 2 LU-decomposition Basic algorithm Cost Stability Pivoting Cholesky decomposition Sparse matrices and reorderings Gerard Sleijpen
More informationMATRICES. a m,1 a m,n A =
MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of
More informationECE295, Data Assimila0on and Inverse Problems, Spring 2015
ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch
More informationTIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets
TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:
More informationProblem set 1 - Solutions
EMPIRICAL FINANCE AND FINANCIAL ECONOMETRICS - MODULE (8448) Problem set 1 - Solutions Exercise 1 -Solutions 1. The correct answer is (a). In fact, the process generating daily prices is usually assumed
More informationInteger Least Squares: Sphere Decoding and the LLL Algorithm
Integer Least Squares: Sphere Decoding and the LLL Algorithm Sanzheng Qiao Department of Computing and Software McMaster University 28 Main St. West Hamilton Ontario L8S 4L7 Canada. ABSTRACT This paper
More information