Estimation of High-dimensional Vector Autoregressive (VAR) models
|
|
- Lorraine Marshall
- 6 years ago
- Views:
Transcription
1 Estimation of High-dimensional Vector Autoregressive (VAR) models George Michailidis Department of Statistics, University of Michigan gmichail CANSSI-SAMSI Workshop, Fields Institute, Toronto May 2014 Joint work with Sumanta Basu George Michailidis (UM) High-dimensional VAR 1 / 47
2 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 2 / 47
3 Vector Autoregressive models (VAR) widely used for structural analysis and forecasting of time-varying systems capture rich dynamics among system components popular in diverse application areas control theory: system identification problems economics: estimate macroeconomic relationships (Sims, 1980) genomics: reconstructing gene regulatory network from time course data neuroscience: study functional connectivity among brain regions from fmri data (Friston, 2009) George Michailidis (UM) High-dimensional VAR 3 / 47
4 VAR models in Economics testing relationship between money and income (Sims, 1972) understanding stock price-volume relation (Hiemstra et al., 1994) dynamic effect of government spending and taxes on output (Blanchard and Jones, 2002) identify and measure the effects of monetary policy innovations on macroeconomic variables (Bernanke et al., 2005) George Michailidis (UM) High-dimensional VAR 4 / 47
5 Feb-60 Aug-60 Feb-61 Aug-61 Feb-62 Aug-62 Feb-63 Aug-63 Feb-64 Aug-64 Feb-65 Aug-65 Feb-66 Aug-66 Feb-67 Aug-67 Feb-68 Aug-68 Feb-69 Aug-69 Feb-70 Aug-70 Feb-71 Aug-71 Feb-72 Aug-72 Feb-73 Aug-73 Feb-74 Aug-74 VAR models in Economics 6 4 Employment Federal Funds Rate Consumer Price Index George Michailidis (UM) High-dimensional VAR 5 / 47
6 VAR models in Functional Genomics technological advances allow collecting huge amount of data DNA microarrays, RNA-sequencing, mass spectrometry capture meaningful biological patterns via network modeling difficult to infer direction of influence from co-expression transition patterns in time course data helps identify regulatory mechanisms George Michailidis (UM) High-dimensional VAR 6 / 47
7 VAR models in Functional Genomics (ctd) HeLa gene expression regulatory network [Courtesy: Fujita et al., 2007] George Michailidis (UM) High-dimensional VAR 7 / 47
8 VAR models in Neuroscience identify connectivity among brain regions from time course fmri data connectivity of VAR generative model (Seth et al., 2013) George Michailidis (UM) High-dimensional VAR 8 / 47
9 Model p-dimensional, discrete time, stationary process X t = {X t 1,...,Xt p} X t = A 1 X t A d X t d + ε t, ε t i.i.d N(0,Σ ε ) (1) A 1,...,A d : p p transition matrices (solid, directed edges) Σ 1 ε : contemporaneous dependence (dotted, undirected edges) stability: Eigenvalues of A (z) := I p d t=1 A tz t outside {z C, z 1} George Michailidis (UM) High-dimensional VAR 9 / 47
10 Why high-dimensional VAR? The parameter space grows quadratically (p 2 edges for p time series) order of the process (d) often unknown Economics: Forecasting with many predictors (De Mol et al., 2008) Understanding structural relationship - price puzzle" (Christiano et al., 1999) Functional Genomics: reconstruct networks among hundreds to thousands of genes experiments costly - small to moderate sample size Finance: structural changes - local stationarity George Michailidis (UM) High-dimensional VAR 10 / 47
11 Literature on high-dimensional VAR models Economics: Bayesian vector autoregression (lasso, ridge penalty; Litterman, Minnesota Prior) Factor model based approach (FAVAR, dynamic factor models) Bioinformatics: Discovering gene regulatory mechanisms using pairwise VARs (Fujita et al., 2007 and Mukhopadhyay and Chatterjee, 2007) Penalized VAR with grouping effects over time (Lozano et al., 2009) Truncated lasso and thesholded lasso variants (Shojaie and Michailidis, 2010 and Shojaie, Basu and Michailidis, 2012) Statistics: lasso (Han and Liu, 2013) and group lasso penalty (Song and Bickel, 2011) low-rank modeling with nuclear norm penalty (Negahban and Wainwright, 2011) sparse VAR modeling via two-stage procedures (Davis et al., 2012) George Michailidis (UM) High-dimensional VAR 11 / 47
12 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 12 / 47
13 Model p-dimensional, discrete time, stationary process X t = {X t 1,...,Xt p} X t = A 1 X t A d X t d + ε t, ε t i.i.d N(0,Σ ε ) (2) A 1,...,A d : p p transition matrices (solid, directed edges) Σ 1 ε : contemporaneous dependence (dotted, undirected edges) stability: Eigenvalues of A (z) := I p d t=1 A tz t outside {z C, z 1} George Michailidis (UM) High-dimensional VAR 13 / 47
14 Detour: VARs and Granger Causality Concept introduced by Granger (1969) A time series X is said to Granger-cause Y if it can be shown, usually through a series of F-tests on lagged values of X (and with lagged values of Y also known), that those X values provide statistically significant information about future values of Y. In the context of a high-dimensional VAR model we have that X T t j is Granger-causal for X T i if A t i,j 0. Granger-causality does not imply true causality; it is built on correlations Also, related to estimating a Directed Acyclic Graph (DAG) with (d + 1) p variables, with a known ordering of the variables George Michailidis (UM) High-dimensional VAR 14 / 47
15 Estimating VARs through regression data: {X 0,X 1,...,X T } - one replicate, observed at T + 1 time points construct autoregression (X T ) (X T 1 ) (X T 2 ) (X T d ) (X T 1 ) (X T 2 ) (X T 3 ) (X T 1 d ) = } (X d ) {{ } } (X d 1 ) (X d 2 ) {{ (X 0 ) } Y X A 1. A d } {{ } B + (ε T ) (ε T 1 ).. (ε d ) } {{ } E vec(y ) = vec(x B ) + vec(e) Y }{{} Np 1 = (I X )vec(b ) + vec(e) = }{{} Z Np q β }{{} q 1 +vec(e) }{{} Np 1 N = (T d + 1), q = dp 2 Assumption : A t are sparse, d t=1 A t 0 k vec(e) N (0,Σ ε I) George Michailidis (UM) High-dimensional VAR 15 / 47
16 Estimates l 1 -penalized least squares (l 1 -LS) 1 argmin β R q N Y Zβ 2 + λ N β 1 l 1 -penalized log-likelihood (l 1 -LL) (Davis et al., 2012) 1 argmin β R q N (Y ( Zβ) Σ 1 ε I ) (Y Zβ) + λ N β 1 George Michailidis (UM) High-dimensional VAR 16 / 47
17 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 17 / 47
18 Detour: Consistency of Lasso Regression Y n 1 LASSO : = X n p β 1 ˆβ := argmin β R p n Y Xβ 2 + λ n β 1 p 1 + ε { } S = j {1,...,p} βj i.i.d. 0, card(s) = k, k n, ε i N(0,σ 2 ) Restricted Eigenvalue (RE): Assume 1 α RE := min v R p, v 1, v S c 1 3 v S 1 n Xv 2 > 0 n 1 Estimation error: ˆβ β Q(X,σ) 1 k logp α RE n with high probability George Michailidis (UM) High-dimensional VAR 18 / 47
19 Verifying Restricted Eigenvalue Condition Raskutti et al. (2010): If the rows of X i.i.d. N(0,Σ X ) and Σ X satisfies RE, then X satisfies RE with high probability. Assumption of independence among rows crucial Rudelson and Zhou (2013): If the design matrix X can be factorized as X = ΨA where A satisfies RE and Ψ acts as (almost) an isometry on the images of sparse vectors under A, then X satisfies RE with high probability. George Michailidis (UM) High-dimensional VAR 19 / 47
20 Back to Vector Autoregression Random design matrix X, correlated with error matrix E (X T ) (X T 1 ) (X T 2 ) (X T d ) (X T 1 ) (X T 2 ) (X T 3 ) (X T 1 d ) A 1 = (X d ) (X d 1 ) (X d 2 ) (X 0 ) A d }{{}}{{}}{{} B Y X vec(y ) = vec(x B ) + vec(e) Y }{{} Np 1 = (I X )vec(b ) + vec(e) = }{{} Z Np q β }{{} q 1 +vec(e) }{{} Np 1 N = (T d + 1), q = dp 2 + vec(e) N (0,Σ ε I) (ε T ) (ε T 1 ). } (ε d ) {{ } E George Michailidis (UM) High-dimensional VAR 20 / 47
21 Vector Autoregression (ctd) Key Questions: How often does RE hold? How small is α RE? How does the cross-correlation affect convergence rates? George Michailidis (UM) High-dimensional VAR 21 / 47
22 Consistency of VAR estimates Restricted Eigenvalue (RE) assumption: (I X ) q q RE(α,τ(N,q)) with α > 0,τ(N,q) > 0 if θ ( I X X /N ) θ α θ 2 2 τ(n,q) θ 2 1 for all θ R q (3) Deviation Condition: There exists a function Q(β,Σ ε ) such that vec ( X E/N ) log d + 2 log p max Q(β,Σ ε ) N Key Result: Estimation Consistency: If (3) and (4) hold with kτ(n, q) α/32, then, for any λ N 4Q(β,Σ ε ) (log d + 2 log p)/n, lasso estimate ˆβ l1 satisfies ˆβ l1 β 64 Q(β,Σ ε ) k (log d + 2 log p) α N (4) George Michailidis (UM) High-dimensional VAR 22 / 47
23 Verifying RE and Deviation Condition Negahban and Wainwright, 2011: for VAR(1) models, assume A 1 < 1, where A := Λ max (A A) For p = 1, d = 1, X t = ρx t 1 + ε t, reduces to ρ < 1 - equivalent to stability Han and Liu, 2013: for VAR(d) models, reformulate as VAR(1): X t = à 1 X t 1 + ε t, where X t = X t X t 1. X t d+1 Assume à 1 < 1 dp 1 à 1 = A 1 A 2 A d 1 A d I p I p I p 0 dp dp ε t = ε t 0. 0 dp 1 George Michailidis (UM) High-dimensional VAR 23 / 47
24 VAR(1): Stability and A 1 < 1 A 1 1 for many stable VAR(1) models [ X t = A 1 X t 1 + ε t α 0, A 1 = β α ] (α, β): {X t } stable 5 β α t 2 X 1 t 1 X 1 α t X β α t 2 t 1 t X 2 X 2 X 2 (α, β): A 1 < 1 5 George Michailidis (UM) High-dimensional VAR 24 / 47
25 VAR(d): Stability and à 1 < 1 à 1 1 for any stable VAR(d) models, if d > 1 [ X t = 2αX t 1 α 2 X t 2 + ε t X t, X t 1 ] [ 2α α 2 = 1 0 ][ X t 1 X t 2 ] [ ε t + 0 ] 3 ~ A 1 α 2 2α X t 2 X t 1 X t 2 ~ A 1 = 1 0 α George Michailidis (UM) High-dimensional VAR 25 / 47
26 Stable VAR models V VAR(2), VAR(3),... VAR(d),... VAR(1) A < 1 George Michailidis (UM) High-dimensional VAR 26 / 47
27 Stable VAR models V VAR(2), VAR(3),... VAR(d),... VAR(1) A < 1 George Michailidis (UM) High-dimensional VAR 27 / 47
28 Quantifying Stability through the Spectral Density Spectral density function of a covariance stationary process {X t }, f X (θ) = 1 2π Γ X (l)e ilθ, l= θ [ π,π] Γ X (l) = E [ X t (X t+l ) ], autocovariance matrix of order l If the VAR process is stable, it has a closed form (Priestley, 1981) f X (θ) = 1 ( ) 1 ( ) 1 A (e iθ ) Σε A (e iθ ) 2π The two sources of dependence factorize in frequency domain George Michailidis (UM) High-dimensional VAR 28 / 47
29 Quantifying Stability by Spectral Density Quantifying Stability by Spectral Density For univariate processes, peak" of the spectral density measures stability of the For process univariate - (sharper processes, peak the = less peak" stable) of the spectral density measures stability of the process - (sharper peak = less stable) Autocovariance Γ(h) ρ=0.1 ρ=0.5 ρ=0.7 f(θ) ρ=0.1 ρ=0.5 ρ= lag (h) θ (f) Autocovariance of AR(1) (g) Spectral Density of AR(1) For multivariate processes, similar role is played by the maximum eigenvalue of For multivariate processes, similar role is played by the maximum eigenvalue of the (matrix-valued) spectral density the (matrix-valued) spectral density Sumanta Basu (UM) Network Estimation in Time Series 52 / 56 George Michailidis (UM) High-dimensional VAR 29 / 47
30 Quantifying Stability by Spectral Density For a stable VAR(d) process {X t }, the maximum eigenvalue of its spectral density captures its stability M (f X ) = max Λ max (f X (θ)) θ [ π,π] The minimum eigenvalue of the spectral density captures dependence among its components m(f X ) = min Λ min (f X (θ)) θ [ π,π] For stable VAR(1) processes, M (f X ) scales with (1 ρ(a 1 )) 2, ρ(a 1 ) is the spectral radius of A 1 m(f X ) scales with the capacity (maximum incoming + outgoing effect at a node) of the underlying graph George Michailidis (UM) High-dimensional VAR 30 / 47
31 Consistency of VAR estimates Theorem Consider a random realization {X 0,...,X T } generated according to a stable VAR(d) process with Λ min (Σ ε ) > 0. Then there exist deterministic functions φ i (A t,σ ε ) > 0 and constants c i > 0 such that for N φ 0 (A t,σ ε ) k(logd + 2logp)/N, the lasso estimate (l 1 -LS) with λ N (2logp + logd)/n satisfies, with probability at least 1 c 1 exp[ c 2 (2logp + logd)], d ( ) Aˆ h A h φ1 (A t,σ ε ) k(logd + 2logp)/N h=1 T 1 d ( ) N ( h A h )X h φ 2 (A t,σ ε ) k(logd + 2logp)/N t=d h=1 ) Further, a thresholded version of lasso à = (Ât,ij 1 {  t,ij > λ N } satisfies supp(ã 1:d )\supp(a 1:d ) φ 3 (A t,σ ε )k φ i (A t,σ ε ) are large when M (f X ) is large and m(f X ) is small. George Michailidis (UM) High-dimensional VAR 31 / 47
32 Some Remarks Convergence rates governed by: dimensionality parameters - dimension of the process (p), order of the process (d), number of parameters (k) in the transition matrices A i and sample size (N = T -d + 1) internal parameters - curvature (α), tolerance (τ) and the deviation bound Q(β,Σ ε ) The squared l 2 -errors of estimation and prediction scale with the dimensionality parameters as k(2logp + logd)/n, similar to the rates obtained when the observations are independent The temporal and cross-sectional dependence affect the rates only through the internal parameters. Typically, the rates are better when α is large and Q(β,Σ ε ),τ are small. This dependence is captured in the next results. George Michailidis (UM) High-dimensional VAR 32 / 47
33 Verifying RE Proposition Consider a random realization {X 0,...,X T } generated according to a stable VAR(d) process. Then there exist universal positive constants c i such that for all N max{1,ω 2 }k log(dp), with probability at least 1 c 1 exp( c 2 N min{ω 2,1}), where I p (X X /N) RE(α,τ), ω = Λ min(σ ε )/Λ max (Σ ε ) µ max (A )/µ min ( A ), α = Λ min(σ ε ) 2µ max (A ), Λ min (Σ ε ) τ(n,q) = c 3 µ max (A ) max{ω 2,1} log(dp). N George Michailidis (UM) High-dimensional VAR 33 / 47
34 Verifying Deviation Condition Proposition If q 2, then, for any A > 0, N logd + 2logp, with probability at least 1 12q A, we have vec ( X E/N ) logd + 2logp max Q(β,Σ ε ), N where Q(β,Σ ε ) = ( (A + 1)) [ Λ max (Σ ε ) + Λ max(σ ε ) µ min (A ) + Λ ] max(σ ε )µ max (A ) µ min (A ) George Michailidis (UM) High-dimensional VAR 34 / 47
35 Some Comments RE: the convergence rates are faster for larger α and smaller τ. From the expressions of ω,α and τ, it is clear that the VAR estimates have lower error bounds when Λ max (Σ ε ), µ max (A ) are smaller and Λ min (Σ ε ), µ min (A ) are larger. Deviation bound: VAR estimates exhibit lower error bounds when Λ max (Σ ε ), µ max (A ) are smaller and Λ min (Σ ε ), µ min (A ) are larger (similar to RE) George Michailidis (UM) High-dimensional VAR 35 / 47
36 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 36 / 47
37 l 1 -LS: Denote the i th column of a matrix M by M i. arg 1 min β R q N Y Zβ 2 + λ N β 1 1 arg min B 1,...,B p N p i X B i i=1 Y 2 p + λ N B i 1 i=1 Amounts to running p separate LASSO programs, each with dp predictors: Y i X, i = 1,...,p. George Michailidis (UM) High-dimensional VAR 37 / 47
38 l 1 -LL: Davis et al, 2012, proposed the following algorithm: arg 1 arg min β R q N 1 min β R q N (Y ( Zβ) Σ 1 ( Σ 1/2 ε ε I ) (Y Zβ) + λ N β 1 ) Σ 1/2 ε X β 2 + λ N β 1 ) ( I Y ( Amounts to ) running a single LASSO program with dp 2 predictors: Σ 1/2 ε I Y Σ 1/2 ε X - cannot be implemented in parallel. σ ij ε := (i,j) th entry of Σ 1. The objective function is 1 N p p i=1 j=1 ε σ ij ε (Y i X B i ) (Y j X B j ) + λ N p k=1 B k 1 George Michailidis (UM) High-dimensional VAR 38 / 47
39 Block Coordinate Descent for l 1 -LL 1 pre-select d. Run l 1 -LS to get ˆB, ˆΣ 1 ε. 2 iterate till convergence: 1 For i = 1,...,p, set r i := (1/2 ˆσ ii ε ) j i update ˆB i = argmin B i ˆσ ε ij ( ) Yj X ˆB j ˆσ ii ε N (Y i + r i ) X B i 2 + λ N B i 1 each iteration amounts to running p separate LASSO programs, each with dp predictors: Y i + r i X, i = 1,...,p. Can be implemented in parallel George Michailidis (UM) High-dimensional VAR 39 / 47
40 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 40 / 47
41 VAR models considered Small Size VAR, p = 10,d = 1,T = 30,50 Medium Size VAR, p = 30,d = 1,T = 80,120,160 In each setting, we generate an adjacency matrix A 1 with 5 10% non-zero edges selected at random and rescale to ensure that the process is stable with SNR = 2. We generate three different error processes with covariance matrix Σ ε from one of the following families: 1 Block-I: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ii = 1, σ ε,ij = ρ if 1 i j p/2, 0 otherwise; 2 Block-II: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ii = 1, σ ε,ij = ρ if 1 i j p/2 or p/2 < i j p, 0 otherwise; 3 bf Toeplitz: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ij = ρ i j. George Michailidis (UM) High-dimensional VAR 41 / 47
42 VAR models considered (ctd) REGUALARIZED ESTIMATION IN TIME SERIES (a) A 1 (b) Σ ɛ : Block-I (c) Σ ɛ : Block-II (d) Σ ɛ : Toeplitz We let ρ vary in {0.5,0.7,0.9}. : Adjacency matrix A 1 and error covariance matrix Σ ɛ of different ty Larger values of ρ indicate that the error processes are more strongly in the simulation studies correlated. Block-I: Σ ɛ = ((σ ɛ,ij )) 1 i,j p with σ ɛ,ii = 1, σ ɛ,ij = ρ if 1 i j George Michailidis (UM) High-dimensional VAR 42 / 47
43 Comparisons and Performance Criteria Different methods for VAR estimation: OLS l 1 -LS l 1 -LL l 1 -LL-O (Oracle version, assuming Σ ε known) Ridge evaluated using the following performance metrics: 1 Model Selection: Area under receiving operator characteristic curve (AUROC) 2 Estimation error: Relative estimation accuracy measured by ˆB B F / B F George Michailidis (UM) High-dimensional VAR 43 / 47
44 Results I Table: VAR(1) model with p = 10, T = 30 BLOCK-I BLOCK-II Toeplitz ρ AUROC l 1 -LS l 1 -LL l 1 -LL-O Estimation OLS Error l 1 -LS l 1 -LL l 1 -LL-O ridge George Michailidis (UM) High-dimensional VAR 44 / 47
45 Results II Table: VAR(1) model with p = 30, T = 120 BLOCK-I BLOCK-II Toeplitz ρ AUROC l 1 -LS l 1 -LL l 1 -LL-O Estimation OLS Error l 1 -LS l 1 -LL l 1 -LL-O Ridge George Michailidis (UM) High-dimensional VAR 45 / 47
46 Summary/Discussion Investigated penalized VAR estimation in high-dimension Established estimation consistency for all stable VAR models, based on novel techniques using spectral representation of stationary processes Developed parallellizable algorithm for likelihood based VAR estimates There is extensive work on characterizing univariate time series, through mixing conditions or functional dependence measures. However, thre is little work for multivariate series, which is needed to be able to provide results in the current setting. George Michailidis (UM) High-dimensional VAR 46 / 47
47 References S. Basu and G. Michailidis, Estimation in High-dimensional Vector Autoregressive Models, arxiv: S. Basu, A. Shojaie and G. Michailidis, Network Granger Causality with Inherent Grouping Structure, revised for JMLR George Michailidis (UM) High-dimensional VAR 47 / 47
REGULARIZED ESTIMATION IN SPARSE HIGH-DIMENSIONAL TIME SERIES MODELS. BY SUMANTA BASU AND GEORGE MICHAILIDIS 1 University of Michigan
The Annals of Statistics 2015, Vol. 43, No. 4, 1535 1567 DOI: 10.1214/15-AOS1315 Institute of Mathematical Statistics, 2015 REGULARIZED ESTIMATION IN SPARSE HIGH-DIMENSIONAL TIME SERIES MODELS BY SUMANTA
More informationHigh-dimensional covariance estimation based on Gaussian graphical models
High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,
More informationarxiv: v1 [stat.me] 28 Jul 2017
Sparse Identification and Estimation of High-Dimensional Vector AutoRegressive Moving Averages Ines Wilms a,b, Sumanta Basu a, Jacob Bien c, and David S. Matteson a arxiv:707.09208v [stat.me] 28 Jul 207
More informationBAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage
BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement
More informationVariable Selection for Highly Correlated Predictors
Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates
More information25 : Graphical induced structured input/output models
10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large
More informationRegularized Estimation of High-dimensional Factor-Augmented Vector Autoregressive (FAVAR) Models
Regularized Estimation of High-dimensional Factor-Augmented Vector Autoregressive (FAVAR) Models George Michailidis University of Florida Department of Statistics and the Informatics Institute February
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationLasso Guarantees for Dependent Data
Lasso Guarantees for Dependent Data by Kam Chung Wong A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy (Statistics) in The University of Michigan
More informationFinite Time Analysis of Vector Autoregressive Models under Linear Restrictions
Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions Yao Zheng collaboration with Guang Cheng Department of Statistics, Purdue University Talk at Math, IUPUI October 23, 2018
More informationLearning discrete graphical models via generalized inverse covariance matrices
Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,
More informationEstimating Structured Vector Autoregressive Models
Estimating Structured Vector Autoregressive Models Igor Melnyk Arindam Banerjee Department of Computer Science and Engineering, University of Minnesota, Twin Cities MELYK@CSUMEDU BAERJEE@CSUMEDU Abstract
More informationVisualizing VAR s: Regularization and Network Tools for High-Dimensional Financial Econometrics
Visualizing VAR s: Regularization and Network Tools for High-Dimensional Financial Econometrics Francis X. Diebold University of Pennsylvania March 7, 2015 1 / 32 DGP: N-Variable VAR(p), t = 1,..., T Φ(L)x
More informationVector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.
Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series
More informationAn algorithm for the multivariate group lasso with covariance estimation
An algorithm for the multivariate group lasso with covariance estimation arxiv:1512.05153v1 [stat.co] 16 Dec 2015 Ines Wilms and Christophe Croux Leuven Statistics Research Centre, KU Leuven, Belgium Abstract
More informationHigh-dimensional Statistics
High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given
More informationFrontiers in Forecasting, Minneapolis February 21-23, Sparse VAR-Models. Christophe Croux. EDHEC Business School (France)
Frontiers in Forecasting, Minneapolis February 21-23, 2018 Sparse VAR-Models Christophe Croux EDHEC Business School (France) Joint Work with Ines Wilms (Cornell University), Luca Barbaglia (KU leuven),
More informationEstimators based on non-convex programs: Statistical and computational guarantees
Estimators based on non-convex programs: Statistical and computational guarantees Martin Wainwright UC Berkeley Statistics and EECS Based on joint work with: Po-Ling Loh (UC Berkeley) Martin Wainwright
More informationFinancial Econometrics
Financial Econometrics Multivariate Time Series Analysis: VAR Gerald P. Dwyer Trinity College, Dublin January 2013 GPD (TCD) VAR 01/13 1 / 25 Structural equations Suppose have simultaneous system for supply
More information11 : Gaussian Graphic Models and Ising Models
10-708: Probabilistic Graphical Models 10-708, Spring 2017 11 : Gaussian Graphic Models and Ising Models Lecturer: Bryon Aragam Scribes: Chao-Ming Yen 1 Introduction Different from previous maximum likelihood
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationFactor models. March 13, 2017
Factor models March 13, 2017 Factor Models Macro economists have a peculiar data situation: Many data series, but usually short samples How can we utilize all this information without running into degrees
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationPenalization method for sparse time series model
Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS020) p.6280 Penalization method for sparse time series model Heewon, Park Chuo University, Department of Mathemetics
More informationNew Introduction to Multiple Time Series Analysis
Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2
More informationVector Auto-Regressive Models
Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More informationAdaptive Thresholding for Reconstructing Regulatory Networks from Time Course Gene Expression Data
Noname manuscript No. (will be inserted by the editor) Adaptive Thresholding for Reconstructing Regulatory Networks from Time Course Gene Expression Data Ali Shojaie Sumanta Basu George Michailidis Received:
More informationVAR Models and Applications
VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions
More information25 : Graphical induced structured input/output models
10-708: Probabilistic Graphical Models 10-708, Spring 2013 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Meghana Kshirsagar (mkshirsa), Yiwen Chen (yiwenche) 1 Graph
More informationInference For High Dimensional M-estimates. Fixed Design Results
: Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and
More informationHigh-dimensional Statistical Models
High-dimensional Statistical Models Pradeep Ravikumar UT Austin MLSS 2014 1 Curse of Dimensionality Statistical Learning: Given n observations from p(x; θ ), where θ R p, recover signal/parameter θ. For
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationInference in VARs with Conditional Heteroskedasticity of Unknown Form
Inference in VARs with Conditional Heteroskedasticity of Unknown Form Ralf Brüggemann a Carsten Jentsch b Carsten Trenkler c University of Konstanz University of Mannheim University of Mannheim IAB Nuremberg
More information2.5 Forecasting and Impulse Response Functions
2.5 Forecasting and Impulse Response Functions Principles of forecasting Forecast based on conditional expectations Suppose we are interested in forecasting the value of y t+1 based on a set of variables
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationWithin Group Variable Selection through the Exclusive Lasso
Within Group Variable Selection through the Exclusive Lasso arxiv:1505.07517v1 [stat.me] 28 May 2015 Frederick Campbell Department of Statistics, Rice University and Genevera Allen Department of Statistics,
More informationRegularized Estimation of High Dimensional Covariance Matrices. Peter Bickel. January, 2008
Regularized Estimation of High Dimensional Covariance Matrices Peter Bickel Cambridge January, 2008 With Thanks to E. Levina (Joint collaboration, slides) I. M. Johnstone (Slides) Choongsoon Bae (Slides)
More informationA Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models
A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los
More informationBayesian Compressed Vector Autoregressions
Bayesian Compressed Vector Autoregressions Gary Koop a, Dimitris Korobilis b, and Davide Pettenuzzo c a University of Strathclyde b University of Glasgow c Brandeis University 9th ECB Workshop on Forecasting
More informationFactor models. May 11, 2012
Factor models May 11, 2012 Factor Models Macro economists have a peculiar data situation: Many data series, but usually short samples How can we utilize all this information without running into degrees
More informationDynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano
Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively
More informationLog Covariance Matrix Estimation
Log Covariance Matrix Estimation Xinwei Deng Department of Statistics University of Wisconsin-Madison Joint work with Kam-Wah Tsui (Univ. of Wisconsin-Madsion) 1 Outline Background and Motivation The Proposed
More informationHigh Dimensional Covariance and Precision Matrix Estimation
High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance
More informationThe Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression
The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression Cun-hui Zhang and Jian Huang Presenter: Quefeng Li Feb. 26, 2010 un-hui Zhang and Jian Huang Presenter: Quefeng The Sparsity
More informationBayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1
Bayesian Graphical Models for Structural Vector Autoregressive Processes Daniel Ahelegbey, Monica Billio, and Roberto Cassin (2014) March 21, 2015 Bayesian Graphical Models for Structural Vector AutoregressiveMarch
More informationSparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda
Sparse regression Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 3/28/2016 Regression Least-squares regression Example: Global warming Logistic
More informationESL Chap3. Some extensions of lasso
ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied
More informationVariable Targeting and Reduction in High-Dimensional Vector Autoregressions
Variable Targeting and Reduction in High-Dimensional Vector Autoregressions Tucker McElroy (U.S. Census Bureau) Frontiers in Forecasting February 21-23, 2018 1 / 22 Disclaimer This presentation is released
More informationSwitching Regime Estimation
Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms
More informationGraphical Model Selection
May 6, 2013 Trevor Hastie, Stanford Statistics 1 Graphical Model Selection Trevor Hastie Stanford University joint work with Jerome Friedman, Rob Tibshirani, Rahul Mazumder and Jason Lee May 6, 2013 Trevor
More informationSparse inverse covariance estimation with the lasso
Sparse inverse covariance estimation with the lasso Jerome Friedman Trevor Hastie and Robert Tibshirani November 8, 2007 Abstract We consider the problem of estimating sparse graphs by a lasso penalty
More informationInference For High Dimensional M-estimates: Fixed Design Results
Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49
More informationarxiv: v1 [math.st] 13 Feb 2012
Sparse Matrix Inversion with Scaled Lasso Tingni Sun and Cun-Hui Zhang Rutgers University arxiv:1202.2723v1 [math.st] 13 Feb 2012 Address: Department of Statistics and Biostatistics, Hill Center, Busch
More informationKatsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract
Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach
More informationTitle. Description. var intro Introduction to vector autoregressive models
Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models
More informationFactor-Adjusted Robust Multiple Test. Jianqing Fan (Princeton University)
Factor-Adjusted Robust Multiple Test Jianqing Fan Princeton University with Koushiki Bose, Qiang Sun, Wenxin Zhou August 11, 2017 Outline 1 Introduction 2 A principle of robustification 3 Adaptive Huber
More informationHigh-dimensional statistics: Some progress and challenges ahead
High-dimensional statistics: Some progress and challenges ahead Martin Wainwright UC Berkeley Departments of Statistics, and EECS University College, London Master Class: Lecture Joint work with: Alekh
More informationConfidence Intervals for Low-dimensional Parameters with High-dimensional Data
Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationGeneralized Concomitant Multi-Task Lasso for sparse multimodal regression
Generalized Concomitant Multi-Task Lasso for sparse multimodal regression Mathurin Massias https://mathurinm.github.io INRIA Saclay Joint work with: Olivier Fercoq (Télécom ParisTech) Alexandre Gramfort
More informationTowards a Regression using Tensors
February 27, 2014 Outline Background 1 Background Linear Regression Tensorial Data Analysis 2 Definition Tensor Operation Tensor Decomposition 3 Model Attention Deficit Hyperactivity Disorder Data Analysis
More informationInference when identifying assumptions are doubted. A. Theory. Structural model of interest: B 1 y t1. u t. B m y tm. u t i.i.d.
Inference when identifying assumptions are doubted A. Theory B. Applications Structural model of interest: A y t B y t B m y tm nn n i.i.d. N, D D diagonal A. Theory Bayesian approach: Summarize whatever
More informationThe lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding
Patrick Breheny February 15 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/24 Introduction Last week, we introduced penalized regression and discussed ridge regression, in which the penalty
More informationGenetic Networks. Korbinian Strimmer. Seminar: Statistical Analysis of RNA-Seq Data 19 June IMISE, Universität Leipzig
Genetic Networks Korbinian Strimmer IMISE, Universität Leipzig Seminar: Statistical Analysis of RNA-Seq Data 19 June 2012 Korbinian Strimmer, RNA-Seq Networks, 19/6/2012 1 Paper G. I. Allen and Z. Liu.
More informationProperties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation
Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Adam J. Rothman School of Statistics University of Minnesota October 8, 2014, joint work with Liliana
More informationRegression with time series
Regression with time series Class Notes Manuel Arellano February 22, 2018 1 Classical regression model with time series Model and assumptions The basic assumption is E y t x 1,, x T = E y t x t = x tβ
More informationCointegration Lecture I: Introduction
1 Cointegration Lecture I: Introduction Julia Giese Nuffield College julia.giese@economics.ox.ac.uk Hilary Term 2008 2 Outline Introduction Estimation of unrestricted VAR Non-stationarity Deterministic
More informationAn iterative hard thresholding estimator for low rank matrix recovery
An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical
More informationCombining Macroeconomic Models for Prediction
Combining Macroeconomic Models for Prediction John Geweke University of Technology Sydney 15th Australasian Macro Workshop April 8, 2010 Outline 1 Optimal prediction pools 2 Models and data 3 Optimal pools
More informationYou must continuously work on this project over the course of four weeks.
The project Five project topics are described below. You should choose one the projects. Maximum of two people per project is allowed. If two people are working on a topic they are expected to do double
More informationAn efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss
An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao
More informationHigh-dimensional Covariance Estimation Based On Gaussian Graphical Models
High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix
More informationAutoregressive distributed lag models
Introduction In economics, most cases we want to model relationships between variables, and often simultaneously. That means we need to move from univariate time series to multivariate. We do it in two
More informationHierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets
Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,
More informationChapter 17: Undirected Graphical Models
Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)
More informationComposite Loss Functions and Multivariate Regression; Sparse PCA
Composite Loss Functions and Multivariate Regression; Sparse PCA G. Obozinski, B. Taskar, and M. I. Jordan (2009). Joint covariate selection and joint subspace selection for multiple classification problems.
More informationVector Autoregressive based Network Models
Vector Autoregressive based Network Models George Michailidis Informatics Institute University of Florida Dynamic Networks Workshop Isaac Newton Institute Cambridge, December 2016 George Michailidis VAR
More informationPenalized Estimation of Panel Vector Autoregressive Models: A Lasso Approach
Penalized Estimation of Panel Vector Autoregressive Models: A Lasso Approach Annika Schnücker Freie Universität Berlin and DIW Berlin Graduate Center, Mohrenstr. 58, 10117 Berlin, Germany October 11, 2017
More information1. Fundamental concepts
. Fundamental concepts A time series is a sequence of data points, measured typically at successive times spaced at uniform intervals. Time series are used in such fields as statistics, signal processing
More informationLecture 1: Information and Structural VARs
Lecture 1: Information and Structural VARs Luca Gambetti 1 1 Universitat Autònoma de Barcelona LBS, May 6-8 2013 Introduction The study of the dynamic effects of economic shocks is one of the key applications
More informationSingle Equation Linear GMM with Serially Correlated Moment Conditions
Single Equation Linear GMM with Serially Correlated Moment Conditions Eric Zivot October 28, 2009 Univariate Time Series Let {y t } be an ergodic-stationary time series with E[y t ]=μ and var(y t )
More informationIntroduction to Algorithmic Trading Strategies Lecture 3
Introduction to Algorithmic Trading Strategies Lecture 3 Pairs Trading by Cointegration Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Distance method Cointegration Stationarity
More informationSparse Learning and Distributed PCA. Jianqing Fan
w/ control of statistical errors and computing resources Jianqing Fan Princeton University Coauthors Han Liu Qiang Sun Tong Zhang Dong Wang Kaizheng Wang Ziwei Zhu Outline Computational Resources and Statistical
More informationLinear regression methods
Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response
More informationInference when identifying assumptions are doubted. A. Theory B. Applications
Inference when identifying assumptions are doubted A. Theory B. Applications 1 A. Theory Structural model of interest: A y t B 1 y t1 B m y tm u t nn n1 u t i.i.d. N0, D D diagonal 2 Bayesian approach:
More informationQuestions and Answers on Unit Roots, Cointegration, VARs and VECMs
Questions and Answers on Unit Roots, Cointegration, VARs and VECMs L. Magee Winter, 2012 1. Let ɛ t, t = 1,..., T be a series of independent draws from a N[0,1] distribution. Let w t, t = 1,..., T, be
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationPermutation-invariant regularization of large covariance matrices. Liza Levina
Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work
More informationRobust high-dimensional linear regression: A statistical perspective
Robust high-dimensional linear regression: A statistical perspective Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics STOC workshop on robustness and nonconvexity Montreal,
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationIEOR165 Discussion Week 5
IEOR165 Discussion Week 5 Sheng Liu University of California, Berkeley Feb 19, 2016 Outline 1 1st Homework 2 Revisit Maximum A Posterior 3 Regularization IEOR165 Discussion Sheng Liu 2 About 1st Homework
More informationIndirect multivariate response linear regression
Biometrika (2016), xx, x, pp. 1 22 1 2 3 4 5 6 C 2007 Biometrika Trust Printed in Great Britain Indirect multivariate response linear regression BY AARON J. MOLSTAD AND ADAM J. ROTHMAN School of Statistics,
More information7. MULTIVARATE STATIONARY PROCESSES
7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability
More informationEstimation of large dimensional sparse covariance matrices
Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)
More informationLesson 17: Vector AutoRegressive Models
Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Vector AutoRegressive models The extension of ARMA models into a multivariate framework
More informationStochastic Processes
Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationA Significance Test for the Lasso
A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical
More information