Estimation of High-dimensional Vector Autoregressive (VAR) models

Size: px
Start display at page:

Download "Estimation of High-dimensional Vector Autoregressive (VAR) models"

Transcription

1 Estimation of High-dimensional Vector Autoregressive (VAR) models George Michailidis Department of Statistics, University of Michigan gmichail CANSSI-SAMSI Workshop, Fields Institute, Toronto May 2014 Joint work with Sumanta Basu George Michailidis (UM) High-dimensional VAR 1 / 47

2 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 2 / 47

3 Vector Autoregressive models (VAR) widely used for structural analysis and forecasting of time-varying systems capture rich dynamics among system components popular in diverse application areas control theory: system identification problems economics: estimate macroeconomic relationships (Sims, 1980) genomics: reconstructing gene regulatory network from time course data neuroscience: study functional connectivity among brain regions from fmri data (Friston, 2009) George Michailidis (UM) High-dimensional VAR 3 / 47

4 VAR models in Economics testing relationship between money and income (Sims, 1972) understanding stock price-volume relation (Hiemstra et al., 1994) dynamic effect of government spending and taxes on output (Blanchard and Jones, 2002) identify and measure the effects of monetary policy innovations on macroeconomic variables (Bernanke et al., 2005) George Michailidis (UM) High-dimensional VAR 4 / 47

5 Feb-60 Aug-60 Feb-61 Aug-61 Feb-62 Aug-62 Feb-63 Aug-63 Feb-64 Aug-64 Feb-65 Aug-65 Feb-66 Aug-66 Feb-67 Aug-67 Feb-68 Aug-68 Feb-69 Aug-69 Feb-70 Aug-70 Feb-71 Aug-71 Feb-72 Aug-72 Feb-73 Aug-73 Feb-74 Aug-74 VAR models in Economics 6 4 Employment Federal Funds Rate Consumer Price Index George Michailidis (UM) High-dimensional VAR 5 / 47

6 VAR models in Functional Genomics technological advances allow collecting huge amount of data DNA microarrays, RNA-sequencing, mass spectrometry capture meaningful biological patterns via network modeling difficult to infer direction of influence from co-expression transition patterns in time course data helps identify regulatory mechanisms George Michailidis (UM) High-dimensional VAR 6 / 47

7 VAR models in Functional Genomics (ctd) HeLa gene expression regulatory network [Courtesy: Fujita et al., 2007] George Michailidis (UM) High-dimensional VAR 7 / 47

8 VAR models in Neuroscience identify connectivity among brain regions from time course fmri data connectivity of VAR generative model (Seth et al., 2013) George Michailidis (UM) High-dimensional VAR 8 / 47

9 Model p-dimensional, discrete time, stationary process X t = {X t 1,...,Xt p} X t = A 1 X t A d X t d + ε t, ε t i.i.d N(0,Σ ε ) (1) A 1,...,A d : p p transition matrices (solid, directed edges) Σ 1 ε : contemporaneous dependence (dotted, undirected edges) stability: Eigenvalues of A (z) := I p d t=1 A tz t outside {z C, z 1} George Michailidis (UM) High-dimensional VAR 9 / 47

10 Why high-dimensional VAR? The parameter space grows quadratically (p 2 edges for p time series) order of the process (d) often unknown Economics: Forecasting with many predictors (De Mol et al., 2008) Understanding structural relationship - price puzzle" (Christiano et al., 1999) Functional Genomics: reconstruct networks among hundreds to thousands of genes experiments costly - small to moderate sample size Finance: structural changes - local stationarity George Michailidis (UM) High-dimensional VAR 10 / 47

11 Literature on high-dimensional VAR models Economics: Bayesian vector autoregression (lasso, ridge penalty; Litterman, Minnesota Prior) Factor model based approach (FAVAR, dynamic factor models) Bioinformatics: Discovering gene regulatory mechanisms using pairwise VARs (Fujita et al., 2007 and Mukhopadhyay and Chatterjee, 2007) Penalized VAR with grouping effects over time (Lozano et al., 2009) Truncated lasso and thesholded lasso variants (Shojaie and Michailidis, 2010 and Shojaie, Basu and Michailidis, 2012) Statistics: lasso (Han and Liu, 2013) and group lasso penalty (Song and Bickel, 2011) low-rank modeling with nuclear norm penalty (Negahban and Wainwright, 2011) sparse VAR modeling via two-stage procedures (Davis et al., 2012) George Michailidis (UM) High-dimensional VAR 11 / 47

12 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 12 / 47

13 Model p-dimensional, discrete time, stationary process X t = {X t 1,...,Xt p} X t = A 1 X t A d X t d + ε t, ε t i.i.d N(0,Σ ε ) (2) A 1,...,A d : p p transition matrices (solid, directed edges) Σ 1 ε : contemporaneous dependence (dotted, undirected edges) stability: Eigenvalues of A (z) := I p d t=1 A tz t outside {z C, z 1} George Michailidis (UM) High-dimensional VAR 13 / 47

14 Detour: VARs and Granger Causality Concept introduced by Granger (1969) A time series X is said to Granger-cause Y if it can be shown, usually through a series of F-tests on lagged values of X (and with lagged values of Y also known), that those X values provide statistically significant information about future values of Y. In the context of a high-dimensional VAR model we have that X T t j is Granger-causal for X T i if A t i,j 0. Granger-causality does not imply true causality; it is built on correlations Also, related to estimating a Directed Acyclic Graph (DAG) with (d + 1) p variables, with a known ordering of the variables George Michailidis (UM) High-dimensional VAR 14 / 47

15 Estimating VARs through regression data: {X 0,X 1,...,X T } - one replicate, observed at T + 1 time points construct autoregression (X T ) (X T 1 ) (X T 2 ) (X T d ) (X T 1 ) (X T 2 ) (X T 3 ) (X T 1 d ) = } (X d ) {{ } } (X d 1 ) (X d 2 ) {{ (X 0 ) } Y X A 1. A d } {{ } B + (ε T ) (ε T 1 ).. (ε d ) } {{ } E vec(y ) = vec(x B ) + vec(e) Y }{{} Np 1 = (I X )vec(b ) + vec(e) = }{{} Z Np q β }{{} q 1 +vec(e) }{{} Np 1 N = (T d + 1), q = dp 2 Assumption : A t are sparse, d t=1 A t 0 k vec(e) N (0,Σ ε I) George Michailidis (UM) High-dimensional VAR 15 / 47

16 Estimates l 1 -penalized least squares (l 1 -LS) 1 argmin β R q N Y Zβ 2 + λ N β 1 l 1 -penalized log-likelihood (l 1 -LL) (Davis et al., 2012) 1 argmin β R q N (Y ( Zβ) Σ 1 ε I ) (Y Zβ) + λ N β 1 George Michailidis (UM) High-dimensional VAR 16 / 47

17 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 17 / 47

18 Detour: Consistency of Lasso Regression Y n 1 LASSO : = X n p β 1 ˆβ := argmin β R p n Y Xβ 2 + λ n β 1 p 1 + ε { } S = j {1,...,p} βj i.i.d. 0, card(s) = k, k n, ε i N(0,σ 2 ) Restricted Eigenvalue (RE): Assume 1 α RE := min v R p, v 1, v S c 1 3 v S 1 n Xv 2 > 0 n 1 Estimation error: ˆβ β Q(X,σ) 1 k logp α RE n with high probability George Michailidis (UM) High-dimensional VAR 18 / 47

19 Verifying Restricted Eigenvalue Condition Raskutti et al. (2010): If the rows of X i.i.d. N(0,Σ X ) and Σ X satisfies RE, then X satisfies RE with high probability. Assumption of independence among rows crucial Rudelson and Zhou (2013): If the design matrix X can be factorized as X = ΨA where A satisfies RE and Ψ acts as (almost) an isometry on the images of sparse vectors under A, then X satisfies RE with high probability. George Michailidis (UM) High-dimensional VAR 19 / 47

20 Back to Vector Autoregression Random design matrix X, correlated with error matrix E (X T ) (X T 1 ) (X T 2 ) (X T d ) (X T 1 ) (X T 2 ) (X T 3 ) (X T 1 d ) A 1 = (X d ) (X d 1 ) (X d 2 ) (X 0 ) A d }{{}}{{}}{{} B Y X vec(y ) = vec(x B ) + vec(e) Y }{{} Np 1 = (I X )vec(b ) + vec(e) = }{{} Z Np q β }{{} q 1 +vec(e) }{{} Np 1 N = (T d + 1), q = dp 2 + vec(e) N (0,Σ ε I) (ε T ) (ε T 1 ). } (ε d ) {{ } E George Michailidis (UM) High-dimensional VAR 20 / 47

21 Vector Autoregression (ctd) Key Questions: How often does RE hold? How small is α RE? How does the cross-correlation affect convergence rates? George Michailidis (UM) High-dimensional VAR 21 / 47

22 Consistency of VAR estimates Restricted Eigenvalue (RE) assumption: (I X ) q q RE(α,τ(N,q)) with α > 0,τ(N,q) > 0 if θ ( I X X /N ) θ α θ 2 2 τ(n,q) θ 2 1 for all θ R q (3) Deviation Condition: There exists a function Q(β,Σ ε ) such that vec ( X E/N ) log d + 2 log p max Q(β,Σ ε ) N Key Result: Estimation Consistency: If (3) and (4) hold with kτ(n, q) α/32, then, for any λ N 4Q(β,Σ ε ) (log d + 2 log p)/n, lasso estimate ˆβ l1 satisfies ˆβ l1 β 64 Q(β,Σ ε ) k (log d + 2 log p) α N (4) George Michailidis (UM) High-dimensional VAR 22 / 47

23 Verifying RE and Deviation Condition Negahban and Wainwright, 2011: for VAR(1) models, assume A 1 < 1, where A := Λ max (A A) For p = 1, d = 1, X t = ρx t 1 + ε t, reduces to ρ < 1 - equivalent to stability Han and Liu, 2013: for VAR(d) models, reformulate as VAR(1): X t = à 1 X t 1 + ε t, where X t = X t X t 1. X t d+1 Assume à 1 < 1 dp 1 à 1 = A 1 A 2 A d 1 A d I p I p I p 0 dp dp ε t = ε t 0. 0 dp 1 George Michailidis (UM) High-dimensional VAR 23 / 47

24 VAR(1): Stability and A 1 < 1 A 1 1 for many stable VAR(1) models [ X t = A 1 X t 1 + ε t α 0, A 1 = β α ] (α, β): {X t } stable 5 β α t 2 X 1 t 1 X 1 α t X β α t 2 t 1 t X 2 X 2 X 2 (α, β): A 1 < 1 5 George Michailidis (UM) High-dimensional VAR 24 / 47

25 VAR(d): Stability and à 1 < 1 à 1 1 for any stable VAR(d) models, if d > 1 [ X t = 2αX t 1 α 2 X t 2 + ε t X t, X t 1 ] [ 2α α 2 = 1 0 ][ X t 1 X t 2 ] [ ε t + 0 ] 3 ~ A 1 α 2 2α X t 2 X t 1 X t 2 ~ A 1 = 1 0 α George Michailidis (UM) High-dimensional VAR 25 / 47

26 Stable VAR models V VAR(2), VAR(3),... VAR(d),... VAR(1) A < 1 George Michailidis (UM) High-dimensional VAR 26 / 47

27 Stable VAR models V VAR(2), VAR(3),... VAR(d),... VAR(1) A < 1 George Michailidis (UM) High-dimensional VAR 27 / 47

28 Quantifying Stability through the Spectral Density Spectral density function of a covariance stationary process {X t }, f X (θ) = 1 2π Γ X (l)e ilθ, l= θ [ π,π] Γ X (l) = E [ X t (X t+l ) ], autocovariance matrix of order l If the VAR process is stable, it has a closed form (Priestley, 1981) f X (θ) = 1 ( ) 1 ( ) 1 A (e iθ ) Σε A (e iθ ) 2π The two sources of dependence factorize in frequency domain George Michailidis (UM) High-dimensional VAR 28 / 47

29 Quantifying Stability by Spectral Density Quantifying Stability by Spectral Density For univariate processes, peak" of the spectral density measures stability of the For process univariate - (sharper processes, peak the = less peak" stable) of the spectral density measures stability of the process - (sharper peak = less stable) Autocovariance Γ(h) ρ=0.1 ρ=0.5 ρ=0.7 f(θ) ρ=0.1 ρ=0.5 ρ= lag (h) θ (f) Autocovariance of AR(1) (g) Spectral Density of AR(1) For multivariate processes, similar role is played by the maximum eigenvalue of For multivariate processes, similar role is played by the maximum eigenvalue of the (matrix-valued) spectral density the (matrix-valued) spectral density Sumanta Basu (UM) Network Estimation in Time Series 52 / 56 George Michailidis (UM) High-dimensional VAR 29 / 47

30 Quantifying Stability by Spectral Density For a stable VAR(d) process {X t }, the maximum eigenvalue of its spectral density captures its stability M (f X ) = max Λ max (f X (θ)) θ [ π,π] The minimum eigenvalue of the spectral density captures dependence among its components m(f X ) = min Λ min (f X (θ)) θ [ π,π] For stable VAR(1) processes, M (f X ) scales with (1 ρ(a 1 )) 2, ρ(a 1 ) is the spectral radius of A 1 m(f X ) scales with the capacity (maximum incoming + outgoing effect at a node) of the underlying graph George Michailidis (UM) High-dimensional VAR 30 / 47

31 Consistency of VAR estimates Theorem Consider a random realization {X 0,...,X T } generated according to a stable VAR(d) process with Λ min (Σ ε ) > 0. Then there exist deterministic functions φ i (A t,σ ε ) > 0 and constants c i > 0 such that for N φ 0 (A t,σ ε ) k(logd + 2logp)/N, the lasso estimate (l 1 -LS) with λ N (2logp + logd)/n satisfies, with probability at least 1 c 1 exp[ c 2 (2logp + logd)], d ( ) Aˆ h A h φ1 (A t,σ ε ) k(logd + 2logp)/N h=1 T 1 d ( ) N ( h A h )X h φ 2 (A t,σ ε ) k(logd + 2logp)/N t=d h=1 ) Further, a thresholded version of lasso à = (Ât,ij 1 {  t,ij > λ N } satisfies supp(ã 1:d )\supp(a 1:d ) φ 3 (A t,σ ε )k φ i (A t,σ ε ) are large when M (f X ) is large and m(f X ) is small. George Michailidis (UM) High-dimensional VAR 31 / 47

32 Some Remarks Convergence rates governed by: dimensionality parameters - dimension of the process (p), order of the process (d), number of parameters (k) in the transition matrices A i and sample size (N = T -d + 1) internal parameters - curvature (α), tolerance (τ) and the deviation bound Q(β,Σ ε ) The squared l 2 -errors of estimation and prediction scale with the dimensionality parameters as k(2logp + logd)/n, similar to the rates obtained when the observations are independent The temporal and cross-sectional dependence affect the rates only through the internal parameters. Typically, the rates are better when α is large and Q(β,Σ ε ),τ are small. This dependence is captured in the next results. George Michailidis (UM) High-dimensional VAR 32 / 47

33 Verifying RE Proposition Consider a random realization {X 0,...,X T } generated according to a stable VAR(d) process. Then there exist universal positive constants c i such that for all N max{1,ω 2 }k log(dp), with probability at least 1 c 1 exp( c 2 N min{ω 2,1}), where I p (X X /N) RE(α,τ), ω = Λ min(σ ε )/Λ max (Σ ε ) µ max (A )/µ min ( A ), α = Λ min(σ ε ) 2µ max (A ), Λ min (Σ ε ) τ(n,q) = c 3 µ max (A ) max{ω 2,1} log(dp). N George Michailidis (UM) High-dimensional VAR 33 / 47

34 Verifying Deviation Condition Proposition If q 2, then, for any A > 0, N logd + 2logp, with probability at least 1 12q A, we have vec ( X E/N ) logd + 2logp max Q(β,Σ ε ), N where Q(β,Σ ε ) = ( (A + 1)) [ Λ max (Σ ε ) + Λ max(σ ε ) µ min (A ) + Λ ] max(σ ε )µ max (A ) µ min (A ) George Michailidis (UM) High-dimensional VAR 34 / 47

35 Some Comments RE: the convergence rates are faster for larger α and smaller τ. From the expressions of ω,α and τ, it is clear that the VAR estimates have lower error bounds when Λ max (Σ ε ), µ max (A ) are smaller and Λ min (Σ ε ), µ min (A ) are larger. Deviation bound: VAR estimates exhibit lower error bounds when Λ max (Σ ε ), µ max (A ) are smaller and Λ min (Σ ε ), µ min (A ) are larger (similar to RE) George Michailidis (UM) High-dimensional VAR 35 / 47

36 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 36 / 47

37 l 1 -LS: Denote the i th column of a matrix M by M i. arg 1 min β R q N Y Zβ 2 + λ N β 1 1 arg min B 1,...,B p N p i X B i i=1 Y 2 p + λ N B i 1 i=1 Amounts to running p separate LASSO programs, each with dp predictors: Y i X, i = 1,...,p. George Michailidis (UM) High-dimensional VAR 37 / 47

38 l 1 -LL: Davis et al, 2012, proposed the following algorithm: arg 1 arg min β R q N 1 min β R q N (Y ( Zβ) Σ 1 ( Σ 1/2 ε ε I ) (Y Zβ) + λ N β 1 ) Σ 1/2 ε X β 2 + λ N β 1 ) ( I Y ( Amounts to ) running a single LASSO program with dp 2 predictors: Σ 1/2 ε I Y Σ 1/2 ε X - cannot be implemented in parallel. σ ij ε := (i,j) th entry of Σ 1. The objective function is 1 N p p i=1 j=1 ε σ ij ε (Y i X B i ) (Y j X B j ) + λ N p k=1 B k 1 George Michailidis (UM) High-dimensional VAR 38 / 47

39 Block Coordinate Descent for l 1 -LL 1 pre-select d. Run l 1 -LS to get ˆB, ˆΣ 1 ε. 2 iterate till convergence: 1 For i = 1,...,p, set r i := (1/2 ˆσ ii ε ) j i update ˆB i = argmin B i ˆσ ε ij ( ) Yj X ˆB j ˆσ ii ε N (Y i + r i ) X B i 2 + λ N B i 1 each iteration amounts to running p separate LASSO programs, each with dp predictors: Y i + r i X, i = 1,...,p. Can be implemented in parallel George Michailidis (UM) High-dimensional VAR 39 / 47

40 Outline 1 Introduction 2 Modeling Framework 3 Theoretical Considerations 4 Implementation 5 Performance Evaluation George Michailidis (UM) High-dimensional VAR 40 / 47

41 VAR models considered Small Size VAR, p = 10,d = 1,T = 30,50 Medium Size VAR, p = 30,d = 1,T = 80,120,160 In each setting, we generate an adjacency matrix A 1 with 5 10% non-zero edges selected at random and rescale to ensure that the process is stable with SNR = 2. We generate three different error processes with covariance matrix Σ ε from one of the following families: 1 Block-I: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ii = 1, σ ε,ij = ρ if 1 i j p/2, 0 otherwise; 2 Block-II: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ii = 1, σ ε,ij = ρ if 1 i j p/2 or p/2 < i j p, 0 otherwise; 3 bf Toeplitz: Σ ε = ((σ ε,ij )) 1 i,j p with σ ε,ij = ρ i j. George Michailidis (UM) High-dimensional VAR 41 / 47

42 VAR models considered (ctd) REGUALARIZED ESTIMATION IN TIME SERIES (a) A 1 (b) Σ ɛ : Block-I (c) Σ ɛ : Block-II (d) Σ ɛ : Toeplitz We let ρ vary in {0.5,0.7,0.9}. : Adjacency matrix A 1 and error covariance matrix Σ ɛ of different ty Larger values of ρ indicate that the error processes are more strongly in the simulation studies correlated. Block-I: Σ ɛ = ((σ ɛ,ij )) 1 i,j p with σ ɛ,ii = 1, σ ɛ,ij = ρ if 1 i j George Michailidis (UM) High-dimensional VAR 42 / 47

43 Comparisons and Performance Criteria Different methods for VAR estimation: OLS l 1 -LS l 1 -LL l 1 -LL-O (Oracle version, assuming Σ ε known) Ridge evaluated using the following performance metrics: 1 Model Selection: Area under receiving operator characteristic curve (AUROC) 2 Estimation error: Relative estimation accuracy measured by ˆB B F / B F George Michailidis (UM) High-dimensional VAR 43 / 47

44 Results I Table: VAR(1) model with p = 10, T = 30 BLOCK-I BLOCK-II Toeplitz ρ AUROC l 1 -LS l 1 -LL l 1 -LL-O Estimation OLS Error l 1 -LS l 1 -LL l 1 -LL-O ridge George Michailidis (UM) High-dimensional VAR 44 / 47

45 Results II Table: VAR(1) model with p = 30, T = 120 BLOCK-I BLOCK-II Toeplitz ρ AUROC l 1 -LS l 1 -LL l 1 -LL-O Estimation OLS Error l 1 -LS l 1 -LL l 1 -LL-O Ridge George Michailidis (UM) High-dimensional VAR 45 / 47

46 Summary/Discussion Investigated penalized VAR estimation in high-dimension Established estimation consistency for all stable VAR models, based on novel techniques using spectral representation of stationary processes Developed parallellizable algorithm for likelihood based VAR estimates There is extensive work on characterizing univariate time series, through mixing conditions or functional dependence measures. However, thre is little work for multivariate series, which is needed to be able to provide results in the current setting. George Michailidis (UM) High-dimensional VAR 46 / 47

47 References S. Basu and G. Michailidis, Estimation in High-dimensional Vector Autoregressive Models, arxiv: S. Basu, A. Shojaie and G. Michailidis, Network Granger Causality with Inherent Grouping Structure, revised for JMLR George Michailidis (UM) High-dimensional VAR 47 / 47

REGULARIZED ESTIMATION IN SPARSE HIGH-DIMENSIONAL TIME SERIES MODELS. BY SUMANTA BASU AND GEORGE MICHAILIDIS 1 University of Michigan

REGULARIZED ESTIMATION IN SPARSE HIGH-DIMENSIONAL TIME SERIES MODELS. BY SUMANTA BASU AND GEORGE MICHAILIDIS 1 University of Michigan The Annals of Statistics 2015, Vol. 43, No. 4, 1535 1567 DOI: 10.1214/15-AOS1315 Institute of Mathematical Statistics, 2015 REGULARIZED ESTIMATION IN SPARSE HIGH-DIMENSIONAL TIME SERIES MODELS BY SUMANTA

More information

High-dimensional covariance estimation based on Gaussian graphical models

High-dimensional covariance estimation based on Gaussian graphical models High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,

More information

arxiv: v1 [stat.me] 28 Jul 2017

arxiv: v1 [stat.me] 28 Jul 2017 Sparse Identification and Estimation of High-Dimensional Vector AutoRegressive Moving Averages Ines Wilms a,b, Sumanta Basu a, Jacob Bien c, and David S. Matteson a arxiv:707.09208v [stat.me] 28 Jul 207

More information

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage Lingrui Gan, Naveen N. Narisetty, Feng Liang Department of Statistics University of Illinois at Urbana-Champaign Problem Statement

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large

More information

Regularized Estimation of High-dimensional Factor-Augmented Vector Autoregressive (FAVAR) Models

Regularized Estimation of High-dimensional Factor-Augmented Vector Autoregressive (FAVAR) Models Regularized Estimation of High-dimensional Factor-Augmented Vector Autoregressive (FAVAR) Models George Michailidis University of Florida Department of Statistics and the Informatics Institute February

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Lasso Guarantees for Dependent Data

Lasso Guarantees for Dependent Data Lasso Guarantees for Dependent Data by Kam Chung Wong A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy (Statistics) in The University of Michigan

More information

Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions

Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions Yao Zheng collaboration with Guang Cheng Department of Statistics, Purdue University Talk at Math, IUPUI October 23, 2018

More information

Learning discrete graphical models via generalized inverse covariance matrices

Learning discrete graphical models via generalized inverse covariance matrices Learning discrete graphical models via generalized inverse covariance matrices Duzhe Wang, Yiming Lv, Yongjoon Kim, Young Lee Department of Statistics University of Wisconsin-Madison {dwang282, lv23, ykim676,

More information

Estimating Structured Vector Autoregressive Models

Estimating Structured Vector Autoregressive Models Estimating Structured Vector Autoregressive Models Igor Melnyk Arindam Banerjee Department of Computer Science and Engineering, University of Minnesota, Twin Cities MELYK@CSUMEDU BAERJEE@CSUMEDU Abstract

More information

Visualizing VAR s: Regularization and Network Tools for High-Dimensional Financial Econometrics

Visualizing VAR s: Regularization and Network Tools for High-Dimensional Financial Econometrics Visualizing VAR s: Regularization and Network Tools for High-Dimensional Financial Econometrics Francis X. Diebold University of Pennsylvania March 7, 2015 1 / 32 DGP: N-Variable VAR(p), t = 1,..., T Φ(L)x

More information

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I. Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series

More information

An algorithm for the multivariate group lasso with covariance estimation

An algorithm for the multivariate group lasso with covariance estimation An algorithm for the multivariate group lasso with covariance estimation arxiv:1512.05153v1 [stat.co] 16 Dec 2015 Ines Wilms and Christophe Croux Leuven Statistics Research Centre, KU Leuven, Belgium Abstract

More information

High-dimensional Statistics

High-dimensional Statistics High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given

More information

Frontiers in Forecasting, Minneapolis February 21-23, Sparse VAR-Models. Christophe Croux. EDHEC Business School (France)

Frontiers in Forecasting, Minneapolis February 21-23, Sparse VAR-Models. Christophe Croux. EDHEC Business School (France) Frontiers in Forecasting, Minneapolis February 21-23, 2018 Sparse VAR-Models Christophe Croux EDHEC Business School (France) Joint Work with Ines Wilms (Cornell University), Luca Barbaglia (KU leuven),

More information

Estimators based on non-convex programs: Statistical and computational guarantees

Estimators based on non-convex programs: Statistical and computational guarantees Estimators based on non-convex programs: Statistical and computational guarantees Martin Wainwright UC Berkeley Statistics and EECS Based on joint work with: Po-Ling Loh (UC Berkeley) Martin Wainwright

More information

Financial Econometrics

Financial Econometrics Financial Econometrics Multivariate Time Series Analysis: VAR Gerald P. Dwyer Trinity College, Dublin January 2013 GPD (TCD) VAR 01/13 1 / 25 Structural equations Suppose have simultaneous system for supply

More information

11 : Gaussian Graphic Models and Ising Models

11 : Gaussian Graphic Models and Ising Models 10-708: Probabilistic Graphical Models 10-708, Spring 2017 11 : Gaussian Graphic Models and Ising Models Lecturer: Bryon Aragam Scribes: Chao-Ming Yen 1 Introduction Different from previous maximum likelihood

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Factor models. March 13, 2017

Factor models. March 13, 2017 Factor models March 13, 2017 Factor Models Macro economists have a peculiar data situation: Many data series, but usually short samples How can we utilize all this information without running into degrees

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

Penalization method for sparse time series model

Penalization method for sparse time series model Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS020) p.6280 Penalization method for sparse time series model Heewon, Park Chuo University, Department of Mathemetics

More information

New Introduction to Multiple Time Series Analysis

New Introduction to Multiple Time Series Analysis Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Adaptive Thresholding for Reconstructing Regulatory Networks from Time Course Gene Expression Data

Adaptive Thresholding for Reconstructing Regulatory Networks from Time Course Gene Expression Data Noname manuscript No. (will be inserted by the editor) Adaptive Thresholding for Reconstructing Regulatory Networks from Time Course Gene Expression Data Ali Shojaie Sumanta Basu George Michailidis Received:

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2013 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Meghana Kshirsagar (mkshirsa), Yiwen Chen (yiwenche) 1 Graph

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

High-dimensional Statistical Models

High-dimensional Statistical Models High-dimensional Statistical Models Pradeep Ravikumar UT Austin MLSS 2014 1 Curse of Dimensionality Statistical Learning: Given n observations from p(x; θ ), where θ R p, recover signal/parameter θ. For

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Inference in VARs with Conditional Heteroskedasticity of Unknown Form Inference in VARs with Conditional Heteroskedasticity of Unknown Form Ralf Brüggemann a Carsten Jentsch b Carsten Trenkler c University of Konstanz University of Mannheim University of Mannheim IAB Nuremberg

More information

2.5 Forecasting and Impulse Response Functions

2.5 Forecasting and Impulse Response Functions 2.5 Forecasting and Impulse Response Functions Principles of forecasting Forecast based on conditional expectations Suppose we are interested in forecasting the value of y t+1 based on a set of variables

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

Within Group Variable Selection through the Exclusive Lasso

Within Group Variable Selection through the Exclusive Lasso Within Group Variable Selection through the Exclusive Lasso arxiv:1505.07517v1 [stat.me] 28 May 2015 Frederick Campbell Department of Statistics, Rice University and Genevera Allen Department of Statistics,

More information

Regularized Estimation of High Dimensional Covariance Matrices. Peter Bickel. January, 2008

Regularized Estimation of High Dimensional Covariance Matrices. Peter Bickel. January, 2008 Regularized Estimation of High Dimensional Covariance Matrices Peter Bickel Cambridge January, 2008 With Thanks to E. Levina (Joint collaboration, slides) I. M. Johnstone (Slides) Choongsoon Bae (Slides)

More information

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los

More information

Bayesian Compressed Vector Autoregressions

Bayesian Compressed Vector Autoregressions Bayesian Compressed Vector Autoregressions Gary Koop a, Dimitris Korobilis b, and Davide Pettenuzzo c a University of Strathclyde b University of Glasgow c Brandeis University 9th ECB Workshop on Forecasting

More information

Factor models. May 11, 2012

Factor models. May 11, 2012 Factor models May 11, 2012 Factor Models Macro economists have a peculiar data situation: Many data series, but usually short samples How can we utilize all this information without running into degrees

More information

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively

More information

Log Covariance Matrix Estimation

Log Covariance Matrix Estimation Log Covariance Matrix Estimation Xinwei Deng Department of Statistics University of Wisconsin-Madison Joint work with Kam-Wah Tsui (Univ. of Wisconsin-Madsion) 1 Outline Background and Motivation The Proposed

More information

High Dimensional Covariance and Precision Matrix Estimation

High Dimensional Covariance and Precision Matrix Estimation High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance

More information

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression

The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression The Sparsity and Bias of The LASSO Selection In High-Dimensional Linear Regression Cun-hui Zhang and Jian Huang Presenter: Quefeng Li Feb. 26, 2010 un-hui Zhang and Jian Huang Presenter: Quefeng The Sparsity

More information

Bayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1

Bayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1 Bayesian Graphical Models for Structural Vector Autoregressive Processes Daniel Ahelegbey, Monica Billio, and Roberto Cassin (2014) March 21, 2015 Bayesian Graphical Models for Structural Vector AutoregressiveMarch

More information

Sparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Sparse regression. Optimization-Based Data Analysis.   Carlos Fernandez-Granda Sparse regression Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 3/28/2016 Regression Least-squares regression Example: Global warming Logistic

More information

ESL Chap3. Some extensions of lasso

ESL Chap3. Some extensions of lasso ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied

More information

Variable Targeting and Reduction in High-Dimensional Vector Autoregressions

Variable Targeting and Reduction in High-Dimensional Vector Autoregressions Variable Targeting and Reduction in High-Dimensional Vector Autoregressions Tucker McElroy (U.S. Census Bureau) Frontiers in Forecasting February 21-23, 2018 1 / 22 Disclaimer This presentation is released

More information

Switching Regime Estimation

Switching Regime Estimation Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms

More information

Graphical Model Selection

Graphical Model Selection May 6, 2013 Trevor Hastie, Stanford Statistics 1 Graphical Model Selection Trevor Hastie Stanford University joint work with Jerome Friedman, Rob Tibshirani, Rahul Mazumder and Jason Lee May 6, 2013 Trevor

More information

Sparse inverse covariance estimation with the lasso

Sparse inverse covariance estimation with the lasso Sparse inverse covariance estimation with the lasso Jerome Friedman Trevor Hastie and Robert Tibshirani November 8, 2007 Abstract We consider the problem of estimating sparse graphs by a lasso penalty

More information

Inference For High Dimensional M-estimates: Fixed Design Results

Inference For High Dimensional M-estimates: Fixed Design Results Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49

More information

arxiv: v1 [math.st] 13 Feb 2012

arxiv: v1 [math.st] 13 Feb 2012 Sparse Matrix Inversion with Scaled Lasso Tingni Sun and Cun-Hui Zhang Rutgers University arxiv:1202.2723v1 [math.st] 13 Feb 2012 Address: Department of Statistics and Biostatistics, Hill Center, Busch

More information

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach

More information

Title. Description. var intro Introduction to vector autoregressive models

Title. Description. var intro Introduction to vector autoregressive models Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models

More information

Factor-Adjusted Robust Multiple Test. Jianqing Fan (Princeton University)

Factor-Adjusted Robust Multiple Test. Jianqing Fan (Princeton University) Factor-Adjusted Robust Multiple Test Jianqing Fan Princeton University with Koushiki Bose, Qiang Sun, Wenxin Zhou August 11, 2017 Outline 1 Introduction 2 A principle of robustification 3 Adaptive Huber

More information

High-dimensional statistics: Some progress and challenges ahead

High-dimensional statistics: Some progress and challenges ahead High-dimensional statistics: Some progress and challenges ahead Martin Wainwright UC Berkeley Departments of Statistics, and EECS University College, London Master Class: Lecture Joint work with: Alekh

More information

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Confidence Intervals for Low-dimensional Parameters with High-dimensional Data Cun-Hui Zhang and Stephanie S. Zhang Rutgers University and Columbia University September 14, 2012 Outline Introduction Methodology

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Generalized Concomitant Multi-Task Lasso for sparse multimodal regression

Generalized Concomitant Multi-Task Lasso for sparse multimodal regression Generalized Concomitant Multi-Task Lasso for sparse multimodal regression Mathurin Massias https://mathurinm.github.io INRIA Saclay Joint work with: Olivier Fercoq (Télécom ParisTech) Alexandre Gramfort

More information

Towards a Regression using Tensors

Towards a Regression using Tensors February 27, 2014 Outline Background 1 Background Linear Regression Tensorial Data Analysis 2 Definition Tensor Operation Tensor Decomposition 3 Model Attention Deficit Hyperactivity Disorder Data Analysis

More information

Inference when identifying assumptions are doubted. A. Theory. Structural model of interest: B 1 y t1. u t. B m y tm. u t i.i.d.

Inference when identifying assumptions are doubted. A. Theory. Structural model of interest: B 1 y t1. u t. B m y tm. u t i.i.d. Inference when identifying assumptions are doubted A. Theory B. Applications Structural model of interest: A y t B y t B m y tm nn n i.i.d. N, D D diagonal A. Theory Bayesian approach: Summarize whatever

More information

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding Patrick Breheny February 15 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/24 Introduction Last week, we introduced penalized regression and discussed ridge regression, in which the penalty

More information

Genetic Networks. Korbinian Strimmer. Seminar: Statistical Analysis of RNA-Seq Data 19 June IMISE, Universität Leipzig

Genetic Networks. Korbinian Strimmer. Seminar: Statistical Analysis of RNA-Seq Data 19 June IMISE, Universität Leipzig Genetic Networks Korbinian Strimmer IMISE, Universität Leipzig Seminar: Statistical Analysis of RNA-Seq Data 19 June 2012 Korbinian Strimmer, RNA-Seq Networks, 19/6/2012 1 Paper G. I. Allen and Z. Liu.

More information

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation Adam J. Rothman School of Statistics University of Minnesota October 8, 2014, joint work with Liliana

More information

Regression with time series

Regression with time series Regression with time series Class Notes Manuel Arellano February 22, 2018 1 Classical regression model with time series Model and assumptions The basic assumption is E y t x 1,, x T = E y t x t = x tβ

More information

Cointegration Lecture I: Introduction

Cointegration Lecture I: Introduction 1 Cointegration Lecture I: Introduction Julia Giese Nuffield College julia.giese@economics.ox.ac.uk Hilary Term 2008 2 Outline Introduction Estimation of unrestricted VAR Non-stationarity Deterministic

More information

An iterative hard thresholding estimator for low rank matrix recovery

An iterative hard thresholding estimator for low rank matrix recovery An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical

More information

Combining Macroeconomic Models for Prediction

Combining Macroeconomic Models for Prediction Combining Macroeconomic Models for Prediction John Geweke University of Technology Sydney 15th Australasian Macro Workshop April 8, 2010 Outline 1 Optimal prediction pools 2 Models and data 3 Optimal pools

More information

You must continuously work on this project over the course of four weeks.

You must continuously work on this project over the course of four weeks. The project Five project topics are described below. You should choose one the projects. Maximum of two people per project is allowed. If two people are working on a topic they are expected to do double

More information

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss arxiv:1811.04545v1 [stat.co] 12 Nov 2018 Cheng Wang School of Mathematical Sciences, Shanghai Jiao

More information

High-dimensional Covariance Estimation Based On Gaussian Graphical Models

High-dimensional Covariance Estimation Based On Gaussian Graphical Models High-dimensional Covariance Estimation Based On Gaussian Graphical Models Shuheng Zhou, Philipp Rutimann, Min Xu and Peter Buhlmann February 3, 2012 Problem definition Want to estimate the covariance matrix

More information

Autoregressive distributed lag models

Autoregressive distributed lag models Introduction In economics, most cases we want to model relationships between variables, and often simultaneously. That means we need to move from univariate time series to multivariate. We do it in two

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Chapter 17: Undirected Graphical Models

Chapter 17: Undirected Graphical Models Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)

More information

Composite Loss Functions and Multivariate Regression; Sparse PCA

Composite Loss Functions and Multivariate Regression; Sparse PCA Composite Loss Functions and Multivariate Regression; Sparse PCA G. Obozinski, B. Taskar, and M. I. Jordan (2009). Joint covariate selection and joint subspace selection for multiple classification problems.

More information

Vector Autoregressive based Network Models

Vector Autoregressive based Network Models Vector Autoregressive based Network Models George Michailidis Informatics Institute University of Florida Dynamic Networks Workshop Isaac Newton Institute Cambridge, December 2016 George Michailidis VAR

More information

Penalized Estimation of Panel Vector Autoregressive Models: A Lasso Approach

Penalized Estimation of Panel Vector Autoregressive Models: A Lasso Approach Penalized Estimation of Panel Vector Autoregressive Models: A Lasso Approach Annika Schnücker Freie Universität Berlin and DIW Berlin Graduate Center, Mohrenstr. 58, 10117 Berlin, Germany October 11, 2017

More information

1. Fundamental concepts

1. Fundamental concepts . Fundamental concepts A time series is a sequence of data points, measured typically at successive times spaced at uniform intervals. Time series are used in such fields as statistics, signal processing

More information

Lecture 1: Information and Structural VARs

Lecture 1: Information and Structural VARs Lecture 1: Information and Structural VARs Luca Gambetti 1 1 Universitat Autònoma de Barcelona LBS, May 6-8 2013 Introduction The study of the dynamic effects of economic shocks is one of the key applications

More information

Single Equation Linear GMM with Serially Correlated Moment Conditions

Single Equation Linear GMM with Serially Correlated Moment Conditions Single Equation Linear GMM with Serially Correlated Moment Conditions Eric Zivot October 28, 2009 Univariate Time Series Let {y t } be an ergodic-stationary time series with E[y t ]=μ and var(y t )

More information

Introduction to Algorithmic Trading Strategies Lecture 3

Introduction to Algorithmic Trading Strategies Lecture 3 Introduction to Algorithmic Trading Strategies Lecture 3 Pairs Trading by Cointegration Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Distance method Cointegration Stationarity

More information

Sparse Learning and Distributed PCA. Jianqing Fan

Sparse Learning and Distributed PCA. Jianqing Fan w/ control of statistical errors and computing resources Jianqing Fan Princeton University Coauthors Han Liu Qiang Sun Tong Zhang Dong Wang Kaizheng Wang Ziwei Zhu Outline Computational Resources and Statistical

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

Inference when identifying assumptions are doubted. A. Theory B. Applications

Inference when identifying assumptions are doubted. A. Theory B. Applications Inference when identifying assumptions are doubted A. Theory B. Applications 1 A. Theory Structural model of interest: A y t B 1 y t1 B m y tm u t nn n1 u t i.i.d. N0, D D diagonal 2 Bayesian approach:

More information

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs Questions and Answers on Unit Roots, Cointegration, VARs and VECMs L. Magee Winter, 2012 1. Let ɛ t, t = 1,..., T be a series of independent draws from a N[0,1] distribution. Let w t, t = 1,..., T, be

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Permutation-invariant regularization of large covariance matrices. Liza Levina

Permutation-invariant regularization of large covariance matrices. Liza Levina Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work

More information

Robust high-dimensional linear regression: A statistical perspective

Robust high-dimensional linear regression: A statistical perspective Robust high-dimensional linear regression: A statistical perspective Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics STOC workshop on robustness and nonconvexity Montreal,

More information

X t = a t + r t, (7.1)

X t = a t + r t, (7.1) Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical

More information

IEOR165 Discussion Week 5

IEOR165 Discussion Week 5 IEOR165 Discussion Week 5 Sheng Liu University of California, Berkeley Feb 19, 2016 Outline 1 1st Homework 2 Revisit Maximum A Posterior 3 Regularization IEOR165 Discussion Sheng Liu 2 About 1st Homework

More information

Indirect multivariate response linear regression

Indirect multivariate response linear regression Biometrika (2016), xx, x, pp. 1 22 1 2 3 4 5 6 C 2007 Biometrika Trust Printed in Great Britain Indirect multivariate response linear regression BY AARON J. MOLSTAD AND ADAM J. ROTHMAN School of Statistics,

More information

7. MULTIVARATE STATIONARY PROCESSES

7. MULTIVARATE STATIONARY PROCESSES 7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability

More information

Estimation of large dimensional sparse covariance matrices

Estimation of large dimensional sparse covariance matrices Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)

More information

Lesson 17: Vector AutoRegressive Models

Lesson 17: Vector AutoRegressive Models Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Vector AutoRegressive models The extension of ARMA models into a multivariate framework

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information