Grouped Network Vector Autoregression

Size: px
Start display at page:

Download "Grouped Network Vector Autoregression"

Transcription

1 Statistica Sinica: Supplement Grouped Networ Vector Autoregression Xuening Zhu 1 and Rui Pan 2 1 Fudan University, 2 Central University of Finance and Economics Supplementary Material We present here the detailed technical proofs of Lemma 1 Lemma 4 in Appendi A. Net, the proofs of Theorem 1 to Theorem 3 are given in Appendi B, C, and D respectively. Appendi A. Four Useful Lemmas Lemma 1. Let X = (X 1,, X n) R n, where X is are independent and identically distributed random variables with mean zero, variance σ 2 X and finite fourth order moment. Let Ỹ t = j=0 Gj UE t j, where G R n n, U R n N, and {E t} satisfy Condition (C1) and are independent of {X i}. Then for a matri A = (a ij) R n n and a vector B = (b 1,, b n) R n, it holds that (a) n 1 B X p 0 if n 2 B B 0 as n. (b) n 1 X AX p σ 2 X lim n n 1 tr(a) if the limit eists, and n 2 tr(aa ) 0 as n. (c) (nt ) 1 T B Ỹ t p 0 if n 1 j=0 (B G j UU (G ) j B) 1/2 0 as n. (d) (nt ) 1 T Ỹ t AỸ t p lim n n 1 tr{aγ(0)} if the limit eists, and n 1 i=0 j=0 [tr{u (G ) i AG j UU (G ) j A G i U}] 1/2 0 as n.

2 (e) (nt ) 1 T X AỸ t p 0 if n 1 j=0 [tr{agj UU (G ) j A }] 1/2 0 as n. Proof: The detailed proof can be found in Lemma 1 of Zhu et al. (2017). Lemma 2. Assume min N = O(N δ ) and the stationary condition c β < 1, where c β = ma ( β 1 + β 2 ). Further assume Conditions (C1)-(C3) hold. For matrices M 1 = (m (1) ij ) R n p and M 2 = (m (2) ij ) Rn p, define M 1 M 2 as m (1) ij m (2) ij for 1 i n and 1 j p. In addition, define M e = ( m ij ) R n p for any arbitrary matri M = (m ij) R n p. Then there eists J > 0, such that (a) for any integer n > 0, we have G n (G ) n e n J c 2n β MM, (A.1) G n Σ Y e αn J c n βmm, (A.2) where M = C1π + J j=0 W j, C > 1 is a constant, π is defined in (C3.1), and α is a finite constant. (b) For positive integers 1 1, 2 1, and j 0, define g j,1, 2 (G, W () ) = (W () ) 1 {G j (G ) j } 2 (W () ) 1 e R N N. In addition, define (W () ) 0 = I = (I N, 0) R N N. For integers 0 1, 2, m 1, m 2 1, as N we have N 1 N 1 j=0 i,j=0 { µ g j,1, 2 (G, W () )µ} 1/2 0, (A.3) [ { 1/2 tr g i,1, 2 (G, W () )g j,m1,m 2 (G, W )}] () 0, (A.4) where µ e c µ1 and c µ is a finite constant. (c) For integers 0 1, 2 1, define f 1, 2 (W (), Q) = (W () ) 1 Q 2 (W () ) 1 e R N N,

3 where Q is given in (C3). Then for integers 0 1, 2, m 1, m 2 1, as N we have N 2 µ f 1, 2 (W (), Q)µ 0, { } N 2 tr f 1, 2 (W (), Q)f m1,m 2 (W (), Q) 0, N 1 j=0 (A.5) (A.6) [ { 1/2 tr f 1, 2 (W (), Q)g j,m1,m 2 (G, W )}] () 0, (A.7) where µ e c µ1 and c µ is a finite constant. Proof: The proof is similar in spirit to Zhu et al. (2017). Therefore, we give the guideline of the proof and sip some similar details. Without loss of generality, we let c β = β 11 + β 21 (i.e., = 1). Consequently, we have G e β 11 W + β 21 I. Let G = β 11 W + β 21 I. Follow similar technique in part (a) in Lemma 2 of Zhu et al. (2017), it can be verified G n e n J c n βm, (A.8) where M = C1π + J j=0 W j is defined in (a) of Lemma 2. Subsequently, the result (A.1) can be readily obtained. Net, recall that Σ Y = (I G) 1 Σ Z (I G ) 1 + j=0 Gj Σ e(g ) j = ( j=0 Gj )Σ Z ( j=0 (G ) j ) + j=0 Gj Σ e(g ) j. Let σ 2 z = ma {γ Σ zγ } and σ 2 e = ma {σ 2 }. Then we have G n Σ Y e σ 2 z( j=0 Gn+j e)( j=0 (G ) j e) + σ 2 e j=0 Gn+j e (G ) j e. Subsequently, (A.2) can by obtained by applying (A.8). Net, we give the proof of (b) in the following. The conclusion (c) can be proved by similar techniques, which is omitted here to save space. Let 1 = 2 = 1. Then we have g j,1,1(g, W () ) = W () G j G j W (). Recall that W () = (w ij : i M, 1 j N) R N N. Since we have µ e c µ1, then it suffices to show j=0 N { 1 1 g j,1,1(g, W () )1 } 1/2 0. We first prove (A.3). By (A.8) we have W () G j e

4 j K ( β 1 + β 2 ) j W () M. As a result, we have W () G j (G ) j W () e j 2K ( β 1 + β 2 ) 2j M, (A.9) where M is defined as M = W () MM W (). As a result, we have j=0 N { 1 1 W () G j (G ) j W () 1 } 1/2 N 1 α1(1 M1) 1/2, where α 1 = j=0 jk c j β <. Then it leads to show N 2 1 M1 0. It can be shown 1 M1 = N 2 C j π2 j + K j=1 1 W () W j (W ) j W () 1 + 2N C j π (W ) j W () 1+ i j 1 W () W i (W ) j W () 1. For the last two terms of 1 M1, by Cauchy inequality, we have ( N π (W ) j W () 1 N j j π 2 j ) 1/2 { 1 W () W j (W ) j W () 1} 1/2, and i j 1 W () W i (W ) j W () 1 i j W () 1 } 1/2. As a result, it leads to show { 1 W () W i (W ) i W () 1 } 1/2{ 1 W () W j (W ) j N j=1 π 2 j 0 and N 2 1 W () W j (W ) j W () 1 0 (A.10) for 1 j K + 1. As the first convergence in (A.10) is implied by (C2.1), we net prove N 2 1 W () W j (W ) j W () 1 0 (1 j K). Recall that W = W + W. Therefore, we have N 2 1 W () W j (W ) j W () 1 N 2 1 W () W 2j W () 1. Then it suffices to show N 2 1 W () W 2j W () 1 0. By eigenvalue-eigenvector decomposition of W we have W = λ (W )u u, where λ (W ) and u R N are the th eigenvalue and eigenvector of W respectively. As a result, we have N 2 1 W () W 2j W () 1 N 2 λma(w ) 2j (1 W () W () 1) (1 j K). Further we have 1 W () W () 1 N λ ma(w () W () ). Note that W () W ()

5 is a sub-matri of W W with row and column inde in M. Therefore, by Cauchy s interlacing Theorem, we have λ ma(w () W () ) λ ma(w W ) = O(N δ ) for δ < δ. Since we have min N = N δ for δ > 0, then we have N 1 λma(w () W () ) 0 as N. As a consequence, the second term in (A.10) holds. Similarly, it can be proved that (A.10) holds for all 0 1, 2 1. As a result, we have (A.3) holds. We net prove (A.4) with 1 = 2 = m 1 = m 2 = 1, and g i,1,1(g, W () )g j,1,1(g, W () ) = W () G i (G ) i W () W () G j (G ) j W () e. Then it can be similarly proved for other cases (i.e., 0 1, 2, m 1, m 2 1). Note that by (A.9), we have [ { tr W () G i (G ) i W () W () G j (G ) j W () }] 1/2 i K j K ( β 1 + β 2 ) i+j tr { M 2} 1/2. It then can be derived that N 1 i,j=0 [tr{w () G i (G ) i W () W () G j (G ) j W () }] 1/2 α 2 N 1 tr { M 2} 1/2. In order to obtain (A.4), it suffices to show that N 2 tr{m 2 } 0. (A.11) Equivalently, by Cauchy inequality, it suffices to prove ( πj 2 ) 2 0, and N 2 tr{w () W j W j W () W () W j W j W () } 0 holds for 1 j K. It can be easily verified the first term holds by (C2.1). For the second one, we have N 2 tr{w () W j W j W () W () W j W j W () } N 2 tr{w () (W ) 4j W () } N 2 λma(w ) 4j tr(w () W () ) N 2 N λ ma(w ) 4K λ ma(w W ). Similarly, due to that λ ma(w ) = O(log N) and λ ma(w W ) = O(N δ ) in (C2.2), we have N 1 λma(w ) 4K λ ma(w W ) 0 as N. Consequently, we have (A.11) and then (A.4) holds. This completes the proof of (b). Lemma 3. Let {X it : 1 t T } and {Y it : 1 t T } be random sub-gaussian time

6 series with mean 0, var(x it) = σ i,, var(y it) = σ i,yy, and cov(x it, Y it) = σ i,y. Let σ i,t1 t 2 = cov(x it1, X it2 ) and Σ i = (σ i,t1 t 2 : 1 t 1, t 2 T ) R T T. Similarly, define σ yi,t1 t 2 and Σ yi R T T. Then we have ( T 1 P T ) X ity it σ i,y > ν { } c 1 ep( c 2σ 2 i T 2 ν 2 ) + ep( c 2σ 2 yi T 2 ν 2 ) (A.12) for ν δ, where σ 2 i = tr(σ 2 i), σ 2 yi = tr(σ 2 yi), c 1, c 2, and δ are finite constants. Proof: Let X i = (X i1,, X it ) R T and Y i = (Y i1,, Y it ) R T. In addition, let Z i = Z i + Y i. Therefore, we have Zi Z i = 2 1 (Zi Z i Xi X i Yi Y i). It can be derived that P { T 1 (X i Y i) σ i,y ν} P { T 1 (Z i Z i) (σ i, + σ i,yy + 2σ i,y) ν 1} + P { T 1 (X i X i) σ i, ν 1} + P { T 1 (Y i Y i) σ i,yy ν 1}, (A.13) where ν 1 = 2ν/3. Net, we derive the upper bound for the right side of (A.13). Note that X i X i, Yi Y i, and Zi Z i all tae quadratic form. Therefore the proofs are similar. For the sae of simplicity, we tae Y i Y i for an eample and derive the upper bound for P { n 1 (Y Y i) σ i,yy ν 1}. Similar results can be obtained for the other two terms. First we have Yi Y i = Yi Σ 1/2 yi Σ yiσ 1/2 yi Y i = Ỹ Σ yi Ỹ i, where i i Ỹi = Σ 1/2 yi Y i follows sub- Gaussian distribution. Let λ 1 λ 2 λ T be the eigenvalues of Σ yi. Since Σ yi is a nonnegative definite matri, The eigenvalue decomposition can be applied to obtain Σ yi = U ΛU, where U = (U 1,, U T ) R T T is an orthogonal matri and Λ = diag{λ 1,, λ T }. As a consequence, we have Yt Y t = t λtζ2 t, where ζ t = Ut Ỹt and ζts are independent and identically distributed as standard sub-gaussian. It can be verified ζ 2 t 1 satisfies sub-eponential distribution and T 1 ( t λt) = σi,yy. In addition, the sub-eponential distribution satisfies condition

7 (P) on page 45 of Saulis and Statuleviveccius (2012). There eists constants c 1, c 2, and δ such that P { T 1 (Y i Y i) σ i,yy ν 1} = P { t λt(ζ2 t 1) T ν 1} c 1 ep{ c 2( t λ2 t ) 1 T 2 ν 2 } = c 1 ep{ c 2σ 1 yi T 2 ν 2 } for ν < δ by the Theorem 3.3 of Saulis and Statuleviveccius (2012). Consequently, (A.12) can be obtained by appropriately chosen c 1, c 2, and δ. Lemma 4. Assume Y it follows the GNAR model (2.4) and c β < 1. Then there eists finite constants c 1, c 2, and δ, for ν < δ we have P { T 1 P { T 1 P { T 1 P { T 1 T Y 2 it µ 2 i e i Σ Y e i > ν } δ T, T Y it(wi Y t) µ Y i(wi µ Y ) wi } Σ Y e i > ν δt T } Y i(t 1) ε it > ν δt, T } Y i(t 1) µ i > ν δt, P { T 1 (A.14) (A.15) T (wi } Y t 1)ε it > ν δt, (A.16) P { T 1 T wi Y t wi } µ Y > ν δt, (A.17) where δ T = c 1 ep( c 2T ν 2 ), e i R N is an N-dimensional vector with all elements being 0 but the ith element being 1, and µ i = e i µ Y. Proof: For the similarity of proof procedure, we only prove (A.14) in the following. Without loss of generality, let µ Y = 0. Recall that the group information is denoted as Z = {z i : 1 i N, 1 K}. Define P ( ) = P ( Z), E ( ) = E( Z), and cov ( ) = cov( Z). Write Y i = (Y i1,, Y it ) R T. Given Z, Y i is a sub-gaussian random vector with cov(y i) = Σ i = (σ i,t1 t 2 ) R T T, where σ i,t1 t 2 = e i G t 1 t 2 Σ Y e i for t 1 t 2, σ i,t1 t 2 = e i Σ Y (G ) t 2 t 1 e i, and G is pre-defined in (2.6) as G = B 1W + B 2. It can be derived var (Y i Y i) ctr(σ 2 i ), where c is a positive constant and tr(σ 2 i ) = T (e i Σ Y e i) T 1 (T t)(e i G t Σ Y e i) 2. It can be derived Σ Y e αmm and G t Σ Y e α 1t J c t βmm by (A.2) of Lemma 2, where c β, J and M are defined in Lemma 2, α and α 1 are finite constants. In addition, it can be

8 verified T 1 (T t)t2j c 2t β α 2T, where α 2 is a finite constant. Therefore we have tr(σ 2 i ) T (α + 2α 1α 2){(e i MM e i) 2 }. Since we have e i MM e i (J + 1)e i M1 (J + 1) 2 = O(1), it can be concluded that tr(σ 2 i ) T α 3, where α 3 = (α + 2α 1α 2)(J + 1) 2. By Lemma 3, the (A.14) can be obtained. Appendi B. Proof of Theorem 1 Let λ i(m) be the ith eigenvalue of M R N N. We first verify that the solution (2.7) is strictly stationary. By Banerjee et al. (2014), we have ma i λ i(w ) 1. Hence we have ma λi(g) ( ma β 1 )( ma λi(w ) ) + ma β 2 < 1. 1 i N 1 K 1 i N 1 K (A.18) Consequently, we have lim m m j=0 Gj E t j eists and {Y t} given by (2.7) is a strictly stationary process. In addition, one could directly verify that {Y t} satisfies the GNAR model (2.4). Net, we verify that the strictly stationary solution (2.7) is unique. Assume {Ỹt} is another strictly stationary solution to the GNAR model (2.4) with E Ỹt <. Then we have Ỹt = m 1 j=1 Gj (B 0 + E t j) + G m Ỹ t m for any positive integer m. Let ρ = ma ( β 1 + β 2 ). Then one could verify E Y t Ỹt = E j=m Gj (B 0 + E t j) G m Ỹ t m Cρ m, where C is a finite constant unrelated to t and m. Note that m can be chosen arbitrarily. As a result, we have that E Y t Ỹt = 0, i.e. Yt = Ỹt with probability one. This completes the proof. Appendi C. Proof of Theorem 2 According to (3.9), θ can be eplicitly written as θ = θ + Σ 1 ζ, where Σ = (N T ) 1 T X () t 1 X() t 1 and ζ = (N T ) 1 T X() t 1 E () t. Without loss of generality, we assume σ 2 = 1 for

9 = 1,, K. Let Σ = lim N E( Σ ). As a result, it suffices to show that Σ p Σ, N T ζ = O p(1), (A.19) (A.20) as min{n, T }. Subsequently, we prove (A.19) in Step 1 and (A.20) in Step 2. Step 1. Proof of (A.19). Define Q = (I G) 1 Σ V (I G ) 1. In this step, we intend to show that Σ = 1 N T T X () t 1 X() t 1 = 1 S 12 S 13 S 14 S 22 S 23 S 24 S 33 S 34 S 44 p 1 c 1β c 2β 0 Σ 1 Σ 2 κ 8γ Σ z Σ 3 κ 3γ Σ z Σ z = Σ, where S 12 = 1 N T T wi Y t 1, S 13 = 1 T Y i(t 1), S 14 = 1 N T N i M i M Vi, i M S 22 = 1 N T T (wi Y t 1) 2, S 23 = 1 T wi Y t 1Y i(t 1), N T i M i M S 24 = 1 N T T wi Y t 1Vi, S 33 = 1 N T i M T Yi(t 1), 2 i M S 34 = (N T ) 1 T i M Y i(t 1) V i, S 44 = N 1 i M V ivi. By (2.7), we have Y t = (I G) 1 b 0 + (I G) 1 b v + Ỹt, (A.21)

10 where b 0 = D B 0, b v = D Vγ, and Ỹt = j=0 Gj E t j. By the law of large numbers, one could directly obtain that S 44 p Σ v and S 14 p 0. Subsequently, we only show the convergence of S 12 and S 23 in Σ as follows. Convergence of S 12. It can be derived that S 12 = 1 N T T 1 W () Y t 1 = 1 W () µ Y N + S 12a + S 12b, where S 12a = N 1 1 W () (I G) 1 b v and S 12b = (N T ) 1 T 1 W () Ỹ t 1. Then by (A.5) and (A.3) in Lemma 2, we have N 2 1 W () QW () 1 0 and N 1 { j=0 1 W () G j (G ) j W () 1 } 1/2 0, as N. As a result, it is implied by Lemma 1 (a) and (c) that S 12a p 0 and S 12b p 0. Convergence of S 23. Note that S 23 = 1 N T T wi Y t 1Y i(t 1) = 1 T N T i M = µ() Y W () µ Y + S 23a + S 23b + S 23c + S 23d + S 23e, N Y () t 1 W () Y t 1 where S 23a = N 1 b v I W () bv, S 23b = N 1 T 1 T ( b v I W () Ỹ t 1 + Ỹ t 1I W () bv), S 23d = N 1 Ỹ() t 1 W () Ỹ t 1 and S 23c = N 1 T 1 T ( b v I µ Y + µ Y I b v), S 23e = N 1 T 1 T (Y () t 1 µy + µ Y I W () Y t 1), where µ Y = W () µ Y and b v = (I G) 1 b v. We net loo at the terms one by one. First we have N 2 tr(i QI W () QW () ) 0 by (A.6) in Lemma 2 (c). Therefore, by (b) in Lemma 1, we have S 23a p s 23a, where s 23a = lim N E(S 23a). Net, for S 23b we have N 1 i,j=0 tr{i G i (G ) i I W () G j (G ) j W () } 0 by (A.4) in Lemma 2 (b). Therefore, by (d) in Lemma 1, we have S 23b p s 23b, where s 23b = lim N E(S 23b ). Net, let S 23c = S (1) 23c +S(2) 23c, where S(1) 23c = N 1 T 1 T b z I W () Ỹ t 1 and

11 23c = N 1 T 1 T Ỹ t 1I W () bv. Note that we have N 1 j=0 tr{w () G j (G ) j W () I Q S (2) I } 0 and N 1 j=0 tr{i G j (G ) j I W () QW () } 0 by (A.7) in Lemma 2 (c). Therefore, S 23c p s 23c by (e) in Lemma 1, where s 23c = lim N E(S 23c). Net, by similar proof to the convergence of S 13, we have that S 23d p 0 and S 23e p 0. As a consequence, we have S 23 p Σ 2. Step 2. Proof of (A.20). It can be verified that N T E( ζ ) = 0. In addition, we have var{ N T ζ } = E( Σ ) Σ as N. Consequently, we have N T ζ = O p(1). Appendi D. Proof of Theorem 3 Let Σ (i) ε it). We then have = T 1 T X i(t 1)X i(t 1) = ( σ,ij) R 3 3 (i), and Σ e = T 1 ( T X i(t 1)δ i bi b i = ( Σ (i) ) 1 Σ (i) e. Let Σ (i) = ( σ,j1 j 2 : 1 l 1, l 2 3) R 3 3, where the inde i of σ,l1 l 2 is omitted. Specifically, σ,11 = 1, σ,12 = T 1 t w i Y t 1, σ,13 = T 1 t e i Y t 1, σ,22 = T 1 t Y 2 i(t 1), σ,23 = T 1 t Y i(t 1)(w i Y t 1), σ,33 = T 1 t (w i Y t 1) 2. Mathematically, it can be computed ( Σ (i) ) 1 = Σ (i) 1 Σ (i), where Σ (i) is the determinant of (i) (i) Σ, and Σ is the adjugate matri of (i) Σ, and Σ (i) = ( σ,l 1 l 2 ), where σ,11 = σ,22 σ,33 σ,23, 2 σ,12 = σ,13 σ,32 σ,12 σ,33 σ,13 = σ,21 σ,32 σ,22 σ,31, σ,22 = σ,11 σ,33 σ 2,13, σ,23 = σ,13 σ,32 σ,12 σ,33, and σ,33 = σ,11 σ,22 σ,12. 2 It can be derived Σ (i) = σ,11( σ,22 σ,33 σ,23) 2 σ,12( σ,12 σ 33 σ 13 σ 23) + σ 13( σ 12 σ 23 σ 22 σ 13). By the maimum inequality, we have P (sup b i b i > ν) i N P ( b i b i > ν). i=1 (A.22)

12 REFERENCES In addition, we have P ( b i b i > ν) P ( Σ(i) σ (i) ) ( δi + P Σ (i) Σ (i) e ) δiν, (A.23) where σ (i) = σ,11(σ,22σ,33 σ 2,23) σ,12(σ,12σ 33 σ 13σ 23) + σ 13(σ 12σ 23 σ 22σ 13) = (e i Σ Y e i)(w i Σ Y w i) (e i Σ Y w i) 2, δ i = σ (i) /2. By lemma 4, for each component of Σ (i) we have P ( σ,l1 l 2 σ,l1 l 2 > ν 0) c 1 ep( c 2T ν 2 0), where σ,l1 l 2 = E( σ,l1 l 2 ) and ν 0 is a finite positive constant. Moreover, by the conditions of Theorem 3, we have σ (i) tending to 1. Consequently, it is not difficult to obtain the result P ( Σ (i) τ with probability σ (i) δ i) c 1 ep( c 2T τ 2 ), where c 1, c 2 are finite constants. Subsequently, we have P ( Σ (i) Σ(i) e δ iν) P ( Σ (i) Σ(i) e τν/2). By similar technique, one could verify that each element of Σ (i) and Σ (i) e converge with probability and the tail probability can be controlled, where the basic results are given in Lemma 4. Consequently, there eists constants c 3 and c 4 such that P ( Σ (i) Σ(i) e τν/2) c 3 ep( c 4T τ 2 ν 2 ). Consequently, we have P ( b i b i > ν) c 1 ep( c 2T τ 2 ) + c 3 ep( c 4T τ 2 ν 2 ) by (A.23). By the condition N = o(ep(t )), the right side of (A.22) goes to 0 as N. This completes the proof. References Banerjee, S., Carlin, B. P., and Gelfand, A. E. (2014), Hierarchical modeling and analysis for spatial data, Crc Press. Saulis, L. and Statuleviveccius, V. (2012), Limit theorems for large deviations, vol. 73, Springer Science & Business Media. Zhu, X., Pan, R., Li, G., Liu, Y., and Wang, H. (2017), Networ vector autoregression, Annals

13 of statistics, 45, REFERENCES

ANOVA: Analysis of Variance - Part I

ANOVA: Analysis of Variance - Part I ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as MAHALANOBIS DISTANCE Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as d E (x, y) = (x 1 y 1 ) 2 + +(x p y p ) 2

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

Assumptions for the theoretical properties of the

Assumptions for the theoretical properties of the Statistica Sinica: Supplement IMPUTED FACTOR REGRESSION FOR HIGH- DIMENSIONAL BLOCK-WISE MISSING DATA Yanqing Zhang, Niansheng Tang and Annie Qu Yunnan University and University of Illinois at Urbana-Champaign

More information

Lectures 2 3 : Wigner s semicircle law

Lectures 2 3 : Wigner s semicircle law Fall 009 MATH 833 Random Matrices B. Való Lectures 3 : Wigner s semicircle law Notes prepared by: M. Koyama As we set up last wee, let M n = [X ij ] n i,j=1 be a symmetric n n matrix with Random entries

More information

7. MULTIVARATE STATIONARY PROCESSES

7. MULTIVARATE STATIONARY PROCESSES 7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability

More information

CONTROL CHARTS FOR MULTIVARIATE NONLINEAR TIME SERIES

CONTROL CHARTS FOR MULTIVARIATE NONLINEAR TIME SERIES REVSTAT Statistical Journal Volume 13, Number, June 015, 131 144 CONTROL CHARTS FOR MULTIVARIATE NONLINEAR TIME SERIES Authors: Robert Garthoff Department of Statistics, European University, Große Scharrnstr.

More information

Hyperbolic Systems of Conservation Laws. in One Space Dimension. II - Solutions to the Cauchy problem. Alberto Bressan

Hyperbolic Systems of Conservation Laws. in One Space Dimension. II - Solutions to the Cauchy problem. Alberto Bressan Hyperbolic Systems of Conservation Laws in One Space Dimension II - Solutions to the Cauchy problem Alberto Bressan Department of Mathematics, Penn State University http://www.math.psu.edu/bressan/ 1 Global

More information

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data Sri Lankan Journal of Applied Statistics (Special Issue) Modern Statistical Methodologies in the Cutting Edge of Science Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations

More information

5: MULTIVARATE STATIONARY PROCESSES

5: MULTIVARATE STATIONARY PROCESSES 5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability

More information

DYNAMIC AND COMPROMISE FACTOR ANALYSIS

DYNAMIC AND COMPROMISE FACTOR ANALYSIS DYNAMIC AND COMPROMISE FACTOR ANALYSIS Marianna Bolla Budapest University of Technology and Economics marib@math.bme.hu Many parts are joint work with Gy. Michaletzky, Loránd Eötvös University and G. Tusnády,

More information

On Expected Gaussian Random Determinants

On Expected Gaussian Random Determinants On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose

More information

A Short Course in Basic Statistics

A Short Course in Basic Statistics A Short Course in Basic Statistics Ian Schindler November 5, 2017 Creative commons license share and share alike BY: C 1 Descriptive Statistics 1.1 Presenting statistical data Definition 1 A statistical

More information

Nonlinear function on function additive model with multiple predictor curves. Xin Qi and Ruiyan Luo. Georgia State University

Nonlinear function on function additive model with multiple predictor curves. Xin Qi and Ruiyan Luo. Georgia State University Statistica Sinica: Supplement Nonlinear function on function additive model with multiple predictor curves Xin Qi and Ruiyan Luo Georgia State University Supplementary Material This supplementary material

More information

MIMO Capacities : Eigenvalue Computation through Representation Theory

MIMO Capacities : Eigenvalue Computation through Representation Theory MIMO Capacities : Eigenvalue Computation through Representation Theory Jayanta Kumar Pal, Donald Richards SAMSI Multivariate distributions working group Outline 1 Introduction 2 MIMO working model 3 Eigenvalue

More information

MATH 205C: STATIONARY PHASE LEMMA

MATH 205C: STATIONARY PHASE LEMMA MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)

More information

11. Regression and Least Squares

11. Regression and Least Squares 11. Regression and Least Squares Prof. Tesler Math 186 Winter 2016 Prof. Tesler Ch. 11: Linear Regression Math 186 / Winter 2016 1 / 23 Regression Given n points ( 1, 1 ), ( 2, 2 ),..., we want to determine

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Section 3.9. Matrix Norm

Section 3.9. Matrix Norm 3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix

More information

Lecture 3 - Expectation, inequalities and laws of large numbers

Lecture 3 - Expectation, inequalities and laws of large numbers Lecture 3 - Expectation, inequalities and laws of large numbers Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 3 - Expectation, inequalities and laws of large numbersapril 19, 2009 1 / 67 Part

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Systems of Algebraic Equations and Systems of Differential Equations

Systems of Algebraic Equations and Systems of Differential Equations Systems of Algebraic Equations and Systems of Differential Equations Topics: 2 by 2 systems of linear equations Matrix expression; Ax = b Solving 2 by 2 homogeneous systems Functions defined on matrices

More information

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 The test lasts 1 hour and 15 minutes. No documents are allowed. The use of a calculator, cell phone or other equivalent electronic

More information

Quadratic Programming

Quadratic Programming Quadratic Programming Quadratic programming is a special case of non-linear programming, and has many applications. One application is for optimal portfolio selection, which was developed by Markowitz

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

Given the vectors u, v, w and real numbers α, β, γ. Calculate vector a, which is equal to the linear combination α u + β v + γ w.

Given the vectors u, v, w and real numbers α, β, γ. Calculate vector a, which is equal to the linear combination α u + β v + γ w. Selected problems from the tetbook J. Neustupa, S. Kračmar: Sbírka příkladů z Matematiky I Problems in Mathematics I I. LINEAR ALGEBRA I.. Vectors, vector spaces Given the vectors u, v, w and real numbers

More information

Notes 6 : First and second moment methods

Notes 6 : First and second moment methods Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative

More information

A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report -

A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report - A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report - Edgar Brunner and Marius Placzek University of Göttingen, Germany 3. August 0 . Statistical Model and Hypotheses Throughout

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

There is a unique function s(x) that has the required properties. It turns out to also satisfy

There is a unique function s(x) that has the required properties. It turns out to also satisfy Numerical Analysis Grinshpan Natural Cubic Spline Let,, n be given nodes (strictly increasing) and let y,, y n be given values (arbitrary) Our goal is to produce a function s() with the following properties:

More information

MP 472 Quantum Information and Computation

MP 472 Quantum Information and Computation MP 472 Quantum Information and Computation http://www.thphys.may.ie/staff/jvala/mp472.htm Outline Open quantum systems The density operator ensemble of quantum states general properties the reduced density

More information

Lecture 7: Weak Duality

Lecture 7: Weak Duality EE 227A: Conve Optimization and Applications February 7, 2012 Lecture 7: Weak Duality Lecturer: Laurent El Ghaoui 7.1 Lagrange Dual problem 7.1.1 Primal problem In this section, we consider a possibly

More information

SUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley

SUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley SUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley A. Proofs A.1. Proof of Proposition.1 e have E T n (r k (X)) r? (X) = E T n (r k (X)) T(r

More information

arxiv: v5 [math.na] 16 Nov 2017

arxiv: v5 [math.na] 16 Nov 2017 RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential. Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample

More information

Hyperbolic Systems of Conservation Laws. I - Basic Concepts

Hyperbolic Systems of Conservation Laws. I - Basic Concepts Hyperbolic Systems of Conservation Laws I - Basic Concepts Alberto Bressan Mathematics Department, Penn State University Alberto Bressan (Penn State) Hyperbolic Systems of Conservation Laws 1 / 27 The

More information

Stein s Method and Characteristic Functions

Stein s Method and Characteristic Functions Stein s Method and Characteristic Functions Alexander Tikhomirov Komi Science Center of Ural Division of RAS, Syktyvkar, Russia; Singapore, NUS, 18-29 May 2015 Workshop New Directions in Stein s method

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

On rate of convergence in distribution of asymptotically normal statistics based on samples of random size

On rate of convergence in distribution of asymptotically normal statistics based on samples of random size Annales Mathematicae et Informaticae 39 212 pp. 17 28 Proceedings of the Conference on Stochastic Models and their Applications Faculty of Informatics, University of Debrecen, Debrecen, Hungary, August

More information

Common-Knowledge / Cheat Sheet

Common-Knowledge / Cheat Sheet CSE 521: Design and Analysis of Algorithms I Fall 2018 Common-Knowledge / Cheat Sheet 1 Randomized Algorithm Expectation: For a random variable X with domain, the discrete set S, E [X] = s S P [X = s]

More information

1 Kernel methods & optimization

1 Kernel methods & optimization Machine Learning Class Notes 9-26-13 Prof. David Sontag 1 Kernel methods & optimization One eample of a kernel that is frequently used in practice and which allows for highly non-linear discriminant functions

More information

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient

More information

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document

More information

Semicircle law on short scales and delocalization for Wigner random matrices

Semicircle law on short scales and delocalization for Wigner random matrices Semicircle law on short scales and delocalization for Wigner random matrices László Erdős University of Munich Weizmann Institute, December 2007 Joint work with H.T. Yau (Harvard), B. Schlein (Munich)

More information

2. Review of Linear Algebra

2. Review of Linear Algebra 2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear

More information

. Find E(V ) and var(v ).

. Find E(V ) and var(v ). Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number

More information

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio

More information

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix. Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric

More information

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means.

More information

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER 1. The method Ashwelde and Winter [1] proposed a new approach to deviation inequalities for sums of independent random matrices. The

More information

Least Squares Based Self-Tuning Control Systems: Supplementary Notes

Least Squares Based Self-Tuning Control Systems: Supplementary Notes Least Squares Based Self-Tuning Control Systems: Supplementary Notes S. Garatti Dip. di Elettronica ed Informazione Politecnico di Milano, piazza L. da Vinci 32, 2133, Milan, Italy. Email: simone.garatti@polimi.it

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

Homework If the inverse T 1 of a closed linear operator exists, show that T 1 is a closed linear operator.

Homework If the inverse T 1 of a closed linear operator exists, show that T 1 is a closed linear operator. Homework 3 1 If the inverse T 1 of a closed linear operator exists, show that T 1 is a closed linear operator Solution: Assuming that the inverse of T were defined, then we will have to have that D(T 1

More information

Lecture 12 Eigenvalue Problem. Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method

Lecture 12 Eigenvalue Problem. Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method Lecture Eigenvalue Problem Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method Eigenvalue Eigenvalue ( A I) If det( A I) (trivial solution) To obtain

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter

More information

ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR

ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR Periodica Mathematica Hungarica Vol. 51 1, 2005, pp. 11 25 ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR István Berkes Graz, Budapest, LajosHorváth Salt Lake City, Piotr Kokoszka Logan Qi-man Shao

More information

Nonlinear Programming Algorithms Handout

Nonlinear Programming Algorithms Handout Nonlinear Programming Algorithms Handout Michael C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 5376 September 9 1 Eigenvalues The eigenvalues of a matrix A C n n are

More information

Non white sample covariance matrices.

Non white sample covariance matrices. Non white sample covariance matrices. S. Péché, Université Grenoble 1, joint work with O. Ledoit, Uni. Zurich 17-21/05/2010, Université Marne la Vallée Workshop Probability and Geometry in High Dimensions

More information

Differential Analysis II Lectures 22-24

Differential Analysis II Lectures 22-24 Lecture : 3 April 015 18.156 Differential Analysis II Lectures - Instructor: Larry Guth Trans.: Kevin Sacel.1 Precursor to Schrödinger Equation Suppose we are trying to find u(, t) with u(, 0) = f() satisfying

More information

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then 1. x S is a global minimum point of f over S if f (x) f (x ) for any x S. 2. x S

More information

Lecture 4: Analysis of MIMO Systems

Lecture 4: Analysis of MIMO Systems Lecture 4: Analysis of MIMO Systems Norms The concept of norm will be extremely useful for evaluating signals and systems quantitatively during this course In the following, we will present vector norms

More information

Linear Algebra: Characteristic Value Problem

Linear Algebra: Characteristic Value Problem Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number

More information

Total Least Squares Approach in Regression Methods

Total Least Squares Approach in Regression Methods WDS'08 Proceedings of Contributed Papers, Part I, 88 93, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Total Least Squares Approach in Regression Methods M. Pešta Charles University, Faculty of Mathematics

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

Statistical signal processing

Statistical signal processing Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable

More information

Concentration Inequalities for Random Matrices

Concentration Inequalities for Random Matrices Concentration Inequalities for Random Matrices M. Ledoux Institut de Mathématiques de Toulouse, France exponential tail inequalities classical theme in probability and statistics quantify the asymptotic

More information

1 Principal component analysis and dimensional reduction

1 Principal component analysis and dimensional reduction Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an

More information

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX Submitted to the Annals of Statistics HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX By Shurong Zheng, Zhao Chen, Hengjian Cui and Runze Li Northeast Normal University, Fudan

More information

arxiv: v2 [math.pr] 16 Aug 2014

arxiv: v2 [math.pr] 16 Aug 2014 RANDOM WEIGHTED PROJECTIONS, RANDOM QUADRATIC FORMS AND RANDOM EIGENVECTORS VAN VU DEPARTMENT OF MATHEMATICS, YALE UNIVERSITY arxiv:306.3099v2 [math.pr] 6 Aug 204 KE WANG INSTITUTE FOR MATHEMATICS AND

More information

Multivariate Normal-Laplace Distribution and Processes

Multivariate Normal-Laplace Distribution and Processes CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by

More information

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality CS369H: Hierarchies of Integer Programming Relaxations Spring 2016-2017 Lecture 5. Max-cut, Expansion and Grothendieck s Inequality Professor Moses Charikar Scribes: Kiran Shiragur Overview Here we derive

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32 ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach

More information

Matrix Solutions to Linear Systems of ODEs

Matrix Solutions to Linear Systems of ODEs Matrix Solutions to Linear Systems of ODEs James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 3, 216 Outline 1 Symmetric Systems of

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Global Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations

Global Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations Global Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations C. David Levermore Department of Mathematics and Institute for Physical Science and Technology

More information

FIXED POINT ITERATIONS

FIXED POINT ITERATIONS FIXED POINT ITERATIONS MARKUS GRASMAIR 1. Fixed Point Iteration for Non-linear Equations Our goal is the solution of an equation (1) F (x) = 0, where F : R n R n is a continuous vector valued mapping in

More information

= 1 2 x (x 1) + 1 {x} (1 {x}). [t] dt = 1 x (x 1) + O (1), [t] dt = 1 2 x2 + O (x), (where the error is not now zero when x is an integer.

= 1 2 x (x 1) + 1 {x} (1 {x}). [t] dt = 1 x (x 1) + O (1), [t] dt = 1 2 x2 + O (x), (where the error is not now zero when x is an integer. Problem Sheet,. i) Draw the graphs for [] and {}. ii) Show that for α R, α+ α [t] dt = α and α+ α {t} dt =. Hint Split these integrals at the integer which must lie in any interval of length, such as [α,

More information

Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling

Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling 1 Introduction Many natural processes can be viewed as dynamical systems, where the system is represented by a set of state variables and its evolution governed by a set of differential equations. Examples

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

Chapter 13: Functional Autoregressive Models

Chapter 13: Functional Autoregressive Models Chapter 13: Functional Autoregressive Models Jakub Černý Department of Probability and Mathematical Statistics Stochastic Modelling in Economics and Finance December 9, 2013 1 / 25 Contents 1 Introduction

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Chapter 3 - Temporal processes

Chapter 3 - Temporal processes STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect

More information

Bayes Decision Theory - I

Bayes Decision Theory - I Bayes Decision Theory - I Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statistical Learning from Data Goal: Given a relationship between a feature vector and a vector y, and iid data samples ( i,y i ), find

More information

(X i,y, i i ), i =1, 2,,n, Y i = min (Y i, C i ), i = I (Y i C i ) C i. Y i = min (C, Xi t θ 0 + i ) E [ρ α (Y t )] t ρ α (y) =(α I (y < 0))y I (A )

(X i,y, i i ), i =1, 2,,n, Y i = min (Y i, C i ), i = I (Y i C i ) C i. Y i = min (C, Xi t θ 0 + i ) E [ρ α (Y t )] t ρ α (y) =(α I (y < 0))y I (A ) Y F Y ( ) α q α q α = if y : F Y (y ) α. α E [ρ α (Y t )] t ρ α (y) =(α I (y < 0))y I (A ) Y 1,Y 2,,Y q α 1 Σ i =1 ρ α (Y i t ) t X = x α q α (x ) Y q α (x )= if y : P (Y y X = x ) α. q α (x ) θ α q α

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Lecture 5 Least-squares

Lecture 5 Least-squares EE263 Autumn 2008-09 Stephen Boyd Lecture 5 Least-squares least-squares (approximate) solution of overdetermined equations projection and orthogonality principle least-squares estimation BLUE property

More information

Chap. 3. Controlled Systems, Controllability

Chap. 3. Controlled Systems, Controllability Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :

More information

Homogeneous Constant Matrix Systems, Part II

Homogeneous Constant Matrix Systems, Part II 4 Homogeneous Constant Matri Systems, Part II Let us now epand our discussions begun in the previous chapter, and consider homogeneous constant matri systems whose matrices either have comple eigenvalues

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

0.1. Linear transformations

0.1. Linear transformations Suggestions for midterm review #3 The repetitoria are usually not complete; I am merely bringing up the points that many people didn t now on the recitations Linear transformations The following mostly

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Lecture 9: SVD, Low Rank Approximation

Lecture 9: SVD, Low Rank Approximation CSE 521: Design and Analysis of Algorithms I Spring 2016 Lecture 9: SVD, Low Rank Approimation Lecturer: Shayan Oveis Gharan April 25th Scribe: Koosha Khalvati Disclaimer: hese notes have not been subjected

More information

4 Inverse function theorem

4 Inverse function theorem Tel Aviv University, 2014/15 Analysis-III,IV 71 4 Inverse function theorem 4a What is the problem................ 71 4b Simple observations before the theorem..... 72 4c The theorem.....................

More information