Grouped Network Vector Autoregression
|
|
- Hector Allen
- 5 years ago
- Views:
Transcription
1 Statistica Sinica: Supplement Grouped Networ Vector Autoregression Xuening Zhu 1 and Rui Pan 2 1 Fudan University, 2 Central University of Finance and Economics Supplementary Material We present here the detailed technical proofs of Lemma 1 Lemma 4 in Appendi A. Net, the proofs of Theorem 1 to Theorem 3 are given in Appendi B, C, and D respectively. Appendi A. Four Useful Lemmas Lemma 1. Let X = (X 1,, X n) R n, where X is are independent and identically distributed random variables with mean zero, variance σ 2 X and finite fourth order moment. Let Ỹ t = j=0 Gj UE t j, where G R n n, U R n N, and {E t} satisfy Condition (C1) and are independent of {X i}. Then for a matri A = (a ij) R n n and a vector B = (b 1,, b n) R n, it holds that (a) n 1 B X p 0 if n 2 B B 0 as n. (b) n 1 X AX p σ 2 X lim n n 1 tr(a) if the limit eists, and n 2 tr(aa ) 0 as n. (c) (nt ) 1 T B Ỹ t p 0 if n 1 j=0 (B G j UU (G ) j B) 1/2 0 as n. (d) (nt ) 1 T Ỹ t AỸ t p lim n n 1 tr{aγ(0)} if the limit eists, and n 1 i=0 j=0 [tr{u (G ) i AG j UU (G ) j A G i U}] 1/2 0 as n.
2 (e) (nt ) 1 T X AỸ t p 0 if n 1 j=0 [tr{agj UU (G ) j A }] 1/2 0 as n. Proof: The detailed proof can be found in Lemma 1 of Zhu et al. (2017). Lemma 2. Assume min N = O(N δ ) and the stationary condition c β < 1, where c β = ma ( β 1 + β 2 ). Further assume Conditions (C1)-(C3) hold. For matrices M 1 = (m (1) ij ) R n p and M 2 = (m (2) ij ) Rn p, define M 1 M 2 as m (1) ij m (2) ij for 1 i n and 1 j p. In addition, define M e = ( m ij ) R n p for any arbitrary matri M = (m ij) R n p. Then there eists J > 0, such that (a) for any integer n > 0, we have G n (G ) n e n J c 2n β MM, (A.1) G n Σ Y e αn J c n βmm, (A.2) where M = C1π + J j=0 W j, C > 1 is a constant, π is defined in (C3.1), and α is a finite constant. (b) For positive integers 1 1, 2 1, and j 0, define g j,1, 2 (G, W () ) = (W () ) 1 {G j (G ) j } 2 (W () ) 1 e R N N. In addition, define (W () ) 0 = I = (I N, 0) R N N. For integers 0 1, 2, m 1, m 2 1, as N we have N 1 N 1 j=0 i,j=0 { µ g j,1, 2 (G, W () )µ} 1/2 0, (A.3) [ { 1/2 tr g i,1, 2 (G, W () )g j,m1,m 2 (G, W )}] () 0, (A.4) where µ e c µ1 and c µ is a finite constant. (c) For integers 0 1, 2 1, define f 1, 2 (W (), Q) = (W () ) 1 Q 2 (W () ) 1 e R N N,
3 where Q is given in (C3). Then for integers 0 1, 2, m 1, m 2 1, as N we have N 2 µ f 1, 2 (W (), Q)µ 0, { } N 2 tr f 1, 2 (W (), Q)f m1,m 2 (W (), Q) 0, N 1 j=0 (A.5) (A.6) [ { 1/2 tr f 1, 2 (W (), Q)g j,m1,m 2 (G, W )}] () 0, (A.7) where µ e c µ1 and c µ is a finite constant. Proof: The proof is similar in spirit to Zhu et al. (2017). Therefore, we give the guideline of the proof and sip some similar details. Without loss of generality, we let c β = β 11 + β 21 (i.e., = 1). Consequently, we have G e β 11 W + β 21 I. Let G = β 11 W + β 21 I. Follow similar technique in part (a) in Lemma 2 of Zhu et al. (2017), it can be verified G n e n J c n βm, (A.8) where M = C1π + J j=0 W j is defined in (a) of Lemma 2. Subsequently, the result (A.1) can be readily obtained. Net, recall that Σ Y = (I G) 1 Σ Z (I G ) 1 + j=0 Gj Σ e(g ) j = ( j=0 Gj )Σ Z ( j=0 (G ) j ) + j=0 Gj Σ e(g ) j. Let σ 2 z = ma {γ Σ zγ } and σ 2 e = ma {σ 2 }. Then we have G n Σ Y e σ 2 z( j=0 Gn+j e)( j=0 (G ) j e) + σ 2 e j=0 Gn+j e (G ) j e. Subsequently, (A.2) can by obtained by applying (A.8). Net, we give the proof of (b) in the following. The conclusion (c) can be proved by similar techniques, which is omitted here to save space. Let 1 = 2 = 1. Then we have g j,1,1(g, W () ) = W () G j G j W (). Recall that W () = (w ij : i M, 1 j N) R N N. Since we have µ e c µ1, then it suffices to show j=0 N { 1 1 g j,1,1(g, W () )1 } 1/2 0. We first prove (A.3). By (A.8) we have W () G j e
4 j K ( β 1 + β 2 ) j W () M. As a result, we have W () G j (G ) j W () e j 2K ( β 1 + β 2 ) 2j M, (A.9) where M is defined as M = W () MM W (). As a result, we have j=0 N { 1 1 W () G j (G ) j W () 1 } 1/2 N 1 α1(1 M1) 1/2, where α 1 = j=0 jk c j β <. Then it leads to show N 2 1 M1 0. It can be shown 1 M1 = N 2 C j π2 j + K j=1 1 W () W j (W ) j W () 1 + 2N C j π (W ) j W () 1+ i j 1 W () W i (W ) j W () 1. For the last two terms of 1 M1, by Cauchy inequality, we have ( N π (W ) j W () 1 N j j π 2 j ) 1/2 { 1 W () W j (W ) j W () 1} 1/2, and i j 1 W () W i (W ) j W () 1 i j W () 1 } 1/2. As a result, it leads to show { 1 W () W i (W ) i W () 1 } 1/2{ 1 W () W j (W ) j N j=1 π 2 j 0 and N 2 1 W () W j (W ) j W () 1 0 (A.10) for 1 j K + 1. As the first convergence in (A.10) is implied by (C2.1), we net prove N 2 1 W () W j (W ) j W () 1 0 (1 j K). Recall that W = W + W. Therefore, we have N 2 1 W () W j (W ) j W () 1 N 2 1 W () W 2j W () 1. Then it suffices to show N 2 1 W () W 2j W () 1 0. By eigenvalue-eigenvector decomposition of W we have W = λ (W )u u, where λ (W ) and u R N are the th eigenvalue and eigenvector of W respectively. As a result, we have N 2 1 W () W 2j W () 1 N 2 λma(w ) 2j (1 W () W () 1) (1 j K). Further we have 1 W () W () 1 N λ ma(w () W () ). Note that W () W ()
5 is a sub-matri of W W with row and column inde in M. Therefore, by Cauchy s interlacing Theorem, we have λ ma(w () W () ) λ ma(w W ) = O(N δ ) for δ < δ. Since we have min N = N δ for δ > 0, then we have N 1 λma(w () W () ) 0 as N. As a consequence, the second term in (A.10) holds. Similarly, it can be proved that (A.10) holds for all 0 1, 2 1. As a result, we have (A.3) holds. We net prove (A.4) with 1 = 2 = m 1 = m 2 = 1, and g i,1,1(g, W () )g j,1,1(g, W () ) = W () G i (G ) i W () W () G j (G ) j W () e. Then it can be similarly proved for other cases (i.e., 0 1, 2, m 1, m 2 1). Note that by (A.9), we have [ { tr W () G i (G ) i W () W () G j (G ) j W () }] 1/2 i K j K ( β 1 + β 2 ) i+j tr { M 2} 1/2. It then can be derived that N 1 i,j=0 [tr{w () G i (G ) i W () W () G j (G ) j W () }] 1/2 α 2 N 1 tr { M 2} 1/2. In order to obtain (A.4), it suffices to show that N 2 tr{m 2 } 0. (A.11) Equivalently, by Cauchy inequality, it suffices to prove ( πj 2 ) 2 0, and N 2 tr{w () W j W j W () W () W j W j W () } 0 holds for 1 j K. It can be easily verified the first term holds by (C2.1). For the second one, we have N 2 tr{w () W j W j W () W () W j W j W () } N 2 tr{w () (W ) 4j W () } N 2 λma(w ) 4j tr(w () W () ) N 2 N λ ma(w ) 4K λ ma(w W ). Similarly, due to that λ ma(w ) = O(log N) and λ ma(w W ) = O(N δ ) in (C2.2), we have N 1 λma(w ) 4K λ ma(w W ) 0 as N. Consequently, we have (A.11) and then (A.4) holds. This completes the proof of (b). Lemma 3. Let {X it : 1 t T } and {Y it : 1 t T } be random sub-gaussian time
6 series with mean 0, var(x it) = σ i,, var(y it) = σ i,yy, and cov(x it, Y it) = σ i,y. Let σ i,t1 t 2 = cov(x it1, X it2 ) and Σ i = (σ i,t1 t 2 : 1 t 1, t 2 T ) R T T. Similarly, define σ yi,t1 t 2 and Σ yi R T T. Then we have ( T 1 P T ) X ity it σ i,y > ν { } c 1 ep( c 2σ 2 i T 2 ν 2 ) + ep( c 2σ 2 yi T 2 ν 2 ) (A.12) for ν δ, where σ 2 i = tr(σ 2 i), σ 2 yi = tr(σ 2 yi), c 1, c 2, and δ are finite constants. Proof: Let X i = (X i1,, X it ) R T and Y i = (Y i1,, Y it ) R T. In addition, let Z i = Z i + Y i. Therefore, we have Zi Z i = 2 1 (Zi Z i Xi X i Yi Y i). It can be derived that P { T 1 (X i Y i) σ i,y ν} P { T 1 (Z i Z i) (σ i, + σ i,yy + 2σ i,y) ν 1} + P { T 1 (X i X i) σ i, ν 1} + P { T 1 (Y i Y i) σ i,yy ν 1}, (A.13) where ν 1 = 2ν/3. Net, we derive the upper bound for the right side of (A.13). Note that X i X i, Yi Y i, and Zi Z i all tae quadratic form. Therefore the proofs are similar. For the sae of simplicity, we tae Y i Y i for an eample and derive the upper bound for P { n 1 (Y Y i) σ i,yy ν 1}. Similar results can be obtained for the other two terms. First we have Yi Y i = Yi Σ 1/2 yi Σ yiσ 1/2 yi Y i = Ỹ Σ yi Ỹ i, where i i Ỹi = Σ 1/2 yi Y i follows sub- Gaussian distribution. Let λ 1 λ 2 λ T be the eigenvalues of Σ yi. Since Σ yi is a nonnegative definite matri, The eigenvalue decomposition can be applied to obtain Σ yi = U ΛU, where U = (U 1,, U T ) R T T is an orthogonal matri and Λ = diag{λ 1,, λ T }. As a consequence, we have Yt Y t = t λtζ2 t, where ζ t = Ut Ỹt and ζts are independent and identically distributed as standard sub-gaussian. It can be verified ζ 2 t 1 satisfies sub-eponential distribution and T 1 ( t λt) = σi,yy. In addition, the sub-eponential distribution satisfies condition
7 (P) on page 45 of Saulis and Statuleviveccius (2012). There eists constants c 1, c 2, and δ such that P { T 1 (Y i Y i) σ i,yy ν 1} = P { t λt(ζ2 t 1) T ν 1} c 1 ep{ c 2( t λ2 t ) 1 T 2 ν 2 } = c 1 ep{ c 2σ 1 yi T 2 ν 2 } for ν < δ by the Theorem 3.3 of Saulis and Statuleviveccius (2012). Consequently, (A.12) can be obtained by appropriately chosen c 1, c 2, and δ. Lemma 4. Assume Y it follows the GNAR model (2.4) and c β < 1. Then there eists finite constants c 1, c 2, and δ, for ν < δ we have P { T 1 P { T 1 P { T 1 P { T 1 T Y 2 it µ 2 i e i Σ Y e i > ν } δ T, T Y it(wi Y t) µ Y i(wi µ Y ) wi } Σ Y e i > ν δt T } Y i(t 1) ε it > ν δt, T } Y i(t 1) µ i > ν δt, P { T 1 (A.14) (A.15) T (wi } Y t 1)ε it > ν δt, (A.16) P { T 1 T wi Y t wi } µ Y > ν δt, (A.17) where δ T = c 1 ep( c 2T ν 2 ), e i R N is an N-dimensional vector with all elements being 0 but the ith element being 1, and µ i = e i µ Y. Proof: For the similarity of proof procedure, we only prove (A.14) in the following. Without loss of generality, let µ Y = 0. Recall that the group information is denoted as Z = {z i : 1 i N, 1 K}. Define P ( ) = P ( Z), E ( ) = E( Z), and cov ( ) = cov( Z). Write Y i = (Y i1,, Y it ) R T. Given Z, Y i is a sub-gaussian random vector with cov(y i) = Σ i = (σ i,t1 t 2 ) R T T, where σ i,t1 t 2 = e i G t 1 t 2 Σ Y e i for t 1 t 2, σ i,t1 t 2 = e i Σ Y (G ) t 2 t 1 e i, and G is pre-defined in (2.6) as G = B 1W + B 2. It can be derived var (Y i Y i) ctr(σ 2 i ), where c is a positive constant and tr(σ 2 i ) = T (e i Σ Y e i) T 1 (T t)(e i G t Σ Y e i) 2. It can be derived Σ Y e αmm and G t Σ Y e α 1t J c t βmm by (A.2) of Lemma 2, where c β, J and M are defined in Lemma 2, α and α 1 are finite constants. In addition, it can be
8 verified T 1 (T t)t2j c 2t β α 2T, where α 2 is a finite constant. Therefore we have tr(σ 2 i ) T (α + 2α 1α 2){(e i MM e i) 2 }. Since we have e i MM e i (J + 1)e i M1 (J + 1) 2 = O(1), it can be concluded that tr(σ 2 i ) T α 3, where α 3 = (α + 2α 1α 2)(J + 1) 2. By Lemma 3, the (A.14) can be obtained. Appendi B. Proof of Theorem 1 Let λ i(m) be the ith eigenvalue of M R N N. We first verify that the solution (2.7) is strictly stationary. By Banerjee et al. (2014), we have ma i λ i(w ) 1. Hence we have ma λi(g) ( ma β 1 )( ma λi(w ) ) + ma β 2 < 1. 1 i N 1 K 1 i N 1 K (A.18) Consequently, we have lim m m j=0 Gj E t j eists and {Y t} given by (2.7) is a strictly stationary process. In addition, one could directly verify that {Y t} satisfies the GNAR model (2.4). Net, we verify that the strictly stationary solution (2.7) is unique. Assume {Ỹt} is another strictly stationary solution to the GNAR model (2.4) with E Ỹt <. Then we have Ỹt = m 1 j=1 Gj (B 0 + E t j) + G m Ỹ t m for any positive integer m. Let ρ = ma ( β 1 + β 2 ). Then one could verify E Y t Ỹt = E j=m Gj (B 0 + E t j) G m Ỹ t m Cρ m, where C is a finite constant unrelated to t and m. Note that m can be chosen arbitrarily. As a result, we have that E Y t Ỹt = 0, i.e. Yt = Ỹt with probability one. This completes the proof. Appendi C. Proof of Theorem 2 According to (3.9), θ can be eplicitly written as θ = θ + Σ 1 ζ, where Σ = (N T ) 1 T X () t 1 X() t 1 and ζ = (N T ) 1 T X() t 1 E () t. Without loss of generality, we assume σ 2 = 1 for
9 = 1,, K. Let Σ = lim N E( Σ ). As a result, it suffices to show that Σ p Σ, N T ζ = O p(1), (A.19) (A.20) as min{n, T }. Subsequently, we prove (A.19) in Step 1 and (A.20) in Step 2. Step 1. Proof of (A.19). Define Q = (I G) 1 Σ V (I G ) 1. In this step, we intend to show that Σ = 1 N T T X () t 1 X() t 1 = 1 S 12 S 13 S 14 S 22 S 23 S 24 S 33 S 34 S 44 p 1 c 1β c 2β 0 Σ 1 Σ 2 κ 8γ Σ z Σ 3 κ 3γ Σ z Σ z = Σ, where S 12 = 1 N T T wi Y t 1, S 13 = 1 T Y i(t 1), S 14 = 1 N T N i M i M Vi, i M S 22 = 1 N T T (wi Y t 1) 2, S 23 = 1 T wi Y t 1Y i(t 1), N T i M i M S 24 = 1 N T T wi Y t 1Vi, S 33 = 1 N T i M T Yi(t 1), 2 i M S 34 = (N T ) 1 T i M Y i(t 1) V i, S 44 = N 1 i M V ivi. By (2.7), we have Y t = (I G) 1 b 0 + (I G) 1 b v + Ỹt, (A.21)
10 where b 0 = D B 0, b v = D Vγ, and Ỹt = j=0 Gj E t j. By the law of large numbers, one could directly obtain that S 44 p Σ v and S 14 p 0. Subsequently, we only show the convergence of S 12 and S 23 in Σ as follows. Convergence of S 12. It can be derived that S 12 = 1 N T T 1 W () Y t 1 = 1 W () µ Y N + S 12a + S 12b, where S 12a = N 1 1 W () (I G) 1 b v and S 12b = (N T ) 1 T 1 W () Ỹ t 1. Then by (A.5) and (A.3) in Lemma 2, we have N 2 1 W () QW () 1 0 and N 1 { j=0 1 W () G j (G ) j W () 1 } 1/2 0, as N. As a result, it is implied by Lemma 1 (a) and (c) that S 12a p 0 and S 12b p 0. Convergence of S 23. Note that S 23 = 1 N T T wi Y t 1Y i(t 1) = 1 T N T i M = µ() Y W () µ Y + S 23a + S 23b + S 23c + S 23d + S 23e, N Y () t 1 W () Y t 1 where S 23a = N 1 b v I W () bv, S 23b = N 1 T 1 T ( b v I W () Ỹ t 1 + Ỹ t 1I W () bv), S 23d = N 1 Ỹ() t 1 W () Ỹ t 1 and S 23c = N 1 T 1 T ( b v I µ Y + µ Y I b v), S 23e = N 1 T 1 T (Y () t 1 µy + µ Y I W () Y t 1), where µ Y = W () µ Y and b v = (I G) 1 b v. We net loo at the terms one by one. First we have N 2 tr(i QI W () QW () ) 0 by (A.6) in Lemma 2 (c). Therefore, by (b) in Lemma 1, we have S 23a p s 23a, where s 23a = lim N E(S 23a). Net, for S 23b we have N 1 i,j=0 tr{i G i (G ) i I W () G j (G ) j W () } 0 by (A.4) in Lemma 2 (b). Therefore, by (d) in Lemma 1, we have S 23b p s 23b, where s 23b = lim N E(S 23b ). Net, let S 23c = S (1) 23c +S(2) 23c, where S(1) 23c = N 1 T 1 T b z I W () Ỹ t 1 and
11 23c = N 1 T 1 T Ỹ t 1I W () bv. Note that we have N 1 j=0 tr{w () G j (G ) j W () I Q S (2) I } 0 and N 1 j=0 tr{i G j (G ) j I W () QW () } 0 by (A.7) in Lemma 2 (c). Therefore, S 23c p s 23c by (e) in Lemma 1, where s 23c = lim N E(S 23c). Net, by similar proof to the convergence of S 13, we have that S 23d p 0 and S 23e p 0. As a consequence, we have S 23 p Σ 2. Step 2. Proof of (A.20). It can be verified that N T E( ζ ) = 0. In addition, we have var{ N T ζ } = E( Σ ) Σ as N. Consequently, we have N T ζ = O p(1). Appendi D. Proof of Theorem 3 Let Σ (i) ε it). We then have = T 1 T X i(t 1)X i(t 1) = ( σ,ij) R 3 3 (i), and Σ e = T 1 ( T X i(t 1)δ i bi b i = ( Σ (i) ) 1 Σ (i) e. Let Σ (i) = ( σ,j1 j 2 : 1 l 1, l 2 3) R 3 3, where the inde i of σ,l1 l 2 is omitted. Specifically, σ,11 = 1, σ,12 = T 1 t w i Y t 1, σ,13 = T 1 t e i Y t 1, σ,22 = T 1 t Y 2 i(t 1), σ,23 = T 1 t Y i(t 1)(w i Y t 1), σ,33 = T 1 t (w i Y t 1) 2. Mathematically, it can be computed ( Σ (i) ) 1 = Σ (i) 1 Σ (i), where Σ (i) is the determinant of (i) (i) Σ, and Σ is the adjugate matri of (i) Σ, and Σ (i) = ( σ,l 1 l 2 ), where σ,11 = σ,22 σ,33 σ,23, 2 σ,12 = σ,13 σ,32 σ,12 σ,33 σ,13 = σ,21 σ,32 σ,22 σ,31, σ,22 = σ,11 σ,33 σ 2,13, σ,23 = σ,13 σ,32 σ,12 σ,33, and σ,33 = σ,11 σ,22 σ,12. 2 It can be derived Σ (i) = σ,11( σ,22 σ,33 σ,23) 2 σ,12( σ,12 σ 33 σ 13 σ 23) + σ 13( σ 12 σ 23 σ 22 σ 13). By the maimum inequality, we have P (sup b i b i > ν) i N P ( b i b i > ν). i=1 (A.22)
12 REFERENCES In addition, we have P ( b i b i > ν) P ( Σ(i) σ (i) ) ( δi + P Σ (i) Σ (i) e ) δiν, (A.23) where σ (i) = σ,11(σ,22σ,33 σ 2,23) σ,12(σ,12σ 33 σ 13σ 23) + σ 13(σ 12σ 23 σ 22σ 13) = (e i Σ Y e i)(w i Σ Y w i) (e i Σ Y w i) 2, δ i = σ (i) /2. By lemma 4, for each component of Σ (i) we have P ( σ,l1 l 2 σ,l1 l 2 > ν 0) c 1 ep( c 2T ν 2 0), where σ,l1 l 2 = E( σ,l1 l 2 ) and ν 0 is a finite positive constant. Moreover, by the conditions of Theorem 3, we have σ (i) tending to 1. Consequently, it is not difficult to obtain the result P ( Σ (i) τ with probability σ (i) δ i) c 1 ep( c 2T τ 2 ), where c 1, c 2 are finite constants. Subsequently, we have P ( Σ (i) Σ(i) e δ iν) P ( Σ (i) Σ(i) e τν/2). By similar technique, one could verify that each element of Σ (i) and Σ (i) e converge with probability and the tail probability can be controlled, where the basic results are given in Lemma 4. Consequently, there eists constants c 3 and c 4 such that P ( Σ (i) Σ(i) e τν/2) c 3 ep( c 4T τ 2 ν 2 ). Consequently, we have P ( b i b i > ν) c 1 ep( c 2T τ 2 ) + c 3 ep( c 4T τ 2 ν 2 ) by (A.23). By the condition N = o(ep(t )), the right side of (A.22) goes to 0 as N. This completes the proof. References Banerjee, S., Carlin, B. P., and Gelfand, A. E. (2014), Hierarchical modeling and analysis for spatial data, Crc Press. Saulis, L. and Statuleviveccius, V. (2012), Limit theorems for large deviations, vol. 73, Springer Science & Business Media. Zhu, X., Pan, R., Li, G., Liu, Y., and Wang, H. (2017), Networ vector autoregression, Annals
13 of statistics, 45, REFERENCES
ANOVA: Analysis of Variance - Part I
ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.
More informationNORMS ON SPACE OF MATRICES
NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system
More informationDef. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as
MAHALANOBIS DISTANCE Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as d E (x, y) = (x 1 y 1 ) 2 + +(x p y p ) 2
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationj=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u
More informationThe Hilbert Space of Random Variables
The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2
More informationAssumptions for the theoretical properties of the
Statistica Sinica: Supplement IMPUTED FACTOR REGRESSION FOR HIGH- DIMENSIONAL BLOCK-WISE MISSING DATA Yanqing Zhang, Niansheng Tang and Annie Qu Yunnan University and University of Illinois at Urbana-Champaign
More informationLectures 2 3 : Wigner s semicircle law
Fall 009 MATH 833 Random Matrices B. Való Lectures 3 : Wigner s semicircle law Notes prepared by: M. Koyama As we set up last wee, let M n = [X ij ] n i,j=1 be a symmetric n n matrix with Random entries
More information7. MULTIVARATE STATIONARY PROCESSES
7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability
More informationCONTROL CHARTS FOR MULTIVARIATE NONLINEAR TIME SERIES
REVSTAT Statistical Journal Volume 13, Number, June 015, 131 144 CONTROL CHARTS FOR MULTIVARIATE NONLINEAR TIME SERIES Authors: Robert Garthoff Department of Statistics, European University, Große Scharrnstr.
More informationHyperbolic Systems of Conservation Laws. in One Space Dimension. II - Solutions to the Cauchy problem. Alberto Bressan
Hyperbolic Systems of Conservation Laws in One Space Dimension II - Solutions to the Cauchy problem Alberto Bressan Department of Mathematics, Penn State University http://www.math.psu.edu/bressan/ 1 Global
More informationAsymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data
Sri Lankan Journal of Applied Statistics (Special Issue) Modern Statistical Methodologies in the Cutting Edge of Science Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations
More information5: MULTIVARATE STATIONARY PROCESSES
5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability
More informationDYNAMIC AND COMPROMISE FACTOR ANALYSIS
DYNAMIC AND COMPROMISE FACTOR ANALYSIS Marianna Bolla Budapest University of Technology and Economics marib@math.bme.hu Many parts are joint work with Gy. Michaletzky, Loránd Eötvös University and G. Tusnády,
More informationOn Expected Gaussian Random Determinants
On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose
More informationA Short Course in Basic Statistics
A Short Course in Basic Statistics Ian Schindler November 5, 2017 Creative commons license share and share alike BY: C 1 Descriptive Statistics 1.1 Presenting statistical data Definition 1 A statistical
More informationNonlinear function on function additive model with multiple predictor curves. Xin Qi and Ruiyan Luo. Georgia State University
Statistica Sinica: Supplement Nonlinear function on function additive model with multiple predictor curves Xin Qi and Ruiyan Luo Georgia State University Supplementary Material This supplementary material
More informationMIMO Capacities : Eigenvalue Computation through Representation Theory
MIMO Capacities : Eigenvalue Computation through Representation Theory Jayanta Kumar Pal, Donald Richards SAMSI Multivariate distributions working group Outline 1 Introduction 2 MIMO working model 3 Eigenvalue
More informationMATH 205C: STATIONARY PHASE LEMMA
MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)
More information11. Regression and Least Squares
11. Regression and Least Squares Prof. Tesler Math 186 Winter 2016 Prof. Tesler Ch. 11: Linear Regression Math 186 / Winter 2016 1 / 23 Regression Given n points ( 1, 1 ), ( 2, 2 ),..., we want to determine
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationSection 3.9. Matrix Norm
3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix
More informationLecture 3 - Expectation, inequalities and laws of large numbers
Lecture 3 - Expectation, inequalities and laws of large numbers Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 3 - Expectation, inequalities and laws of large numbersapril 19, 2009 1 / 67 Part
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationSystems of Algebraic Equations and Systems of Differential Equations
Systems of Algebraic Equations and Systems of Differential Equations Topics: 2 by 2 systems of linear equations Matrix expression; Ax = b Solving 2 by 2 homogeneous systems Functions defined on matrices
More informationMidterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015
Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 The test lasts 1 hour and 15 minutes. No documents are allowed. The use of a calculator, cell phone or other equivalent electronic
More informationQuadratic Programming
Quadratic Programming Quadratic programming is a special case of non-linear programming, and has many applications. One application is for optimal portfolio selection, which was developed by Markowitz
More informationMATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018
MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S
More informationGiven the vectors u, v, w and real numbers α, β, γ. Calculate vector a, which is equal to the linear combination α u + β v + γ w.
Selected problems from the tetbook J. Neustupa, S. Kračmar: Sbírka příkladů z Matematiky I Problems in Mathematics I I. LINEAR ALGEBRA I.. Vectors, vector spaces Given the vectors u, v, w and real numbers
More informationNotes 6 : First and second moment methods
Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative
More informationA Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report -
A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report - Edgar Brunner and Marius Placzek University of Göttingen, Germany 3. August 0 . Statistical Model and Hypotheses Throughout
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationThere is a unique function s(x) that has the required properties. It turns out to also satisfy
Numerical Analysis Grinshpan Natural Cubic Spline Let,, n be given nodes (strictly increasing) and let y,, y n be given values (arbitrary) Our goal is to produce a function s() with the following properties:
More informationMP 472 Quantum Information and Computation
MP 472 Quantum Information and Computation http://www.thphys.may.ie/staff/jvala/mp472.htm Outline Open quantum systems The density operator ensemble of quantum states general properties the reduced density
More informationLecture 7: Weak Duality
EE 227A: Conve Optimization and Applications February 7, 2012 Lecture 7: Weak Duality Lecturer: Laurent El Ghaoui 7.1 Lagrange Dual problem 7.1.1 Primal problem In this section, we consider a possibly
More informationSUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley
SUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley A. Proofs A.1. Proof of Proposition.1 e have E T n (r k (X)) r? (X) = E T n (r k (X)) T(r
More informationarxiv: v5 [math.na] 16 Nov 2017
RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem
More informationPrincipal Components Theory Notes
Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationHyperbolic Systems of Conservation Laws. I - Basic Concepts
Hyperbolic Systems of Conservation Laws I - Basic Concepts Alberto Bressan Mathematics Department, Penn State University Alberto Bressan (Penn State) Hyperbolic Systems of Conservation Laws 1 / 27 The
More informationStein s Method and Characteristic Functions
Stein s Method and Characteristic Functions Alexander Tikhomirov Komi Science Center of Ural Division of RAS, Syktyvkar, Russia; Singapore, NUS, 18-29 May 2015 Workshop New Directions in Stein s method
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationOn rate of convergence in distribution of asymptotically normal statistics based on samples of random size
Annales Mathematicae et Informaticae 39 212 pp. 17 28 Proceedings of the Conference on Stochastic Models and their Applications Faculty of Informatics, University of Debrecen, Debrecen, Hungary, August
More informationCommon-Knowledge / Cheat Sheet
CSE 521: Design and Analysis of Algorithms I Fall 2018 Common-Knowledge / Cheat Sheet 1 Randomized Algorithm Expectation: For a random variable X with domain, the discrete set S, E [X] = s S P [X = s]
More information1 Kernel methods & optimization
Machine Learning Class Notes 9-26-13 Prof. David Sontag 1 Kernel methods & optimization One eample of a kernel that is frequently used in practice and which allows for highly non-linear discriminant functions
More informationIn particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with
Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient
More informationDuke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014
Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document
More informationSemicircle law on short scales and delocalization for Wigner random matrices
Semicircle law on short scales and delocalization for Wigner random matrices László Erdős University of Munich Weizmann Institute, December 2007 Joint work with H.T. Yau (Harvard), B. Schlein (Munich)
More information2. Review of Linear Algebra
2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear
More information. Find E(V ) and var(v ).
Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number
More informationRandom Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications
Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio
More informationQuadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.
Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric
More informationJackknife Empirical Likelihood Test for Equality of Two High Dimensional Means
Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means.
More informationA NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER
A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER 1. The method Ashwelde and Winter [1] proposed a new approach to deviation inequalities for sums of independent random matrices. The
More informationLeast Squares Based Self-Tuning Control Systems: Supplementary Notes
Least Squares Based Self-Tuning Control Systems: Supplementary Notes S. Garatti Dip. di Elettronica ed Informazione Politecnico di Milano, piazza L. da Vinci 32, 2133, Milan, Italy. Email: simone.garatti@polimi.it
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More informationHomework If the inverse T 1 of a closed linear operator exists, show that T 1 is a closed linear operator.
Homework 3 1 If the inverse T 1 of a closed linear operator exists, show that T 1 is a closed linear operator Solution: Assuming that the inverse of T were defined, then we will have to have that D(T 1
More informationLecture 12 Eigenvalue Problem. Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method
Lecture Eigenvalue Problem Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method Eigenvalue Eigenvalue ( A I) If det( A I) (trivial solution) To obtain
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR
Periodica Mathematica Hungarica Vol. 51 1, 2005, pp. 11 25 ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR István Berkes Graz, Budapest, LajosHorváth Salt Lake City, Piotr Kokoszka Logan Qi-man Shao
More informationNonlinear Programming Algorithms Handout
Nonlinear Programming Algorithms Handout Michael C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 5376 September 9 1 Eigenvalues The eigenvalues of a matrix A C n n are
More informationNon white sample covariance matrices.
Non white sample covariance matrices. S. Péché, Université Grenoble 1, joint work with O. Ledoit, Uni. Zurich 17-21/05/2010, Université Marne la Vallée Workshop Probability and Geometry in High Dimensions
More informationDifferential Analysis II Lectures 22-24
Lecture : 3 April 015 18.156 Differential Analysis II Lectures - Instructor: Larry Guth Trans.: Kevin Sacel.1 Precursor to Schrödinger Equation Suppose we are trying to find u(, t) with u(, 0) = f() satisfying
More informationLecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then
Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then 1. x S is a global minimum point of f over S if f (x) f (x ) for any x S. 2. x S
More informationLecture 4: Analysis of MIMO Systems
Lecture 4: Analysis of MIMO Systems Norms The concept of norm will be extremely useful for evaluating signals and systems quantitatively during this course In the following, we will present vector norms
More informationLinear Algebra: Characteristic Value Problem
Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number
More informationTotal Least Squares Approach in Regression Methods
WDS'08 Proceedings of Contributed Papers, Part I, 88 93, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Total Least Squares Approach in Regression Methods M. Pešta Charles University, Faculty of Mathematics
More informationSHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan
A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,
More informationStatistical signal processing
Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable
More informationConcentration Inequalities for Random Matrices
Concentration Inequalities for Random Matrices M. Ledoux Institut de Mathématiques de Toulouse, France exponential tail inequalities classical theme in probability and statistics quantify the asymptotic
More information1 Principal component analysis and dimensional reduction
Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an
More informationHYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX
Submitted to the Annals of Statistics HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX By Shurong Zheng, Zhao Chen, Hengjian Cui and Runze Li Northeast Normal University, Fudan
More informationarxiv: v2 [math.pr] 16 Aug 2014
RANDOM WEIGHTED PROJECTIONS, RANDOM QUADRATIC FORMS AND RANDOM EIGENVECTORS VAN VU DEPARTMENT OF MATHEMATICS, YALE UNIVERSITY arxiv:306.3099v2 [math.pr] 6 Aug 204 KE WANG INSTITUTE FOR MATHEMATICS AND
More informationMultivariate Normal-Laplace Distribution and Processes
CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by
More informationLecture 5. Max-cut, Expansion and Grothendieck s Inequality
CS369H: Hierarchies of Integer Programming Relaxations Spring 2016-2017 Lecture 5. Max-cut, Expansion and Grothendieck s Inequality Professor Moses Charikar Scribes: Kiran Shiragur Overview Here we derive
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32
ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach
More informationMatrix Solutions to Linear Systems of ODEs
Matrix Solutions to Linear Systems of ODEs James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 3, 216 Outline 1 Symmetric Systems of
More informationOptimization Theory. A Concise Introduction. Jiongmin Yong
October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization
More informationGlobal Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations
Global Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations C. David Levermore Department of Mathematics and Institute for Physical Science and Technology
More informationFIXED POINT ITERATIONS
FIXED POINT ITERATIONS MARKUS GRASMAIR 1. Fixed Point Iteration for Non-linear Equations Our goal is the solution of an equation (1) F (x) = 0, where F : R n R n is a continuous vector valued mapping in
More information= 1 2 x (x 1) + 1 {x} (1 {x}). [t] dt = 1 x (x 1) + O (1), [t] dt = 1 2 x2 + O (x), (where the error is not now zero when x is an integer.
Problem Sheet,. i) Draw the graphs for [] and {}. ii) Show that for α R, α+ α [t] dt = α and α+ α {t} dt =. Hint Split these integrals at the integer which must lie in any interval of length, such as [α,
More informationExamples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling
1 Introduction Many natural processes can be viewed as dynamical systems, where the system is represented by a set of state variables and its evolution governed by a set of differential equations. Examples
More informationTesting Some Covariance Structures under a Growth Curve Model in High Dimension
Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping
More informationChapter 13: Functional Autoregressive Models
Chapter 13: Functional Autoregressive Models Jakub Černý Department of Probability and Mathematical Statistics Stochastic Modelling in Economics and Finance December 9, 2013 1 / 25 Contents 1 Introduction
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationChapter 3 - Temporal processes
STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect
More informationBayes Decision Theory - I
Bayes Decision Theory - I Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statistical Learning from Data Goal: Given a relationship between a feature vector and a vector y, and iid data samples ( i,y i ), find
More information(X i,y, i i ), i =1, 2,,n, Y i = min (Y i, C i ), i = I (Y i C i ) C i. Y i = min (C, Xi t θ 0 + i ) E [ρ α (Y t )] t ρ α (y) =(α I (y < 0))y I (A )
Y F Y ( ) α q α q α = if y : F Y (y ) α. α E [ρ α (Y t )] t ρ α (y) =(α I (y < 0))y I (A ) Y 1,Y 2,,Y q α 1 Σ i =1 ρ α (Y i t ) t X = x α q α (x ) Y q α (x )= if y : P (Y y X = x ) α. q α (x ) θ α q α
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More informationLecture 5 Least-squares
EE263 Autumn 2008-09 Stephen Boyd Lecture 5 Least-squares least-squares (approximate) solution of overdetermined equations projection and orthogonality principle least-squares estimation BLUE property
More informationChap. 3. Controlled Systems, Controllability
Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :
More informationHomogeneous Constant Matrix Systems, Part II
4 Homogeneous Constant Matri Systems, Part II Let us now epand our discussions begun in the previous chapter, and consider homogeneous constant matri systems whose matrices either have comple eigenvalues
More informationStatistics & Data Sciences: First Year Prelim Exam May 2018
Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More information0.1. Linear transformations
Suggestions for midterm review #3 The repetitoria are usually not complete; I am merely bringing up the points that many people didn t now on the recitations Linear transformations The following mostly
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationLecture 9: SVD, Low Rank Approximation
CSE 521: Design and Analysis of Algorithms I Spring 2016 Lecture 9: SVD, Low Rank Approimation Lecturer: Shayan Oveis Gharan April 25th Scribe: Koosha Khalvati Disclaimer: hese notes have not been subjected
More information4 Inverse function theorem
Tel Aviv University, 2014/15 Analysis-III,IV 71 4 Inverse function theorem 4a What is the problem................ 71 4b Simple observations before the theorem..... 72 4c The theorem.....................
More information