Testing Block-Diagonal Covariance Structure for High-Dimensional Data Under Non-normality

Size: px
Start display at page:

Download "Testing Block-Diagonal Covariance Structure for High-Dimensional Data Under Non-normality"

Transcription

1 Testing Block-Diagonal Covariance Structure for High-Dimensional Data Under Non-normality Yuki Yamada, Masashi Hyodo and Takahiro Nishiyama Department of Mathematical Information Science, Tokyo University of Science, 1-3, Kagurazaka, Shinjuku-ku, Tokyo, , Japan, Department of Mathematical Sciences, Graduate School of Engineering, Osaka prefecture university, 1-1, Gakuen-cho, Naka-ku, Sakai-shi, Osaka , Japan. Department of Business Administration, Senshu University, -1-1 Higashimita, Tama-ku, Kawasaki-shi, Kanagawa , Japan. Abstract In this article, we propose a test for making an inference about the blockdiagonal covariance structure of a covariance matrix in non-normal highdimensional data. We prove that the limiting null distribution of the proposed test is normal under mild conditions when its dimension is substantially larger than its sample size. We further study the local power of the proposed test. Finally, we study the finite sample performance of the proposed test via Monte Carlo simulations. We demonstrate the relevance and benefits of the proposed approach for a number of alternative covariance structures. AMS 000 subject classification: Primary 6H15; secondary 6F05. Key words: Tests of covariance structure, high dimension, statistical hypothesis testing, non-normality. 1. Introduction In recent years, a number of statistical methods for high-dimensional data such as DNA microarray gene expressions, where the number of feature variables, p, exceeds the sample size, N, have been discussed by many Hiroshima Statistical Research Group Technical Report March 30, 016

2 aouthors. For analysing high-dimensional data, the p p covariance matrix Σ is an important measure for dependence among the components of a high-dimensional random vector x, and thus plays a major role in many statistical procedures. Accordingly, the statistical inference of Σ is an important problem. Incidentally, a number of procedures for testing a high-dimensional covariance matrix, which include testing identity, sphericity and diagonal structure, have been proposed by Schott 005), Srivastava 005), Akita, Jin and Wakaki 010), Srivastava and Reid 01), and so on. All these tests assume either normality or moderate dimensionality such that p/n c for a finite constant c or both assumptions. There has also been substantial research on the inference of hte covariance matrix under non-normal highdimensional distributions: see, for example, Chen, Zhang and Zhong 010), Srivastava, Kollo and Rosen 011), Zhong and Chen 011), Li and Chen 01), Qiu and Chen 01), Yata and Aoshima 013), Srivastava, Yanagihara and Kubokawa 014) and Himeno and Yamada 014). In this paper, we provide tests for covariance matrices without the normality assumption while allowing the dimension p to be much larger than the sample size N. Let x 1, x,..., x N N 4) be independent and identically distributed i.i.d.) p-dimensional random vectors with E[x i ] µ and Var[x i ] Σ. Now, we partition x i, µ, and Σ into q components: x 1) i µ 1) Σ 11 Σ 1 Σ 1q x ) ị x i., µ µ )., Σ Σ 1 Σ Σ q......, x q) µ q) Σ q1 Σ q Σ qq i where x g) i and µ g) are p g 1 vectors and Σ gh is a p g p h matrix, g, h 1,..., q. Note that p q g1 p g. Our interest is to test structure for covariance: H 0 : Σ Σ d vs. H A : Σ Σ d, 1.1) where Σ d diagσ 11, Σ,..., Σ qq ). When p N, to verify the block-diagonal sturucture is very important; if H 0 holds, then the total number of unknown parameters to be estimated for Σ reduces to q i1 p ip i +1)/ instead of pp+1)/. Further, it is interesting to note that the block-diagonal covariance structure attracts much attention across the fields of convex optimization and machine learning, with the main forcus being the Gaussian graph structure learning procedures.

3 For testing hypothesis 1.1), Hyodo et al. 015) proposed a test statistic based on the normalized Frobenious matrix norm under normal assumption. They used an estimator of trσ proposed by Bai and Saranadasa 1996) and Srivastava 005). Under normality assumption, this estimator has unbaisedness, but this estimator is not generally unbiased. Thus, in this paper, we propose a new test statistic based on an unbiased estimator proposed by Himeno and Yamada 014) and Srivastava, Yanagihara and Kubokawa 014) that does not need the normality assumption. Hypothesis 1.1) includes some important hypotheses. For p g 1, that is, for a test of the covariance being diagonal H 0 : Σ diagσ 1,, σ p )), Schott 005) and Srivastava 005) proposed an approximate test for a highdimensional diagonal matrix under the multivariate normal assumption. Besides, Chen, Zhang and Zhong 010) proposed approximate test without the normality assumption. Further, for q, that is, for a test of the independence of two sub-vectors H 0 : Σ diagσ 11, Σ )), Srivastava and Reid 01) proposed a test procedure under the multivariate normal assumption, while Yata and Aoshima 015) proposed a test procedure without the normality assumption. Our testing procedure extends the results derived by Chen, Zhang and Zhong 010) and Yata and Aoshima 015). The rest of the paper is organized as follows. Section presents test procedure after establishing the asymptotic normality of the test statistics, and derives asymptotic powers in Section 3. Further, in Section 4, the attained significance levels and powers of the suggested test are empirically analyzed. Finally, Section 5 concludes this paper. Some technical details are relegated to the Appendix.. Test procedure.1. Assumptions In this section, we mention the several assumptions to made construct a test for 1.1). Suppose x 1, x,..., x N are i.i.d. p-dimensional random vectors such that x i Γz i + µ for i 1,..., N,.1) where Γ Γ 1), Γ ),, Γ q) ) is a p m constant matrix so that ΓΓ Σ, and z 1, z,..., z N are i.i.d. m-dimensional random vectors such that E[z i ] 0 and Var[z i ] I m, an m m identity matrix. Note that x g) i Γ g) z i +µ g). Further, we use the following assumptions as necessary. 3

4 A1) Let z i z i1,..., z im ), with each z ij having a uniformly bounded 4th moment, and there existing finite constants κ 3, κ 4 such that E[zij] 3 κ 3, E[zij] 4 κ and for any positive integers r and α l such that α l 4, r l1 α l 8, E[z α 1 ij 1 z α ij z αr ij r ] E[z α 1 ij 1 ]E[z α ij ] E[z αr ij r ], whenever j 1, j,..., j r are distinct indices. A1 ) Let z i z i1,..., z im ), with each z ij having a uniformly bounded 8th moment, and there existing finite constants κ 3, κ 4 such that E[zij] 3 κ 3, E[zij] 4 κ 4 +3 and for any positive integers r and α l such that r l1 α l 8, E[z α 1 ij 1 z α ij z α r ij r ] E[z α 1 ij 1 ]E[z α ij ] E[z α r ij r ], whenever j 1, j,..., j r are distinct indices. A) We assume one of the following asymptotic frameworks: i) N, q is fixed and at least one of p 1,..., p q goes to infinity. ii) N, q and p 1,..., p q are fixed. A3) Let τ trσ ggtrσ hh. Then under A), g,h1 τ 1 min trσ4, A4) Under A), τ 1 τ 1 g g h h g g,g,h 1 g g h g A5) Under A), trσ 4 gg) 1/4 trσ 4 hh) 1/4 o1). g,h1 trσ gh Σ hg )trσ g h Σ h g ) o1), τ 1 Ntr Σ Σ d o1), trσ gg)trσ g h Σ h g ) o1). 1 N {trσ Σ d } trσ ghσ hg)trσ g h Σ h g ) o1). g g,h h 4

5 We state some remarks on the conditions. In assumption A1), the requirement of the common third and fourth moments of z ij, is not essential and is purely for the sake of simpler notation. Our theory allows different third and fourth moments as long as they are uniformly bounded, which is assured by z ij having a uniformly bounded fourth moment. We sometimes assume A1 ) instead of A1) to investigate the power properties of the test. Assumption A)-i) means that number of blocks is small but the number of dimensions is large. Conversely, assumption A)-ii) means that the number of blocks is small, but the matrix is large-dimensional. We discuss some specific examples of assumptions A3) and A4). Assumption A3) is used to derive the asymptotic null distribution, and assumption A4) is used to derive the asymptotic non-null distribution. Let us consider the case q. If q, H 0 means the hypothesis that testing the correlation coefficient matrix between x 1) i and x ) i. For this hypothesis, Yata and Aoshima 015) proposed a test statistic using the extended cross-data-matrix methodology and showed the asymptotic normality of this statistic under assumptions A1), C1) C) { } trσ 4 min gg g1, 0 as p, trσ gg lim sup m { N trσ 1 Σ 1 trσ 11 trσ } <, where m min{n, p}. Note that if C1) holds, min { trσ 4, trσ 4 11) 1/ trσ 4 ) 1/} τ trσ4 11) 1/ trσ 4 ) 1/ τ as p. We also note that if C1) and C) hold, 0 Ntr Σ Σ d ) τ 4Nλ maxσ 11 )λ max Σ )trσ 1 Σ 1 trσ 11trΣ 0 as p. Here, λ max A) denotes the largest eigenvalue of matrix A. From these results, A3) and A4) quite similar to but somewhat weaker than conditions C1) and C) formulated by Yata and Aoshima 015). Thus, we can compare our statistics with those proposed by Yata and Aoshima 015) under assumptions A1), C1), and C). We now consider specific types of the alternative covariance structure under p 1 p p q. The following covariance structure is called the blocked 5

6 compound symmetric BCS) structure: Σ I q Σ R) + 1 q 1 q R, where Σ is a p 1 p 1 positive definite symmetric matrix, and R is a p 1 p 1 symmetric matrix. Here, 1 q denotes a q-dimensional vector with all elements being one. If q p, Σ σ and R σ ρ, where 1/p 1) ρ 1, Σ σ {1 ρ)i p + ρ1 p 1 p}. This covariance structure is called the intraclass correlation structure. Note that τ qq 1)trΣ, trσ 4 gg) 1/4 trσ 4 hh) 1/4 q q 1 trσ 4, g,h1 Ntr Σ Σ d N{qq 1)trRΣ + qq 1)q )trr 3 Σ ) +qq 1)q 3q + 3)trR 4 }. Using the results for the correlation matrix in Yata and Aoshima 015), we get Suppose that trrσ λ max Σ ) trr, trr 3 Σ trrσ trr 4 λ max Σ ) trr, trr 4 λ max Σ ) trr. A3 ) A4 ) trσ 4 0 as p, trσ { } lim sup N trr m <. trσ If A)-i), A3 ), and A4 ) are assumed, A3) and A4) holds. Further, if we assume A)-ii) and r ij 1, Nq A3) and A4) hold. Here, a b means that a Ob) and b Oa), and r ij denotes each element of the correlation matrix R. 6

7 .. Test statistic We note that H 0 is valid if and only if Σ gh O for g h, and the latter implies that trσ Σ d 0. A test for the hypothesis H 0 vs. H A can be based on the unbiased estimator of trσ Σ d trσ trσ d and the same can be used to develop the test statistic. We assume.1), and let x 1 N N i1 x i, S 1 N 1 N x i x)x i x), Q 1 N x i x) x i x) ; N 1 i1 then, the unbiased estimator of trσ is trσ N 1 NN )N 3) {N 1)N )trs + trs NQ}. This estimator has been proposed in Himeno and Yamada 014) and Srivastava, Yanagihara and Kubokawa 014). From these results, it is straightforward to propose the unbiased estimator of trσ d as trσ d where and x g) 1 N N 1 NN )N 3) N i1 i1 {N 1)N )trsg + trs g NQ g }, g1 x g) i, S g 1 N 1 Q g 1 N 1 N i1 {x g) i N i1 x g) i x g) )x g) i x g) ) x g) ) x g) i x g) )}. From these facts, we construct the test statistic as follows: T trσ trσ d which is an unbiased estimator of trσ Σ d under.1). The following theorem shows that statistic T is the rate-consistent estimator of trσ Σ d. Theorem.1. Under assumptions A1), A3), and A5), T trσ Σ d p 1. Here, p denotes convergence in probability. Proof) The assertion follows immediately from Lemma A.3. 7

8 .3. Asymptotic normality of the test statistic We first introduce the asymptotic variance of T : σ 4 NN 1) trσ ghσ hg)trσ g h Σ h g ) + g g,h h + 4κ 4 N tr {Γ Σ Σ d )Γ} {Γ Σ Σ d )Γ}), 8 N tr ) Σ Σ d where denotes the Hadamard product, that is, A B) ij A) ij B) ij. See Lemma A.3 for details.) Remark.1. Under H 0 : Σ Σ d, σ H 0 4 NN 1) g,h1 trσ ggtrσ hh. Note that for any symmetric matrix A, tra A) tra. Hence, tr {Γ Σ Σ d )Γ} {Γ Σ Σ d )Γ}) tr Σ Σ d. Thus, under assumptions A1) - A4), σ σ H 0 1. The following theorems establish the asymptotic normality of T. Theorem.. Under assumptions A1) - A4), T trσ Σ d σ H0 d N 0, 1), where d denotes the convergence in distribution. Proof) See Appendix A.. To test H 0, it is necessary to estimate σ H 0. Using the unbiased estimator of trσ gg, trσ gg N 1 NN )N 3) {N 1)N )trs g + trs g NQ g }, 8

9 we estimate σh 0 by σ H 0 4 NN 1) g,h1 trσ gg trσ hh. The following theorem shows that σ H 0 σh 0. is the ratio-consistent estimator of Theorem.3. Under H 0 and assumptions A1) and A), Proof) See Appendix A.3. σ H 0 σ H 0 p 1. Applying Theorems. and.3, under H 0, T σ H0 d N 0, 1). Hence, we propose a test based on the test statistic T and reject H 0 if T > σ H0 z α, where z α is the upper critical value of the standard normal distribution at significance level α. 3. Asymptotic power In this section, we discuss the asymptotic power of our statistic T. We define the power of the test under H A : Σ Σ d as Power Y HN Pr T > σ H0 z α Σ Σ d ). First, we compare the asymptotic power of our statistic and the asymptotic power of Yata and Aoshima s statistic. From Theorems. and.3, we obtain the following corollary. Corollary 3.1. We assume q, C1), and C). Then, T trσ Σ d σ H0 d N 0, 1). 9

10 Remark 3.1. Let δ trσ 1 Σ 1 ). /{NN 1)}trΣ 11 trσ From Corollary 3.1, under q, A1), A), C1), and C), { 1 Φ z α δ 0 ) if δ δ 0 <, Power Y HN 1 if δ. From the above results and Theorem 4.1 in Yata and Aoshima 015), the power of statistic T when q is equivalent to that of Yata and Aoshima s statistic asymptotically. Second, we derive the asymptotic power of our test under a specific type of alternative covariance structure, the BSC structure. Let qq 1)NN 1)trR δ. trσ Using Theorem., under A1), A)-i), A3 ), and A4 ); or A1), A)-ii), and r ij 1/ nq, T qq 1)trR σ H0 d N 0, 1). From the above results, we obtain the asymptotic power of our statistic as follows: { 1 Φ z α Power Y HN δ ) 0 if δ δ 0 <, 1 if δ. Finally, we derive the asymptotic distribution of T when we cannot assume A4). Moreover, we consider the asymptotic power of our test under H A : Σ Σ d. To derive the asymptotic distribution of T, we assume assumption A1 ) instead of assumption A1). In assumption A1 ), the requirement of common third and fourth moments of z ij is not essential and is purely for the sake of simpler notation. Our theory allows different third and fourth moments as long as they are uniformly bounded, which is assured by z ij having a uniformly bounded eighth moment. The following theorem establishes the asymptotic normality of T. 10

11 Theorem 3.1. Under assumptions A1 ), A), and A3), Proof) See Appendix A.4. We define T trσ Σ d σ d N 0, 1). ν trσ Σ d, σ which may be viewed as a signal-to-noise ratio for the testing problem. This is because trσ Σ d is the square of the Frobenius norm of the difference between Σ and Σ d. The following corollary expresses the asymptotic power of the test more clearly. Corollary 3.. Under assumptions A1 ),A), and A3), { 1 Φ σ H0 σ Power Y HN z ) α ν 0 if ν ν0 <, 1 if ν. Thus, if the difference between Σ and Σ d is not too small in that trσ Σ d is of the same order as σ or of a larger order, the test will be powerful. Conversely, if the difference between Σ and Σ d is so small that trσ Σ d is of a smaller order than σ, the test will not be powerful and cannot distinguish H 0 from H A. 4. Simulation studies We now investigate the effectiveness of the proposed test statistic by numerical studies. We first provide the numerical evaluation by simulating the attained significant level ASL), or size of the test. ASL is defined by ASLα) T 0 > σ H0 z α ), r where T 0 is the value of the test statistic T calculated from N independent observations under the null hypothesis H 0, r is the number of replications, and z α is the upper 100α percentiles of the standard normal distribution. In 11

12 our simulation studies, we set N 10, 50, 100 and r We evaluate ASL under the diagonal structure: { Σ diag } 1,, {0.5 + p } ) 1, p + 1 p + 1 for p 100, 00, 400, 1000, and under the block diagonal structure: Here, Σ B0.3 i j 1/3 )B B Σ I q Σ Σ q ) q, 5) / d 1/ / d 1/3 0.3 d 1 1/3 0.3 d 1/3 1 B, where B diag{ /d + 1)} 1,, {0.5 + d/d + 1)} 1 ). Through simulations, for j 1,..., N, z j z ij ) in assumption A1) emerges for the following cases: Case1) z j N p 0, I p ), Case) z ij u ij 1)/ for u ij χ 1, Case3) Further, we set α 0.05 and z ij SN1), Case4) z ij u ij / 3 for u ij t 3, Case5) z ij u ij / 5 for u ij t 5. q, d), 50),, 100),, 00),, 500), 5, 0), 5, 40), 5, 80), 5, 00). Tables 1-3 give the results for ASL. Overall, it may be noted that even if power is low for N 10, when the values of p and N become large, approximation accuracy becomes good. The procedures proposed by Srivasta 005) and Srivasta and Reid 01) need the assumption of normality. Thus, these procedures are not helpful for Cases -5. However, we note that the accuracy of approximation for our procedure is almost the same for Cases 1-5. Further, because Yata and Aoshima 015) provide a procedure under 1

13 HDLSS framework, for this simulation setting, we observe that our procedure is better than that of Yata and Aoshima 015). Further, we compute empirical power EP) under certain alternative covariance structure. EP is defined as EPα) T 1 > σ H0 z α Σ Σ d ), r where T 1 is the value of test statistic T calculated from N independent observations under the alternative hypothesis H 1. When H 0 : Σ is a diagonal covariance, we set the alternative hypothesis as follows: { i) H 1 : Σ σ ij ), σ ij δ ij i } I i j 0), p + 1 for i, j 1,,..., p, and { ii) H 1 : Σ diag } 1,, {0.5 + p } ) p 1 p I p ). p + 1 p + 1 When H 0 : Σ Σ q q, 5), we set the following alternative hypothesis: i) H 1 : Σ I q Σ d 1 d) + 1 q 1 q) d 1 d), and ii) H 1 : Σ I i j 1)) d 1 d) + I q Σ d 1 d)). Tables 4-9 give the results for EP. As a whole, it may be noted when the values of p and N become large, power becomes high. However, for case ii) and q, power becomes low when p Besides, for q, we observe that the power of our procedure is almost the same as that of Yata and Aoshima 015) and Srivasta and Reid 01) for Case 1 and 3. Further, we note that the power for q is higher than that for q 5 in the case i), and the power for q 5 is higher than that for q in case ii). When H 0 : Σ is a diagonal covariance, we observe that the power of our procedure is almost the same as that of Srivasta 005) for Case 1 and 3. The proposed test are slightly more powerful than Srivasta 005) in Case, 4 and 5. 13

14 5. Concluding remarks In this paper, we discuss the test for a block-diagonal covariance structure of the covariance matrix in non-normal and high-dimensional settings. This hypothesis includes some important hypotheses, for example, when each block size is one, that is, in a test for the covariance being diagonal, and when there are two blocks, that is, in a test for the independence of two subvectors. Further, the validity of commonly used high-dimensional statistical procedures including diagonal or block diagonal) linear classifiers requires the assumption of diagonal or block diagonal) covariance matrices. A test for the block-diagonal covariance structure of the covariance matrix usually used the likelihood ratio test. Akita, Jin and Wakaki 010) derived the Edgeworth expansion of a likelihood ratio test statistic when both sample size and dimension are large, and simulate the goodness of fit of its approximation. However, the likelihood ratio cannot use p > n. Motivated by an unbiased estimator of the Frobenius norm of Σ Σ d, Hyodo et al. 015) proposed an approximate test under a multivariate normal population. The difference between our study and theirs lies in the population distribution. Under the non-normal population, we proposed an approximate test based on the asymptotic normality of the unbiased estimator of Σ Σ d that does not need the normality assumption, and also studied its asymptotic power. The type-i error and power of our test have been investigated under some non-normal settings by simulation. Through the simulation results, we have confirmed that the approximate test is not bad in most cases except for in the case of N 10. However, some multivariate non-normal distributions will not satisfy assumption A1) or A1 )). For instance, elliptical distributions except for the multivariate normal model seldom satisfy assumption A1) or A1 )). For these reasons, it is interesting to investigate the robustness of our test under elliptical distributions. This is an important topic for future research that is of both theoretical and practical interest. Acknowledgments. The research of the third author was supported in part by a Grant-in-Aid for Young Scientists B673000) from the Japan Society for the Promotion of Science. 14

15 References [1] T. Akita, J. Jin, H. Wakaki, High-dimensional Edgeworth expansion of a test statistic on independence and its error bound. J. Multivariate Anal., ) [] Z. Bai, H. Saranadasa, Effect of high dimension: by an example of a two sample problem. Statist. Sinica, ) [3] S. X. Chen, L. X. Zhang, P. S. Zhong, Tests for high-dimensional covariance matrices. J. Amer. Statist. Assoc., ) [4] T. Himeno, T. Yamada, Estimations for some functions of covariance matrix in high dimension under non-normality and its applications. J. Multivariate Anal., [5] M. Hyodo, N. Shutoh, T. Nishiyama, T. Pavlenko, Testing blockdiagonal covariance structure for high-dimensional data. Stat. Neerl., ) [6] J. Li, S. X. Chen, Two sample tests for high-dimensional covariance matrices. Ann. Statist., 40 01) [7] Y. Qiu, S. X. Chen, Test for bandedness of high-dimensional covariance matrices and bandwidth estimation. Ann. Statist., 40 01) [8] J. R. Schott, Testing for complete independence in high dimensions. Biometrika, 9 005) [9] M. S. Srivastava, Some tests concerning the covariance matrix in highdimensional data. J. Japan Statist. Soc., [10] M. S. Srivastava, T. Kollo, D. von Rosen, Some tests for the covariance matrix with fewer observations than the dimension under non-normality. J. Multivariate Anal., ) [11] M. S. Srivastava, N. Reid, Testing the structure of the covariance matrix with fewer observations than the dimension. J. Multivariate Anal., 11 01)

16 [1] M. S. Srivastava, H. Yanagihara, T. Kubokawa, Tests for covariance matrices in high dimension with less sample size. J. Multivariate Anal., [13] K. Yata, M. Aoshima, High-dimensional inference on covariance structures via the extended cross-data-matrix methodology. Submitted arxiv: ), 015). [14] K. Yata, M. Aoshima, Correlation tests for high-dimensional data using extended cross-data-matrix methodology. J. Multivariate Anal., ) [15] P. S. Zhong, S. X. Chen, Tests for high-dimensional regression coefficients with factorial designs. J. Amer. Statist. Assoc.,

17 A. Appendix A.1. Preliminary results Lemma A.1. Let A, B, C and D are m m symmetric matrices and E, F are m m matrices. Under assumption A1), we have: i) E[z 1Ez 1 z 1F z 1 ] κ 4 tre F ) + tretrf + tref ) + tref ), ii) E[z 1Az z 1Bz z 1Ez ] κ 3tr A B)E ), iii) iv) E[z 1Az z 1Bz z 1Cz z 1Dz ] κ 4tr A B)C D)) + κ 4 tr AB) CD)) + κ 4 tr AC) BD)) + κ 4 tr AD) BC)) + trab)trcd) + trac)trbd) + trad)trbc) + trabcd) + trabdc) + tracbd) + tracdb) + tradbc) + tradcb), m ) 4 m E a ij z 1i z 1j const. aij. i j Proof) See supplemental material. Lemma A.. Let Γ g) Γ g) A g) a g) ij ) and q g1 Ag) Γ Γ A a ij ). It holds that i) tr A g) A h) ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4, 17

18 ii) iii) iv) v) vi) tra g) A g ) A h) A h ) ),g h ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4, tr g g,h h A g) A h)) A g ) A h ) )) ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4, ) tr A g) A g ) )A h ) A h),g h ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, 4trΣ 4, tr g 1 h 1,g h g 3 h 3,g 4 h 4 A g 1 ) A g 3) ) A h 1) A h 3) ) ) A g ) A g 4) ) A h ) A h 4) ) ) ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, 18trΣ 4, g tra 1 ) A g) A g3) A g4) ) g 1 h 1,g h g 3 h 3,g 4 h 4 tra h1) A h) A h3) A h4) ) ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, 18trΣ 4, 18

19 vii) viii) tr A g) Γ Σ Σ d )ΓA h) ) min 4trΣ4, trσ 4 gg) 1/4 trσ 4 hh) 1/4 trσσ Σ d), { [diagγσ Σ d )Γ)] A g) A h)) g g,h h )} A g ) A h ) [diagγσ Σ d )Γ)] ) min trσ4, trσ 4 gg) 1/4 trσ 4 hh) 1/4 trσσ Σ d). Here, [diaga)] a 11, a,, a mm ) for m m matrix A a ij ). Proof) See supplemental material. Lemma A.3. Under assumptions A1)-A3), Proof) Let U 1 U U 3 E[T ] trσ Σ d, Var[T ] σ 1 + o1)). 1 NN 1) g,h1 N z ia g) z j z ia h) z j, i j 1 NN 1)N ) g,h1 1 NN 1)N )N 3) N i,j,k1 i j k i g,h1 z ia g) z j z ia h) z k, N i,j,k,l1 i j k l j l i k z ia g) z j z ka h) z l. Then T can be expressed as T U 1 U +U 3. It is straightforward to show that E[T ] E[U 1 ] trσ Σ d. By using Lemma A.1, Var[U 1 ] is calculate as Var[U 1 ] 19 6 i1 V 1) i,

20 where V 1) 1 4 NN 1) 4 NN 1) tra g g,h h ) V 1) 8 N tr A g) A h) g) A h) )tra g ) A h ) ) trσ ghσ hg)trσ g h Σ h g ), g g,h h 8 N tr Σ Σd, ) )) V 1) 3 4κ 4 N tr A g) A h) A g) A h) 4κ 4 N tr ΓΣ Σ d)γ ΓΣ Σ d )Γ), V 1κ 4 4 NN 1) tr A g) A h), V 1) 5 V 1) 6 8κ 4 NN 1) 4 NN 1) tr g g,h h tra,g h A g) A h)) )) A g ) A h ), g) A g ) A h) A h ) ). From Lemma A., V 1) 4, V 1) 5 and V 1) 6 are negligible under A1)-A3). Hence, Var[U 1 ] σ 3 t1 V 1) t σ + o1) A.1) under A1)-A3). From A.1) and 3 t1 V 1) t σ, we have Var[U 1 ] σ 1. A.) 0

21 By using Lemma A.1, Var[U ] is calculate as Var[U ] 5 i1 V ) i, where V ) 1 V ) V ) 3 V ) 4 V ) 5 NN 1)N ) NN 1)N ) tra g g,h h g) A h) )tra g ) A h ) ) trσ ghσ hg)trσ g h Σ h g ), g g,h h NN 1) tr A g) A h), NN 1) tr Σ Σ d κ 4 NN 1)N ) 4κ 3 NN 1)N ) NN 1)N ) tr g g,h h tr,g h tra,g h A g) A h)) )) A g ) A h ), A g) A g ) )A h ) A h) ), g) A g ) A h) A h ) ). From Lemma A., under A1)-A3), and V ) 3 σ 0, V ) 4 σ 0, V ) 5 σ 0, A.3) V ) 1 V 1) 1 0, V ) V 1) 1 0 A.4)

22 as N. From A.3) and A.4), under A1)-A3), Var[U ] σ By using Lemma A.1, Var[U 3 ] is calculate as Var[U 3 ] 5 t1 V ) t σ 0. A.5) 16 N N 1 N N 3 N E[z ia g1) z j z ka h1) z l z ia g) z k z ja h) z l ] g 1,h 1,g,h 1 i,j,k,l1 g 1 h 1,g h i j k l j l i k 8 + N N 1 N N 3 N E[z ia g1) z j z ka h1) z l z ia g) z j z ka h) z l ] g 1,h 1,g,h 1 i,j,k,l1 g 1 h 1,g h i j k l j l i k 8 NN 1)N )N 3) ) +tra g) A g ) )tra h) A h ) ).,g h From above result, under assumptions A1)-A3), tra g) A g ) A h) A h ) ) Var[U 3 ] σ 0. A.6) Combining A.), A.5) and A.6), Var[T ] σ {1 + o1)}. Lemma A.4. Under assumptions A1)-A4), it holds that T trσ Σ d 1 NN 1) N i j k 1 k,l 1 l β k1 l 1 k l z ik1 z ik z jl1 z jl + o p σ H0 ), where β k1 l 1 k l a g) k 1 l 1 a h) k l.

23 And under assumptions A1)-A3), it holds that T trσ Σ d 1 NN 1) + N N j1 N i j k 1 k,l 1 l β k1 l 1 k l z ik1 z ik z jl1 z jl { z j Γ Σ Σ d )Γz j tr )} Σ Σ d + op σ). Proof) From A.5) and A.6), U and U 3 are negligible under A1)-A3). So, we only consider U 1 in T trσ Σ d. Let U 11 U 1 N U 13 N U 14 U 15 1 NN 1) N j1 k,l1 N j1 N i j β k1l1klz ik1z ikz jl1z jl, k 1,k,l 1,l 1 k 1 k,l 1 l β klkl zjl 1), k,l 1,l 1 l 1 l 1 NN 1) NN 1) N β kl1klz jl1z jl, i j k,l1 N i j β klkl zik 1)zjl 1), k 1,k,l1 k 1 k β k1lklz ik1z ikzjl 1). Then these random variables are uncorrelated, namely, CovU 1t1, U 1t ) 0 for t 1 t, and U 1 can be expressed as U 1 trσ Σ d + 5 U 1t. At first we show first statement in Lemma A.4. For the proof of first statement, we need to show that Var[ 5 t U 1t] σ H t1

24 under A1)-A4). The variance of U 11 is caluclated as ) Var[U 11 ] NN 1) E β k1 l 1 k l z ik1 z ik z jl1 z jl k 1 k,l 1 l 4 NN 1) k 1 k,l 1 l β k 1 l 1 k l + β k1 l 1 k l β k l 1 k 1 l ) 4 NN 1) + 4 tra g g,h h tr g g,h h tra g g,h h g) A h) )tra g ) A h ) ) g) A h) A g ) A h ) ) + tr A g) A h) A g) A h)) A g ) A h ) )). Thus under A1)-A4), Var[U 11 ] V 1) 1 We also note that under A1)-A4), 1. A.7) Var[U 1 ] V 1) 1 1, V 1) 1 σ H 0 1. A.8) From A.7) and A.8), under A1)-A4), Var[ 5 t U 1t] σ H 0 Var[U 1] Var[U 11 ] V 1) 1 This proves the first statement of the lemma. 4 V 1) 1 0. σh 0

25 Next we show first statement in Lemma A.4. For the proof of first statement, we need to show that under A1)-A3). Note that Var[ 5 t4 U 1t] σ 0 U 1 + U 13 N N j1 { z j Γ Σ Σ d )Γz j tr )} Σ Σ d. By using Lemma A.1, the variance of U 1 + U 13 is caluclated as [ Var[U 1 + U 13 ] 4 N Var N [z j{γ Σ Σ d )Γ}z j tr ] ) Σ Σ d ] Thus under A1)-A3), j1 4κ 4 N tr[{γ Σ Σ d )Γ} {Γ Σ Σ d )Γ}] + 8 N tr ) Σ Σ d. Var[ 3 t1 U 1t] σ Var[U 11] + Var[U 1 + U 13 ] σ 1. A.9) We also note that under A1)-A3), Var[U 1 ] σ 1. A.10) From A.9) and A.10), under A1)-A3), Var[ 5 t4 U 1t] Var[U 1] Var[ 3 t1 U 1t] 0. σ σ This proves the second statement of the lemma. A.. Proof of Theorem.. From Lemma A.4, Under assumptions A1)-A4), the leading order term in Var[T ] is contributed by U 11. Hence, we only need to show the asymptotic normality of U 11. Let ε j j 1 β k1 l NN 1) 1 k l z ik1 z ik z jl1 z jl. i1 k 1 k,l 1 l 5

26 Then U 11 N j1 ε j. Define F 0 {, Ω}, F j σ{z 1, z,..., z j } for j 1,..., N be the σ-algebra generated by the sequence of random vectors z 1, z,..., z j. Then it is straightforward to show that E[ε j ] 0 and E[ε j F j 1 ] 0. So ε j is a martingale difference sequence. To apply martingale central limit theorem, we need to show that N j1 E[ε j F j 1 ] p 1 and σh 0 N j1 E[ε4 j] σ 4 H 0 0. A.11) We first show the first part of A.11). Note that E[ε j F j 1 ] 4 N N 1 j 1 i 1,i 1 k 1 k,k 3 k 4 γ k1 k k 3 k 4 z i1 k 1 z i1 k z i k 3 z i k 4, where γ k1 k k 3 k 4 l 1 l β k1 l 1 k l β k3 l 1 k 4 l + β k1 l 1 k l β k3 l k 4 l 1. To show the first part of A.11) we decompose N j1 E[ε j F j 1 ] into sum of ten parts N σj Var[U 11 ] + j1 9 R t, t1 where R 1 R R 3 R 4 8 N N 1 8 N N 1 8 N N 1 8 N N 1 j 1 N j γ k1 k k 1 k zik 1 1)zik 1), i1 k 1 k j 1 N γ k1 k k 1 k zik 1 1), i1 k 1 k j 1 N γ k1 k k 1 k zik 1), i1 k 1 k j 1 N γ k1 k k 1 k 3 zik 1 1)z ik z ik3, i1 k 1 k k 3 k 1 j j j 6

27 R 5 R 6 R 7 R 8 R 9 8 N N 1 8 N N 1 8 N N 1 4 N N 1 8 N N 1 j 1 N j γ k1 k k 1 k 3 z ik z ik3, i1 k 1 k k 3 k 1 j 1 N γ k1 k k k 3 zik 1)z ik1 z ik3, i1 k 1 k k 3 k 1 j 1 N γ k1 k k k 3 z ik1 z ik3, i1 k 1 k k 3 k 1 j 1 N γ k1kkk3z ik1z ikz ik3z ik4, i1 k 1 k k 3 k 4 k k 4 k 1 k 3 j 1 N γ k1 k k k 3 z i1 k 1 z i1 k z i k 3 z i k 4. k 1 k,k 3 k 4 j j j j i 1 <i Then we need to show E[Rt ] oσh 4 0 ) for t 1,..., 9 since Var[U 11 ]/σh 0 1. By applying Cauchy-Schwarz inequality, we obtain E[R1] 3κ 4 + γ N N 1) 3 k 1 k k 1 k + γ k1 k k 1 k γ k k 1 k k 1 ) k 1 k 64κ 4 + γ N N 1) 3 k 1 k k 1 k k 1 k 64κ 4 + m β N N 1) 3 k 1 l 1 k l + β k1 l 1 k l β k1 l k 1 l 1 k 1 k l 1 l l 1 l 56κ 4 + m N N 1) 3 56κ 4 + N N 1) 3 56κ 4 + N 3 N 1 k 1 k l 1 l β k 1 l 1 k l k 1,k,l 1,l 1 β k 1 l 1 k l trσ ghσ hg)trσ g h Σ h g ) g g,h h 7.

28 Simiraly, we obtain E[R] 51κ 4 + ) N 3 N 1 E[R 3] 51κ 4 + ) N 3 N 1 E[R 4] 56κ 4 + κ 3 + ) N 3 N 1 E[R 5] 560 N 3 N 1 E[R 6] 56κ 4 + κ 3 + ) N 3 N 1 E[R 7] E[R 8] 560 N 3 N N 3 N 1 trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ), g g,h h trσ ghσ hg)trσ g h Σ h g ). g g,h h From these results and V 1) 1 σ 4, under A1)-A3), E[R t ] σ 4 E[R t ] V 1) 1 ON 1 ) t 1,..., 8). A.1) 8

29 Also, we evaluate E[R 9]. The expectation of R 9 is evaluated as following: E[R9] 3N ) γ 3N N 1) 3 k 1 k k 3 k 4 + γ k1 k k 3 k 4 γ k1 k k 4 k 3 k 1 k,k 3 k 4 +γ k1 k k 3 k 4 γ k k 1 k 3 k 4 + γ k1 k k 3 k 4 γ k k 1 k 4 k 3 ) 18N ) γ 3N N 1) 3 k 1 k k 3 k 4 k 1,k,k 3,k N ) m β 3N N 1) 3 k1 l 1 k l β k3 l 1 k 4 l + β k1 l 1 k l β k3 l k 4 l 1 k 1,k,k 3,k 4 1 l 1 l 56N ) m β 3N N 1) 3 k1 l 1 k l β k3 l 1 k 4 l k 1,k,k 3,k 4 1 l 1 l 56N ) m + β 3N N 1) 3 k1 l 1 k l β k3 l k 4 l 1 k 1,k,k 3,k 4 1 l 1 l 51N ) m β 3N N 1) 3 k1 l 1 k l β k3 l 1 k 4 l k 1,k,k 3,k 4 1 l 1,l 1 51N ) m + β 3N N 1) 3 k1 l 1 k l β k3 l k 4 l 1 k 1,k,k 3,k 4 1 l 1,l 1 104N ) m + β 3N N 1) 3 k1 lk lβ k3 lk 4 l k 1,k,k 3,k 4 1 l1 104N ) g tra 1 ) A g) A g3) A g4) )tra h1) A h) A h3) A h4) ) 3N N 1) 3 g 1 h 1,g h g 3 h 3,g 4 h 4 104N ) + 3N N 1) 3 A g 1 ) A g3) ) A h1) A h3) ) ) A g) A g4) ) A h) A h4) ) ). tr g 1 h 1,g h g 3 h 3,g 4 h 4 9

30 From Lemma A., under A1)-A3), E[R 9] σ 4 o1). A.13) From A.1), A.13) and σ /σ H 0 1 under A1)-A4), we obtain E[R t ] oσ 4 H 0 ). This proves first part of A.11). Next we show the second part of A.11). There exists constants c 1 and c such that E[ε 4 j] c 1j 1)j ) m ) E β N 4 N 1) 4 k1 l 1 k l z jl1 z jl k 1,k 1 l 1 l c j 1)j ) N 4 N 1) 4 trσ ghσ hg)trσ g h Σ h g ). g g,h h From the above results, N j1 E[ε4 j] σ 4 N j1 E[ε4 j] V 1) 1 From A.14) and σ /σ H 0 1 under A1)-A4), we obtain N E[ε 4 j] oσh 4 0 ). j1 ON 1 ). A.14) This proves second part of A.11) and completes the proof of Theorem.1. A.3. Proof of Theorem.3. Under assumptions A1) and A), it holds that Hence trσ gg trσ gg1 + O p N 1/ )). σ H 0 σ H O p N 1/ )) 30

31 and it hold σ H 0 σ H O p N 1/ ). From the above results, it completes the proof of Theorem.. A.4. Proof of Theorem 3.1. From Lemma A.4, under assumptions A1 )-A3), the leading order term in Var[T ] is contributed by U 11 + U 1 + U 13. Hence, we only need to show the asymptotic normality of U 11 + U 1 + U 13. Let ψ j ε j + N { z j Γ Σ Σ d )Γz j tr )} Σ Σ d. Then N j1 ψ j U 11 + U 1 + U 13, and it is straightforward to show that E[ψ j ] 0 and E[ψ j F j 1 ] 0. So ψ j is a martingale difference sequence. To apply martingale central limit theorem, we need to show that N j1 E[ψ j F j 1 ] p 1 and σ N j1 E[ψ4 j ] σ 4 0. A.15) We first show the first part of A.15). To show the first part of A.15) we decompose N j1 E[ψ j F j 1 ] into sum of ten parts E[ψ j F j 1 ] Var[U 11 ] + V 1) + V 1) 3 + where R t t 1,,..., 9) is defined in A.1), and 10 t1 R t, R 10 j 1 8 N N 1) i1 k 1,k,k 3 1 k 1 k γ k1kk3k3z ik1z ik. Then we need to show E[R t ] oσ 4 ) for t 1,..., 10 since under A1 )-A3), Var[U 11 ] + V 1) + V 1) 3 1. σ From A.1) and A.13), under A1 )-A3), E[R t ] σ 4 o1) t 1,,..., 9). A.16) 31

32 Also, we evaluate E[R10]. The expectation of R10 is evaluated as following: E[R10] 3 m m ) m ) γ N N 1) k1 k k 3 k 3 + γ k1 k k 3 k 3 γ k k 1 k 3 k 3 k 1 k k 3 1 k 3 1 k m γ N k1 k N 1) k 3 k 3 k 1 k k m β N k1 l N 1) 1 k l β k3 l 1 k 3 l β k1 lk lβ k3 lk 3 l k 1 k l 1,l,k 3 1 l,k m β N k1 l N 1) 1 k l β k3 l 1 k 3 l k 1,k 1 l 1,l,k m + β N k1 lk N 1) lβ k3 lk 3 l k 1,k 1 l,k N N 1) tr A g) Γ Σ Σ d )ΓA h) 56 + N N 1) [diagγσ Σ d )Γ)] A g) A h)) A i) A j)) diagγσ Σ d )Γ). g g,h h From above result and Lemma A., E[R10] 51 ) N N 1) min 4trΣ4, trσ 4 gg) 1/4 trσ 4 hh) 1/4 trσ Σ d. From above result, under A1 )-A3), E[R 10] σ 4 o1). A.17) Combining A.16) and A.17), we prove the first part of A.15). 3

33 Next we show the second part of A.15). By applying Hölder s inequality, E[ψj 4 ] 8E[ε 4 j] + 18 [ z N E 4 j Γ Σ Σ d )Γz j tr )) ] Σ Σ 4 d. A.18) Note that for any symmetric matrix A, E[z 1Az 1 tra) 4 ] const. {tra }. Hence, there exists a constant c 3 such that E[ { z jγ Σ Σ d )Γz j trσ Σ d) } 4 ] c3 tr Σ Σ d. A.19) From A.18) and A.19), N j1 E[ψ4 j ] oσ 4 ). This proves second part of A.15) and completes the proof of Theorem.3. 33

34 Table 1: Attained significant level in the case of diagonal matrix. our procedure p N Case 1 Case Case 3 Case 4 Case Srivastava 005)

35 Table : Attained significant level in the case of q. our procedure p N Case 1 Case Case 3 Case 4 Case Yata and Aoshima 015) , Srivastava and Reid 01)

36 Table 3: Attained significant level in the case of q 5. p N Case 1 Case Case 3 Case 4 Case Table 4: Empirical power in the case of diagonal matrix for i). our procedure p N Case 1 Case Case 3 Case 4 Case Srivastava 005)

37 Table 5: Empirical power in the case of diagonal matrix for ii). our procedure p N Case 1 Case Case 3 Case 4 Case Srivastava 005)

38 Table 6: Empirical power in the case of q for i). our procedure p N Case 1 Case Case 3 Case 4 Case Yata and Aoshima 015) Srivastava and Reid 01)

39 Table 7: Empirical power in the case of q for ii). our procedure p N Case 1 Case Case 3 Case 4 Case Yata and Aoshima 015) Srivastava and Reid 01)

40 Table 8: Empirical power in the case of q 5 for i). p N Case 1 Case Case 3 Case 4 Case Table 9: Empirical power in the case of q 5 for ii). p N Case 1 Case Case 3 Case 4 Case

41 Supplemental Material:Testing Block-Diagonal Covariance Structure for High-Dimensional Under Non-normality Yuki Yamada, Masashi Hyodo and Takahiro Nishiyama March 30, 016 Abstract The supplement provides full proof of Lemma A.1 and, which are omitted in the original paper. 1 Proof of Lemma A.1 First, we show i). E[z 1Ez 1 z 1F z 1 ] E e ii z1i + i1 e ii f ii E[z1i] 4 + i1 i j e ij z 1i z 1j f ii z1i + i1 e ii f jj E[z1iz 1j] + i j f ij z 1i z 1j i j e ij f ij + e ij f ji )E[z1iz 1j] i j κ 4 tre F ) + tretrf + tref ) + tref ). Second, we show ii). [ m ) m ) m )] E[z 1Az z 1Bz z 1Ez ] E a ij z 1i z j b ij z 1i z j e ij z 1i z j a ij b ij e ij E[z1iz 3 j] 3 κ 3tr A B)E ) Department of Mathematical Information Science, Tokyo University of Science, 1-3, Kagurazaka, Shinjuku-ku, Tokyo, , Japan, y.y.rfpfg.t.o.d.006.s.m.e@gmail.com. Department of Mathematical Sciences, Graduate School of Engineering, Osaka Prefecture University, 1-1 Gakuen-cho, Naka-ku, Sakai, Osaka , Japan. hyodo@ms.osakafu-u.ac.jp Department of Business Administration, Senshu University, -1-1, Higashimita, Tama-ku, Kawasakishi, Kanagawa , Japan. 1

42 Third, we show iii). E [z 1Az z 1Bz z 1Cz z 1Dz ] [ m ) m ) m ) m )] E a ij z 1i z j b ij z 1i z j c ij z 1i z j d ij z 1i z j E a ij b ij z1iz j + + i,j,k,l1 i j,k l + i,j,k1 i j i,j,k1 j k a ij b ik z 1iz i z k + i,j,k1 i j a ij b kl z 1i z 1j z k z l c ij d ij z1iz j + c ik d jk z 1i z 1j z k + a ij b ij c ij d ij E[z1iz 4 j] i,j,k1 j k i,j,k1 j k i,j,k,l1 i j,k l i,j,k1 i j i,j,k,l1 i j,k l i,j,k,l1 i j,k l i,j,k,l1 i j,k l i,j,k,l1 i j,k l i,j,k1 i k a ij b ij c ik d ik E[z 4 1iz jz k] + a ij b ik c ij d ik E[z 4 1iz jz k] + a ik b jk z 1i z 1j z k i,j,k1 j k c ij d kl z 1i z 1j z k z l a ij b ij c kj d kj E[z 1iz 1kz 4 j] i,j,k,l1 i k,j l i,j,k1 j k a ik b il c jk d jl E[z 1iz 1kz jz l] + a ik b jk c ik d jk E[z 1iz 1jz 4 k] + a ik b jk c il d jl E[z 1iz 1jz kz l] + a ik b jl c ik d jl E[z 1iz 1kz jz l] + a ik b jl c jk d il E[z 1iz 1kz jz l] + c ij d ik z 1iz i z k a ij b ij c kl d kl E[z 1iz 1kz jz l] a ij b ik c ik d ij E[z 4 1iz jz k] i,j,k,l1 i j,k l i,j,k1 i j a ik b il c jl d jk E[z 1iz 1kz jz l] a ik b jk c jk d ik E[z 1iz 1jz 4 k] i,j,k,l1 i j,k l i,j,k,l1 i j,k l i,j,k,l1 i j,k l a ik b jk c jl d il E[z 1iz 1jz kz l] a ik b jl c il d ik E[z 1iz 1kz jz l] a ik b jl c jl d ik E[z 1iz 1kz jz l]

43 κ 4 a ij b ij c ij d ij + κ 4 +κ 4 m + + i,j,k1 i,j,k,l1 i,j,k,l1 a ij b ik c ik d ij + a ik b jk c il d jl + a ik b jl c il d ik + m i,j,k1 i,j,k,l1 i,j,k,l1 i,j,k,l1 a ij b ij c kj d kj + κ 4 a ij b ij c kl d kl + a ik b jk c jl d il + a ik b jl c jk d il + m i,j,k1 i,j,k,l1 i,j,k,l1 i,j,k,l1 a ij b ik c ij d ik a ik b il c jk d jl + a ik b jl c ik d jl a ik b jl c jl d ik i,j,k,l1 κ 4tr A B)C D)κ 4 tr AB) CD)) + κ 4 tr AC) BD)) +κ 4 tr AD) BC)) + trab)trcd) + trac)trbd) + trad)trbc) a ik b il c jl d jk +trabcd) + trabdc) + tracbd) + tracdb) + tradbc) + tradcb). Finally, we show iv). m ) 4 [ E a ij z 1i z 1j E i j + a ijz1iz 1j + 4 i j i j k l j l i k i j k i a ij a kl z 1i z 1j z 1k z 1l a ij a ik z 1iz 1j z 1k m ) m 1E a ij z1iz 1j + 48E +3E i j i j k l j l i k a ij a kl z 1i z 1j z 1k z 1l i j k i ) a ij a ik z1iz 1j z 1k. 1.1) The each term on right hand side in 1.1) is evaluated as m E a ij a kl z 1i z 1j z 1k z 1l 8a ija kl + 16a ij a kl a ik a jl )E[z1iz 1jz 1kz 1l] i j k l j l i k 4 i j k l j l i k i j k l j l i k a ija kl m 4 aij, 1.) 3 1.3)

44 m E i j k i ) a ij a ik z1iz 1j z 1k +4 + i j k i i j k i i j k l j l i k κ 4 + 3) + i j k l j l i k a ija ike[z 4 1iz 1jz 1k] a ij a ik a ji a jk E[z 3 1iz 3 1jz 1k] a ij a ik a lj a lk E[z 1iz 1jz 1kz 1l] i j k i a ij a ik a lj a lk a ija ik + 4κ 3 i j k i a ij a ik a ji a jk m κ 4 + 4κ 3 + 8) aij, 1.4) and m E i j a ijz 1iz 1j + a 4 ije[z1iz 4 1j] i j i j k l j l i k i j k i a ija kle[z 1iz 1jz 1kz 1l] a ija ike[z 4 1iz 1jz 1k] { κ κ 4 + 3) + 1 } m aij. 1.5) Combining 1.1)-1.4), we obtain iv). 4

45 Proof of Lemma A. First, we show i). By applying Cauchy-Schwarz inequality tr A g) A h),g h,g h m a g ) 4 ij,g h a g) ij ah) ij m m a g) ij a h) ij a g)4 ij ) 1/4 m ) 1/ m ) 1/4 m a h ) 4 ij ) 1/4 a h)4 ij a g ij a h ij ) 1/4 ) 1/ tra g)4) 1/4 tra h)4) 1/4 tra g ) 4) 1/4 tra h ) 4) 1/4 trσ 4 gg) 1/4 trσ 4 hh) 1/4..1) On the other hand, we note that tr A g) A h) since g1 g) aij ah) ) ij ag ij a h ) ij,g h a 4 ij g1 a ija g) ij + g1 a g) ij trσ 4,.) a 4 ij tra A tra A ) trσ 4, a g) ij g,h1 g,h1 tr A g) A g) )A h) A h) ) ) tra g) A h) ) trγ Σ d Γ Γ Σ d Γ) trσσ d trσ Σ d trσ 4 ) 1/ trσ 4 d) 1/ trσ 4. 5

46 From.1) and.), we obtain ) ) tr A g) A h) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4. Second, we show ii). By applying Cauchy-Schwarz inequality tra g) A g ) A h) A h ),g h On the other hand, we note that g) tra A g ) 1/ tra h) A h ) 1/,g h trσ 4 gg) 1/4 trσ 4 hh) 1/4..3) tra g) A g ) A h) A h ),g h tra 4 tra g) A + g1 tra 4 + tra g) A h) g,h1 tra g) A h) g,h1 trσ 4,.4) since tra g) A h) tra g) A h) trσσ d trσ 4. g,h1 g,h1 From.3) and.4), we obtain ) tra g) A g ) A h) A h ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4.,g h 6

47 Third, we show iii). By applying Cauchy-Schwarz inequality A g) A h)) )) m A g ) A h ) tr g g,h h On the other hand, we note that A g) A h)) )) A g ) A h ) tr g g,h h since tra A ) trσ 4, tr A g) A h) ) A g) A h) ) ) g,h1 From.5) and.6), we obtain tr g g,h h A g) A h)) )) A g ) A h ) g 1,h 1,g,h 1 g 1 h 1,g h g 1,h 1,g,h 1 g 1 h 1,g h i1 m i1 g 1,h 1,g,h 1 g 1 h 1,g h m j1 i1 j1 j1 a g 1) ij a g ) ij ) m j1 m ) a g 1) ij a g ) ij m a h 1) ij a h ) ij a h 1 ij a h ij a g 1 ij a g ij ) 1/ 1/ ) 1/ a h 1) ij a h ) ij 1/ trσ 4 gg) 1/4 trσ 4 hh) 1/4..5) tra A ) + g,h1 tra A ) + tr A g) A) A g) A) ) g1 tr A g) A h) ) A g) A h) ) ) g,h1 tr A g) A h) ) A g) A h) ) ) trσ 4,.6) g,h1 tra g) A h) trσσ d trσ 4. ) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, trσ 4. 7 )

48 Fourth, we show iv). By applying Cauchy-Schwarz inequality ) ) 1/ A g) A g ) )A h ) A h) tra g) A g ) ) tra h A h)) 1/,g h ) 1/ tra g) A g ) tr,g h,g h ) 1/ tra h ) 4 ) 1/ tra h)4 ) 1/ ) 1/ tra g)4 ) 1/ tra g ) 4 ) 1/,g h ) 1/ tra h ) 4 ) 1/ tra h)4 ) 1/ trσ 4 gg) 1/4 trσ 4 hh) 1/4..7) On the other hand, we note that ) A g) A g ) )A h ) A h),g h tr tr A A)A ) tr A g) A)AA g)) g1 tr A g) A h) )A h) A g)) g,h1 4trΣ 4,.8) 8

49 since ) tr A A)A tra A) 1/ tra 4 ) 1/ trσ 4, tr A g) A)AA g)) a g) ij a ija jk a g) ki g1 i,j,k1 ) 1/ ) 1/ a g) ij a ij a jk a g) ki ) g1 g1 m a 4 ij trσ 4, tr A g) A h) )A h) A g)) g,h1 g,h1 g,h1 ) 1/4 m g1 g1 k1 trσ 4) 1/4 trσσd ) 1/4 trσ 3 Σ d ) 1/ i,j,k1 g,h1 a g) ij ah) ij ah) a g) ij ah) ij jk ag) ki a g) ij a h) ij trσσ d trσ 4. ) 1/4 ) 1/ a g) ij tra g) A g1 a h) jk ag) ki k1 ) 1/ ) 1/ a h) jk ag) ki ) g,h1 k1 From.7) and.8), we obtain ) ) A g) A g ) )A h ) A h) min trσ 4 gg) 1/4 trσ 4 hh) 1/4, 4trΣ 4.,g h tr Fifth, we show v). Note that tr {A g 1) A g 3) ) A h 1) A h 3) )}{A g ) A g 4) ) A h ) A h 4) )} ) tr {A g 1) A g 3) ) A h 1) A h 3) )}{A g 3) A g 1) ) A h 3) A h 1) )} )) 1/ tr {A g) A g4) ) A h) A h4) )}{A g4) A g) ) A h4) A h) )} )) 1/ ) 1/ tra g 1 A g 3 )tra h 1 A h 3 ) tra g A g 4 )tra h A h 4 ) 4 t1 tra gt)4) 1/4 tra ht)4) 1/4. ) 1/ 9

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Takahiro Nishiyama a,, Masashi Hyodo b, Takashi Seo a, Tatjana Pavlenko c a Department of Mathematical

More information

Testing Block-Diagonal Covariance Structure for High-Dimensional Data

Testing Block-Diagonal Covariance Structure for High-Dimensional Data Testing Block-Diagonal Covariance Structure for High-Dimensional Data MASASHI HYODO 1, NOBUMICHI SHUTOH 2, TAKAHIRO NISHIYAMA 3, AND TATJANA PAVLENKO 4 1 Department of Mathematical Sciences, Graduate School

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

High-dimensional two-sample tests under strongly spiked eigenvalue models

High-dimensional two-sample tests under strongly spiked eigenvalue models 1 High-dimensional two-sample tests under strongly spiked eigenvalue models Makoto Aoshima and Kazuyoshi Yata University of Tsukuba Abstract: We consider a new two-sample test for high-dimensional data

More information

On testing the equality of mean vectors in high dimension

On testing the equality of mean vectors in high dimension ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.

More information

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data Sri Lankan Journal of Applied Statistics (Special Issue) Modern Statistical Methodologies in the Cutting Edge of Science Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations

More information

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means.

More information

Statistica Sinica Preprint No: SS R2

Statistica Sinica Preprint No: SS R2 Statistica Sinica Preprint No: SS-2016-0063R2 Title Two-sample tests for high-dimension, strongly spiked eigenvalue models Manuscript ID SS-2016-0063R2 URL http://www.stat.sinica.edu.tw/statistica/ DOI

More information

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work Hiroshima Statistical Research Group: Technical Report Approximate interval estimation for PMC for improved linear discriminant rule under high dimensional frame work Masashi Hyodo, Tomohiro Mitani, Tetsuto

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

Estimation of multivariate 3rd moment for high-dimensional data and its application for testing multivariate normality

Estimation of multivariate 3rd moment for high-dimensional data and its application for testing multivariate normality Estimation of multivariate rd moment for high-dimensional data and its application for testing multivariate normality Takayuki Yamada, Tetsuto Himeno Institute for Comprehensive Education, Center of General

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors Takashi Seo a, Takahiro Nishiyama b a Department of Mathematical Information Science, Tokyo University

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX Submitted to the Annals of Statistics HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX By Shurong Zheng, Zhao Chen, Hengjian Cui and Runze Li Northeast Normal University, Fudan

More information

High-dimensional asymptotic expansions for the distributions of canonical correlations

High-dimensional asymptotic expansions for the distributions of canonical correlations Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic

More information

Tests for Covariance Matrices, particularly for High-dimensional Data

Tests for Covariance Matrices, particularly for High-dimensional Data M. Rauf Ahmad Tests for Covariance Matrices, particularly for High-dimensional Data Technical Report Number 09, 200 Department of Statistics University of Munich http://www.stat.uni-muenchen.de Tests for

More information

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Nobumichi Shutoh, Masashi Hyodo and Takashi Seo 2 Department of Mathematics, Graduate

More information

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control Takahiro Nishiyama a a Department of Mathematical Information Science, Tokyo University of Science,

More information

Effects of data dimension on empirical likelihood

Effects of data dimension on empirical likelihood Biometrika Advance Access published August 8, 9 Biometrika (9), pp. C 9 Biometrika Trust Printed in Great Britain doi:.9/biomet/asp7 Effects of data dimension on empirical likelihood BY SONG XI CHEN Department

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

The Third International Workshop in Sequential Methodologies

The Third International Workshop in Sequential Methodologies Area C.6.1: Wednesday, June 15, 4:00pm Kazuyoshi Yata Institute of Mathematics, University of Tsukuba, Japan Effective PCA for large p, small n context with sample size determination In recent years, substantial

More information

A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report -

A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report - A Box-Type Approximation for General Two-Sample Repeated Measures - Technical Report - Edgar Brunner and Marius Placzek University of Göttingen, Germany 3. August 0 . Statistical Model and Hypotheses Throughout

More information

TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA

TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA Statistica Sinica (014): Preprint 1 TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA Long Feng 1, Changliang Zou 1, Zhaojun Wang 1, Lixing Zhu 1 Nankai University and Hong Kong Baptist University

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA Kazuyuki Koizumi Department of Mathematics, Graduate School of Science Tokyo University of Science 1-3, Kagurazaka,

More information

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Tetsuro Sakurai and Yasunori Fujikoshi Center of General Education, Tokyo University

More information

On corrections of classical multivariate tests for high-dimensional data

On corrections of classical multivariate tests for high-dimensional data On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics

More information

Assessing the dependence of high-dimensional time series via sample autocovariances and correlations

Assessing the dependence of high-dimensional time series via sample autocovariances and correlations Assessing the dependence of high-dimensional time series via sample autocovariances and correlations Johannes Heiny University of Aarhus Joint work with Thomas Mikosch (Copenhagen), Richard Davis (Columbia),

More information

On the Testing and Estimation of High- Dimensional Covariance Matrices

On the Testing and Estimation of High- Dimensional Covariance Matrices Clemson University TigerPrints All Dissertations Dissertations 1-009 On the Testing and Estimation of High- Dimensional Covariance Matrices Thomas Fisher Clemson University, thomas.j.fisher@gmail.com Follow

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Empirical Likelihood Tests for High-dimensional Data

Empirical Likelihood Tests for High-dimensional Data Empirical Likelihood Tests for High-dimensional Data Department of Statistics and Actuarial Science University of Waterloo, Canada ICSA - Canada Chapter 2013 Symposium Toronto, August 2-3, 2013 Based on

More information

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University

More information

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR Introduction a two sample problem Marčenko-Pastur distributions and one-sample problems Random Fisher matrices and two-sample problems Testing cova On corrections of classical multivariate tests for high-dimensional

More information

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large Tetsuji Tonda 1 Tomoyuki Nakagawa and Hirofumi Wakaki Last modified: March 30 016 1 Faculty of Management and Information

More information

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Bull. Korean Math. Soc. 5 (24), No. 3, pp. 7 76 http://dx.doi.org/34/bkms.24.5.3.7 KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Yicheng Hong and Sungchul Lee Abstract. The limiting

More information

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data Yujun Wu, Marc G. Genton, 1 and Leonard A. Stefanski 2 Department of Biostatistics, School of Public Health, University of Medicine

More information

Statistica Sinica Preprint No: SS

Statistica Sinica Preprint No: SS Statistica Sinica Preprint No: SS-2017-0275 Title Testing homogeneity of high-dimensional covariance matrices Manuscript ID SS-2017-0275 URL http://www.stat.sinica.edu.tw/statistica/ DOI 10.5705/ss.202017.0275

More information

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS Communications in Statistics - Simulation and Computation 33 (2004) 431-446 COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS K. Krishnamoorthy and Yong Lu Department

More information

Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions

Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions Tiefeng Jiang 1 and Fan Yang 1, University of Minnesota Abstract For random samples of size n obtained

More information

Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population

Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population (Last Modified: April 24, 2008) Hirokazu Yanagihara 1 and Ke-Hai Yuan 2 1 Department of Mathematics, Graduate

More information

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017 Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION By Degui Li, Peter C. B. Phillips, and Jiti Gao September 017 COWLES FOUNDATION DISCUSSION PAPER NO.

More information

Multivariate-sign-based high-dimensional tests for sphericity

Multivariate-sign-based high-dimensional tests for sphericity Biometrika (2013, xx, x, pp. 1 8 C 2012 Biometrika Trust Printed in Great Britain Multivariate-sign-based high-dimensional tests for sphericity BY CHANGLIANG ZOU, LIUHUA PENG, LONG FENG AND ZHAOJUN WANG

More information

On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly Unrelated Regression Model and a Heteroscedastic Model

On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly Unrelated Regression Model and a Heteroscedastic Model Journal of Multivariate Analysis 70, 8694 (1999) Article ID jmva.1999.1817, available online at http:www.idealibrary.com on On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly

More information

On Expected Gaussian Random Determinants

On Expected Gaussian Random Determinants On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose

More information

A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices

A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices Natalia Bailey 1 M. Hashem Pesaran 2 L. Vanessa Smith 3 1 Department of Econometrics & Business Statistics, Monash

More information

Assumptions for the theoretical properties of the

Assumptions for the theoretical properties of the Statistica Sinica: Supplement IMPUTED FACTOR REGRESSION FOR HIGH- DIMENSIONAL BLOCK-WISE MISSING DATA Yanqing Zhang, Niansheng Tang and Annie Qu Yunnan University and University of Illinois at Urbana-Champaign

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

A distance-based, misclassification rate adjusted classifier for multiclass, high-dimensional data

A distance-based, misclassification rate adjusted classifier for multiclass, high-dimensional data Ann Inst Stat Math (2014) 66:983 1010 DOI 10.1007/s10463-013-0435-8 A distance-based, misclassification rate adjusted classifier for multiclass, high-dimensional data Makoto Aoshima Kazuyoshi Yata Received:

More information

A new test for the proportionality of two large-dimensional covariance matrices. Citation Journal of Multivariate Analysis, 2014, v. 131, p.

A new test for the proportionality of two large-dimensional covariance matrices. Citation Journal of Multivariate Analysis, 2014, v. 131, p. Title A new test for the proportionality of two large-dimensional covariance matrices Authors) Liu, B; Xu, L; Zheng, S; Tian, G Citation Journal of Multivariate Analysis, 04, v. 3, p. 93-308 Issued Date

More information

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA J. Japan Statist. Soc. Vol. 37 No. 1 2007 53 86 MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA M. S. Srivastava* In this article, we develop a multivariate theory for analyzing multivariate datasets

More information

SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS

SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS Hisayuki Tsukuma and Yoshihiko Konno Abstract Two-sample problems of estimating p p scale matrices

More information

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Inference for High Dimensional Robust Regression

Inference for High Dimensional Robust Regression Department of Statistics UC Berkeley Stanford-Berkeley Joint Colloquium, 2015 Table of Contents 1 Background 2 Main Results 3 OLS: A Motivating Example Table of Contents 1 Background 2 Main Results 3 OLS:

More information

Composite Hotelling s T-Square Test for High-Dimensional Data. 1. Introduction

Composite Hotelling s T-Square Test for High-Dimensional Data. 1. Introduction A^VÇÚO 1 33 ò 1 4 Ï 2017 c 8 Chinese Journal of Applied Probability and Statistics Aug., 2017, Vol. 33, No. 4, pp. 349-368 doi: 10.3969/j.issn.1001-4268.2017.04.003 Composite Hotelling s T-Square Test

More information

Mi-Hwa Ko. t=1 Z t is true. j=0

Mi-Hwa Ko. t=1 Z t is true. j=0 Commun. Korean Math. Soc. 21 (2006), No. 4, pp. 779 786 FUNCTIONAL CENTRAL LIMIT THEOREMS FOR MULTIVARIATE LINEAR PROCESSES GENERATED BY DEPENDENT RANDOM VECTORS Mi-Hwa Ko Abstract. Let X t be an m-dimensional

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12 Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

arxiv: v1 [stat.ml] 5 Jan 2015

arxiv: v1 [stat.ml] 5 Jan 2015 To appear on the Annals of Statistics SUPPLEMENTARY MATERIALS FOR INNOVATED INTERACTION SCREENING FOR HIGH-DIMENSIONAL NONLINEAR CLASSIFICATION arxiv:1501.0109v1 [stat.ml] 5 Jan 015 By Yingying Fan, Yinfei

More information

More powerful tests for sparse high-dimensional covariances matrices

More powerful tests for sparse high-dimensional covariances matrices Accepted Manuscript More powerful tests for sparse high-dimensional covariances matrices Liu-hua Peng, Song Xi Chen, Wen Zhou PII: S0047-259X(16)30004-5 DOI: http://dx.doi.org/10.1016/j.jmva.2016.03.008

More information

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order

More information

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi Faculty of Science and Engineering, Chuo University, Kasuga, Bunkyo-ku,

More information

Persistence and global stability in discrete models of Lotka Volterra type

Persistence and global stability in discrete models of Lotka Volterra type J. Math. Anal. Appl. 330 2007 24 33 www.elsevier.com/locate/jmaa Persistence global stability in discrete models of Lotka Volterra type Yoshiaki Muroya 1 Department of Mathematical Sciences, Waseda University,

More information

Tests for High Dimensional Generalized Linear Models

Tests for High Dimensional Generalized Linear Models MPRA Munich Personal RePEc Archive Tests for High Dimensional Generalized Linear Models Song Xi Chen and Bin Guo Peking Univeristy and Iowa State University, Peking Univeristy 2014 Online at http://mpra.ub.uni-muenchen.de/59816/

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Stein s Method and Characteristic Functions

Stein s Method and Characteristic Functions Stein s Method and Characteristic Functions Alexander Tikhomirov Komi Science Center of Ural Division of RAS, Syktyvkar, Russia; Singapore, NUS, 18-29 May 2015 Workshop New Directions in Stein s method

More information

Networks: Lectures 9 & 10 Random graphs

Networks: Lectures 9 & 10 Random graphs Networks: Lectures 9 & 10 Random graphs Heather A Harrington Mathematical Institute University of Oxford HT 2017 What you re in for Week 1: Introduction and basic concepts Week 2: Small worlds Week 3:

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

Grouped Network Vector Autoregression

Grouped Network Vector Autoregression Statistica Sinica: Supplement Grouped Networ Vector Autoregression Xuening Zhu 1 and Rui Pan 2 1 Fudan University, 2 Central University of Finance and Economics Supplementary Material We present here the

More information

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION J. Japan Statist. Soc. Vol. 32 No. 2 2002 165 181 A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION Hisayuki Tsukuma* The problem of estimation in multivariate linear calibration

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

Chapter 7, continued: MANOVA

Chapter 7, continued: MANOVA Chapter 7, continued: MANOVA The Multivariate Analysis of Variance (MANOVA) technique extends Hotelling T 2 test that compares two mean vectors to the setting in which there are m 2 groups. We wish to

More information

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives TR-No. 14-06, Hiroshima Statistical Research Group, 1 11 Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives Mariko Yamamura 1, Keisuke Fukui

More information

CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL

CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL J. Japan Statist. Soc. Vol. 37 No. 1 2007 135 155 CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL Hirokazu Yanagihara* This paper presents the conditions for robustness

More information

A Squared Correlation Coefficient of the Correlation Matrix

A Squared Correlation Coefficient of the Correlation Matrix A Squared Correlation Coefficient of the Correlation Matrix Rong Fan Southern Illinois University August 25, 2016 Abstract Multivariate linear correlation analysis is important in statistical analysis

More information

Pattern correlation matrices and their properties

Pattern correlation matrices and their properties Linear Algebra and its Applications 327 (2001) 105 114 www.elsevier.com/locate/laa Pattern correlation matrices and their properties Andrew L. Rukhin Department of Mathematics and Statistics, University

More information

Inference For High Dimensional M-estimates: Fixed Design Results

Inference For High Dimensional M-estimates: Fixed Design Results Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49

More information

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Takeshi Emura and Hisayuki Tsukuma Abstract For testing the regression parameter in multivariate

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

Freeness and the Transpose

Freeness and the Transpose Freeness and the Transpose Jamie Mingo (Queen s University) (joint work with Mihai Popa and Roland Speicher) ICM Satellite Conference on Operator Algebras and Applications Cheongpung, August 8, 04 / 6

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

f-divergence Estimation and Two-Sample Homogeneity Test under Semiparametric Density-Ratio Models

f-divergence Estimation and Two-Sample Homogeneity Test under Semiparametric Density-Ratio Models IEEE Transactions on Information Theory, vol.58, no.2, pp.708 720, 2012. 1 f-divergence Estimation and Two-Sample Homogeneity Test under Semiparametric Density-Ratio Models Takafumi Kanamori Nagoya University,

More information

Advanced Statistics II: Non Parametric Tests

Advanced Statistics II: Non Parametric Tests Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney

More information

Notes on Random Vectors and Multivariate Normal

Notes on Random Vectors and Multivariate Normal MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case The largest eigenvalues of the sample covariance matrix 1 in the heavy-tail case Thomas Mikosch University of Copenhagen Joint work with Richard A. Davis (Columbia NY), Johannes Heiny (Aarhus University)

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Lecture 20: Linear model, the LSE, and UMVUE

Lecture 20: Linear model, the LSE, and UMVUE Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Estimation of the Conditional Variance in Paired Experiments

Estimation of the Conditional Variance in Paired Experiments Estimation of the Conditional Variance in Paired Experiments Alberto Abadie & Guido W. Imbens Harvard University and BER June 008 Abstract In paired randomized experiments units are grouped in pairs, often

More information

Gaussian vectors and central limit theorem

Gaussian vectors and central limit theorem Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables

More information

COMPLETE SPACELIKE HYPERSURFACES IN THE DE SITTER SPACE

COMPLETE SPACELIKE HYPERSURFACES IN THE DE SITTER SPACE Chao, X. Osaka J. Math. 50 (203), 75 723 COMPLETE SPACELIKE HYPERSURFACES IN THE DE SITTER SPACE XIAOLI CHAO (Received August 8, 20, revised December 7, 20) Abstract In this paper, by modifying Cheng Yau

More information

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL Jrl Syst Sci & Complexity (2009) 22: 483 494 TESTIG SERIAL CORRELATIO I SEMIPARAMETRIC VARYIG COEFFICIET PARTIALLY LIEAR ERRORS-I-VARIABLES MODEL Xuemei HU Feng LIU Zhizhong WAG Received: 19 September

More information

Assignment 10. Arfken Show that Stirling s formula is an asymptotic expansion. The remainder term is. B 2n 2n(2n 1) x1 2n.

Assignment 10. Arfken Show that Stirling s formula is an asymptotic expansion. The remainder term is. B 2n 2n(2n 1) x1 2n. Assignment Arfken 5.. Show that Stirling s formula is an asymptotic expansion. The remainder term is R N (x nn+ for some N. The condition for an asymptotic series, lim x xn R N lim x nn+ B n n(n x n B

More information

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals Satoshi AOKI and Akimichi TAKEMURA Graduate School of Information Science and Technology University

More information