CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL

Size: px
Start display at page:

Download "CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL"

Transcription

1 J. Japan Statist. Soc. Vol. 37 No CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL Hirokazu Yanagihara* This paper presents the conditions for robustness to the nonnormality on three test statistics for a general multivariate linear hypothesis, which were proposed under the normal assumption in a generalized multivariate analysis of variance (GMANOVA) model. The proposed conditions require the cumulants of an unknown population s distribution to vanish in the second terms of the asymptotic expansions for both the mean and variance of the test statistics. With the proposed conditions, the test statistic can be investigated for robustness to nonnormality of the population s distribution. When the conditions are satisfied, the Bartlett correction and the modified Bartlett correction in the normal case improve the quality of the chi-square approximation even under nonnormality. Key words and phrases: Actual test size, asymptotic expansion, Bartlett correction, chi-square approximation, general multivariate linear hypothesis, modified Bartlett correction. 1. Introduction During the last several decades, there have been many studies on the influences of nonnormality on several statistical test procedures proposed under the normal assumption. For the test size (or the significant level), Hotelling s two-sample test is known to be robust to nonnormality of the population s distribution (i.e., the true distribution), while Hotelling s one-sample test is not. Chase and Bulgren (1971) and Everitt (1979) showed these facts through numerical experiments. Figure 1 shows the results of a similar numerical experiment in the case of sample size n = 20 and the dimension of observation p = 2. Normal, Laplace, uniform, skew-laplace, chi-square and log-normal distributions were used as the true distributions (for details of the setting of true distributions in numerical studies, see Appendix A.1). This figure explains actual test sizes of the two test statistics under the six true distributions when we believe that the data came from the normal distribution, i.e., when we use the F -distribution as the null distribution. From this figure, it is revealed that a variation of the actual test size of the two-sample test, caused by the difference in distributions, is small as opposed to the case of the one-sample test. In the case of the one-sample test, the larger the skewness of the true distributions, the more the difference between the actual and nominal test size grows. If the test size, which we believe to be 10%, becomes 15% or 25%, the corresponding test statistic becomes difficult to use. Received February 27,2006. Revised May 8,2006. Accepted May 15,2006. *Department of Mathematics,Graduate School of Science,Hiroshima University,1-3-1 Kagamiyama, Higashi-Hiroshima,Hiroshima ,Japan.

2 136 HIROKAZU YANAGIHARA 25 Nominal 10% Test 20 Nominal 5% Test Actual Size (%) Actual Size (%) Distributions Nominal Size One Sample Test Two Sample Test Distributions are: 1. Normal 2. Laplace 3. Uniform 4. Skew Laplace 5. Chi Square 6. Log Normal Actual Size (%) Distributions Nominal 1% Test Distributions Figure 1. Actual test sizes of Hotelling s one- and two-sample tests. Although the statistical model in both tests is essentially a multivariate linear model, a large difference is observed in the influence of nonnormality on these two test statistics. A serious difference between these two test statistics appears in the second term of an asymptotic expansion of the mean of the test statistic. Although there are skewnesses of the true distribution in the second term of this expansion of the one-sample test statistic, there are no skewnesses in that of the two-sample test statistic. In this paper, we investigate the difference of the influence theoretically, not numerically. We extend the test statistic for a general multivariate linear hypothesis in the generalized multivariate analysis of variance (GMANOVA) model (Potthoff and Roy (1964)) from Hotelling s one- and two-sample test statistics. Here we use the robustness to nonnormality for the case where the cumulants of the true distribution, i.e., a quantity denoting nonnormality of the true distribution, do not remain in the second terms of asymptotic expansions of the moments for the test statistics. We search for the condition that the cumulants of the true distribution vanish in the second terms of asymptotic expansions for the mean and variance of the test statistic. The condition is clarified through the coefficients of an asymptotic expansion of the null distribution of the test statistic. With the proposed condition, the used test statistic can be investigated for robustness to nonnormality of the true distribution. Furthermore, when the proposed condition is satisfied, the Bartlett correction (Bartlett (1937)) in the normal case, or the modified Bartlett correction (Fujikoshi (2000)), improves the quality of the chi-square approximation under any distributions. Improvement in the quality of the chi-square approximation without estimating higher-order cumulants is important because the bias

3 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 137 of estimators for the cumulants becomes large without having adequate sample size. Especially, the bias of the ordinary estimator of kurtosis proposed by Mardia (1970) becomes serious unless the sample size is huge (see Yanagihara (2007)). The paper is organized as follows: In Section 2, we elaborate on Yanagihara s asymptotic expansions of the null distributions of the three test statistics in a GMANOVA model (Yanagihara (2001)). By using the coefficients of these expansions, the conditions for the test statistics to be robust to nonnormality of the true distribution are obtained in Section 3, followed by the discussion and conclusion in Section 4. Technical details are provided in the Appendix. 2. Asymptotic expansions of null distributions of test statistics in the GMANOVA model The GMANOVA model proposed by Potthoff and Roy (1964) is defined by (2.1) Y = AΞX + EΣ 1/2, where Y =(y 1,...,y n ) is an n p observation matrix of response variables, A =(a 1,...,a n ) is an n k between-individuals design matrix of explanatory variables with the full rank k (< n), X is a p q within-individuals design matrix of explanatory variables with the full rank q ( p), Ξ is a k q unknown parameter matrix, and E =(ε 1,...,ε n ) is an n p unobserved error matrix. It is assumed that ε 1,...,ε n are independent random vectors from ε =(ε 1,...,ε p ), and that the true distribution of ε follows an unknown distribution with a mean E[ε] =0 p and variance-covariance matrix Cov[ε] =I p. Notice that 0 p is a p 1 vector with all elements equal 0. Since this model can be frequently applied to the analysis of growth curve data, it is also called the growth curve model. We consider testing for a general linear hypothesis as (2.2) H 0 : CΞD = O c d, where C is a known c k matrix with the rank c ( k), D is a known q d matrix with the rank d ( q) and O c d is a c d matrix with all elements equal 0. Let S h and S e be the variation matrices due to the hypothesis and the error respectively, i.e., where S h =(C ΞD) (CRC ) 1 (C ΞD), S e = D (X S 1 X) 1 D, Ξ =(A A) 1 A YS 1 X(X S 1 X) 1, R =(A A) 1 A (I n + YS 1 Y )A(A A) 1 Ξ(X S 1 X) Ξ, and S = Y (I n P A )Y. Here, P A is the projection matrix to the linear space R(A) generated by the column vectors of A, i.e., P A = A(A A) 1 A. Then,

4 138 HIROKAZU YANAGIHARA the three criteria proposed for testing (2.2) under normality, in particular, are as follows: (2.3) (i) the likelihood ratio statistic (LR): T LR = {n k (p q)} log( S e / S e + S h ), (ii) the Lawley-Hotelling trace criterion (LH): T LH = {n k (p q)} tr(s h S 1 e ), (iii) the Bartlett-Nanda-Pillai trace criterion (BNP): T BNP = {n k (p q)} tr{s h (S h + S e ) 1 }, (for these results, see e.g., Kariya (1978), von Rosen (1991), Fujikoshi (1993), and Srivastava and von Rosen (1999)). By changing the known matrices C and D, it is possible to express several linear hypotheses. For these examples, see Kshirsagar and Smith (1995, p. 40) and Yanagihara (2001). In order to simplify the formula of asymptotic expansions of the null distributions of the three test statistics, we state some notations of the moments and cumulants of the true distribution, and define some projection matrices. Let µ j1 j m be an mth multivariate moment of ε defined by µ j1 j m = E[ε j1 ε jm ]. Similarly, the corresponding mth multivariate cumulant of ε is denoted by κ j1 j m, e.g., κ abc = µ abc, κ abcd = µ abcd δ ab δ cd δ ac δ bd δ ad δ bc, where δ ab is the Kronecker delta, i.e., δ aa = 1 and δ ab = 0 for a b. Let e 1 and e 2 be independent random vectors from ε, and H, L and M be p p symmetric matrices whose (a, b)th elements are h ab, l ab and m ab, respectively. Then, we define the following multivariate skewness and kurtosis of the transformed ε by H, L and M. κ(h, M) =E e1 [(e 1He 1 )(e 1Me 1 )] {tr(h) tr(m) + 2 tr(hm)} p = κ abcd h ab m cd, a,b,c,d γ1(h, 2 L, M) =E e1 E e2 [(e 1He 2 )(e 1Le 2 )(e 1Me 2 )] p = κ abc κ def h ad l be m cf, a,b,c,d,e,f γ2(h, 2 L, M) =E e1 E e2 [(e 1He 1 )(e 1Le 2 )(e 2Me 2 )] p = κ abc κ def h ab l cd m ef, a,b,c,d,e,f

5 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 139 where the notation p a,b,... means p p a=1 b=1. The ordinary multivariate skewnesses and kurtosis (see e.g., Mardia (1970) and Isogai (1983)) can be expressed as κ (1) 4 = κ(i p, I p ), κ (1) 3,3 = γ2 1(I p, I p, I p ), κ (2) 3,3 = γ2 2(I p, I p, I p ). Let an n n idempotent matrix B whose (i, j)th element is b ij be given by B = A(A A) 1 C {C(A A) 1 C } 1 C(A A) 1 A. Then, the matrices B (d) and B (3) are defined by b ij as b 3 11 b3 1n B (d) = diag(b 11,...,b nn ), B (3) = b 3 n1 b3 nn Furthermore, the projection matrices Ψ and Φ are given by Ψ = Σ 1/2 X(X Σ 1 X) 1 D{D (X Σ 1 X) 1 D} 1 D (X Σ 1 X) 1 X Σ 1/2, Φ = I p Σ 1/2 X(X Σ 1 X) 1 X Σ 1/2. By using these skewnesses, kurtosis and matrices, asymptotic expansions of the null distributions of the test statistics in (2.3) are expressed as in the following lemma (for the proof of the lemma, see Yanagihara (2001)). Lemma 2.1. Suppose that the between-individuals design matrix A and the true distribution of the error matrix E satisfy the following five assumptions: 1 n lim sup a i 4 <, where denotes the Euclidean norm, n n i=1 λ n lim inf n For some constant δ (0, 1/2], n > 0, where λ n is the smallest eigenvalue of A A, max a i = O(n 1/2 δ ), i=1,...,n E[ ε 8 ] <, The Cramér s condition for the joint distribution of ε and εε holds. Then, the null distributions of the three test statistics in (2.3) are expanded as (2.4) P (T x) =G cd (x)+ 1 n 3 β j G cd+2j (x)+o(n 1 ), j=0 where G f (x) is the distribution function of the chi-square distribution with f degrees of freedom and the coefficients β j are given by β 0 = 1 8 m(1) 4 κ(ψ, Ψ) 1 24 {2m(1) 3,3 +3(c 2)(c +1)m(1) 1,1 }γ2 1(Ψ, Ψ, Ψ) 1 8 {m(2) 3,3 2cm(1) 3,1 (c 2)m(1) 1,1 }γ2 2(Ψ, Ψ, Ψ)

6 140 HIROKAZU YANAGIHARA 1 2 {2m(1) 3,1 (2c +1)m(1) 1,1 }γ2 1(Ψ, Ψ, Φ) 1 2 {m(1) 3,1 cm(1) 1,1 }γ2 2(Ψ, Ψ, Φ) 1 2 {m(1) 3,1 (c +1)m(1) 1,1 }γ2 2(Ψ, Φ, Ψ) 1 2 m(1) 1,1 {3γ2 1(Ψ, Φ, Φ)+2γ 2 2(Ψ, Φ, Φ)+γ 2 2(Φ, Ψ, Φ)} + 1 cd(c d 1), 4 β 1 = 1 4 m(1) 4 κ(ψ, Ψ)+1 8 {2m(1) 3,3 4m(1) 3,1 +(3c2 + c 6)m (1) 1,1 }γ2 1(Ψ, Ψ, Ψ) {3m(2) 3,3 2(3c +2)m(1) 3,1 +(c +6)m(1) 1,1 }γ2 2(Ψ, Ψ, Ψ) (2.5) {4m(1) 3,1 (4c +5)m(1) 1,1 }γ2 1(Ψ, Ψ, Φ) + {m (1) 3,1 (c +1)m(1) 1,1 }γ2 2(Ψ, Ψ, Φ) {2m(1) 3,1 (2c +3)m(1) 1,1 }γ2 2(Ψ, Φ, Ψ) m(1) 1,1 {3γ2 1(Ψ, Φ, Φ)+2γ 2 2(Ψ, Φ, Φ)+γ 2 2(Φ, Ψ, Φ)} 1 cd{c + r(c + d +1)}, 2 β 2 = 1 8 m(1) 4 κ(ψ, Ψ) 1 8 {2m(1) 3,3 8m(1) 3,1 +(c + 2)(3c 1)m(1) 1,1 }γ2 1(Ψ, Ψ, Ψ) 1 8 {3m(2) 3,3 2(3c +4)m(1) 3,1 +5(c +2)m(1) 1,1 }γ2 2(Ψ, Ψ, Ψ) 1 2 {m(1) 3,1 (c +2)m(1) 1,1 }{2γ2 1(Ψ, Ψ, Φ)+γ2(Ψ, 2 Ψ, Φ)+γ2(Ψ, 2 Φ, Ψ)} + 1 cd(c + d +1)(1+2r), 4 β 3 = 1 24 {2m(1) 3,3 12m(1) 3,1 +3(c + 1)(c +2)m(1) 1,1 }γ2 1(Ψ, Ψ, Ψ) {m(2) 3,3 2(c +2)m(1) 3,1 +3(c +2)m(1) 1,1 }γ2 2(Ψ, Ψ, Ψ). Here, the constant r is determined by the test statistics as 1/2 (when T is T LR ) r = 0 (when T is T LH ) 1 (when T is T BNP ) and the coefficients m are defined by the between-individuals design matrix A and

7 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 141 the coefficient matrix for hypothesis C as (2.6) m (1) 1,1 = 1 n 1 nb1 n, m (1) 4 = n tr(b 2 (d)) c(c +2), m (1) 3,1 = 1 nb (d) B1 n, m (1) 3,3 = n1 nb (3) 1 n, m (2) 3,3 = n1 nb (d) BB (d) 1 n, where 1 n is an n 1 vector with all elements equal 1. Let u i be a c 1 vector given by (2.7) u i = n{c(a A) 1 C } 1/2 C(A A) 1 a i, (i =1,...,n). Then the coefficients in (2.6) are rewritten as (2.8) m (1) 1,1 = 1 n 2 m (1) 3,1 = 1 n 2 m (2) 3,3 = 1 n 2 n i,j n i,j u iu j, m (1) 4 = 1 n n u i 4 c(c +2), i=1 u i 2 (u iu j ), m (1) 3,3 = 1 n 2 n u i 2 u j 2 (u iu j ). i,j n (u iu j ) 3, i,j Therefore, the coefficients β j (j =0, 1, 2, 3) are determined by the cumulants of ε and sample cumulants of u i (i =1,...,n). Lemma 2.1 points out that the second term in the asymptotic expansion of the null distribution of (2.3) depends on the skewnesses and kurtosis of the true distribution. Therefore, we can generally say that the second terms in asymptotic expansions of the mean and variance of the test statistic depend on the cumulants of ε. However, in the next section, we show conditions that these cumulants vanish in the second terms of the expansions. 3. Conditions for robustness to nonnormality 3.1. Condition for the mean In this subsection, we consider the condition that the cumulants of ε vanish in the second term of an asymptotic expansion for the mean of the test statistic T. Notice that 0 x s dg f (x) =f(f +2) (f +2(s 1)). From the formula (2.4) in Lemma 2.1, we obtain an asymptotic expansion of the mean of T as E[T ]=cd + 1 n 3 β j (cd +2j)+o(n 1 ). j=0

8 142 HIROKAZU YANAGIHARA By using the relation 3 j=0 β j = 0, we can write a concrete form of the expansion of E[T ]as (3.1) E[T ]=cd + 1 n m(1) 1,1 [γ2 1(Ψ, Ψ, Ψ)+γ 2 2(Ψ, Ψ, Ψ) +3{γ1(Ψ, 2 Ψ, Φ)+γ1(Ψ, 2 Φ, Φ)} +2{γ2(Ψ, 2 Ψ, Φ)+γ2(Ψ, 2 Φ, Φ)} + γ2(ψ, 2 Φ, Ψ)+γ2(Φ, 2 Ψ, Φ)] + 1 n cd{r(c + d +1)+d +1} + o(n 1 ) ( = cd 1+ η ) 1 + o(n 1 ). n From (3.1), we can see that η 1 is independent of the cumulants of ε if the coefficient m (1) 1,1 is 0 or all the γ2 j (,, ) are 0 (j =1, 2). It means that a location of the null distribution of the test statistic does not easily change due to the nonnormality of the true distribution. Then, the test statistic becomes robust to nonnormality of the true distribution. Notice that m (1) 1,1 = 0 is equivalent to 1 n 1 na(a A) 1 C {C(A A) 1 C } 1 C(A A) 1 A 1 n =0. Since C(A A) 1 C is a positive definite matrix, the above equation can be rewritten as C(A A) 1 A 1 n = 0 c. Therefore, we obtain the conditions for robustness to nonnormality as in the following theorem. Theorem 3.1. If a considered test satisfies either of the following, Condition 1 or 2: Condition 1: C(A A) 1 A 1 n = 0 c, Condition 2: All the γj 2 (,, ) (j =1, 2) are 0, the second term in the asymptotic expansion of the mean of the test statistic T becomes independent of the cumulants of ε, i.e., E[T ]=cd [1+ 1n ] (3.2) {r(c + d +1)+d +1} + o(n 1 ). Consequently, the location of the null distribution of T does not easily change due to nonnormality of the true distribution. From Appendix A.2, we can see that all the γj 2 (,, ) are 0 if the distribution of ε possesses orthant symmetry defined by Efron (1969). Therefore, we obtain the following corollary (the proof is given in Appendix A.2). Corollary 3.1. Condition 2 holds. If the distribution of ε possesses orthant symmetry, then If the distribution possesses orthant symmetry, this distribution is not skewed. It is well known that the elliptical distribution in which the mean is zero and variance-covariance matrix is diagonal possesses orthant symmetry. Many

9 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 143 authors have reported that a test statistic is robust when the distribution of observation is the elliptical distribution, e.g., Efron (1969), Eaton and Efron (1970), Dawid (1977), Kariya (1981a, 1981b) and Khatri (1988). Robustness to nonnormality, which is caused by Condition 2, corresponds to their results. Although Condition 2 depends on the unknown true distribution, Condition 1 only depends on the setting of the hypothesis testing. By using Condition 1, we can judge whether or not a used test statistic is robust to nonnormality of the true distribution. Here, we can consider two types of the between-individuals design matrix A. Let A be an n (k 1) matrix satisfying P A 1 n 1 n and rank(a )=k 1. When the matrix A has a segment, i.e., A =(1 n A ), then we call such a design matrix the type I between-individuals design matrix. On the other hand, when the matrix A is given by 1 n1 0 n1 (3.3) A =....., 0 nk 1 nk where k i=1 n i = n, then we call such a design matrix the type II betweenindividuals design matrix. Through these matrices, we can obtain settings which satisfy Condition 1 as in the following corollary (the proof is given in Appendix A.3). Corollary 3.2. We consider the two cases of the between-individuals design matrix A, i.e., A is type I or type II. Then, Condition 1 holds if the following conditions of the coefficient matrix for hypothesis C are satisfied. (I) In the case that A is type I, if C =(0 c C ), where C is any c (k 1) matrix with the rank c, Condition 1 holds. (II) In the case that A is type II, if the equation C1 k = 0 c is satisfied, Condition 1 holds. Corollary 3.2 points out that Hotelling s two-sample test satisfies Condition 1 and Hotelling s one-sample test does not. Even if the between-individuals design matrix A satisfying P A 1 n 1 n is not type I or II, the test can be made to satisfy Condition 1 through the transformations A (1 n A), C (0 c C) and Ξ (ξ Ξ ). Furthermore, we can see that testing for the equality, e.g., the equality for the means between k-groups, i.e., H 0 : µ 1 = = µ k, satisfies Condition 1, because matrix A is type II and C1 k = 0 c in this case. Practical tests are testing for the equality of parameters, e.g., the testing for the equality of several means in a multi-way MANOVA model; therefore, it seems that our result is useful for actual applications. If the mean of the test statistic T is expanded as (3.1), then the Bartlett correction can be obtained as ( T = 1 ˆη ) 1 T, n

10 144 HIROKAZU YANAGIHARA where ˆη 1 is a consistent estimator of η 1 in (3.1). If the true distribution is the normal distribution, the Bartlett correction is transformation by the constant coefficient. On the other hand, in general, it is necessary to estimate skewnesses in order to use the Bartlett correction. However, from Theorem 3.1, we can see that the Bartlett correction in the normal case can improve the quality of the chisquare approximation, even if the true distribution is not the normal distribution; this is because η 1 is independent of the cumulants of ε. Therefore, we obtain the following corollary. Corollary 3.3. Let T denote the test statistics adjusted by the Bartlett correction in the normal case as ( 1+ c d 1 ) T LR 2n ( T = 1 d +1 ) T LH. n ( 1+ c ) T BNP n Suppose that either Condition 1 or 2 holds. Then, the asymptotic expansion of the mean of T becomes E[ T ]=cd + o(n 1 ). This naturally implies that the Bartlett correction in the normal case can improve the quality of the chi-square approximation even under nonnormality. Before concluding this subsection, we verify the validity of our theorem through a numerical experiment. The true distributions used are described in Appendix A.1. We dealt with the test for a given matrix in the case of p = 2 and n = 30. The p (p 1) within-individuals design matrix X was generated from the uniform distribution on ( 1, 1) and the coefficient matrix for the hypothesis was D = I q. We prepared two test statistics specified by A and C, and their adjusted versions by the Bartlett correction in the normal case as follows: T 0 : A = A and C = I 3, T 0 : the adjusted version of T 0 by the Bartlett correction in the normal case, T 1 : A =(1 n A ) and C =(0 3 I 3 ), T 1 : the adjusted version of T 1 by the Bartlett correction in the normal case, where an n 3 matrix A was generated from the uniform distribution on (0, 1). It is easy to see that T 1 satisfies Condition 1, and T 0 and T 1 satisfy Condition 2 when the true distributions are 1, 2 and 3. Table 1 shows the actual test sizes of these test statistics when we used the chi-square distribution as the null distribution. The average and standard deviation (S.D.) of actual sizes for every true distribution are also given in the table. From this table, we can see that the S.D. becomes small when either Condition 1 or 2 holds. Therefore, we can say that a variation of the actual test size by the difference in distributions is small in these cases. Then, the Bartlett correction in the normal case improves the quality of the chi-square approximation. From the numerical results, we suggest that the test statistic satisfying Condition 1 should be adjusted by the Bartlett correction in the normal case.

11 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 145 Table 1. Actual test sizes of several test statistics for testing a given matrix. (LR) T 0 T0 T 1 T1 Nominal Sizes Nominal Sizes Nominal Sizes Nominal Sizes Distribution 10% 5% 1% 10% 5% 1% 10% 5% 1% 10% 5% 1% Normal Laplace Uniform Skew-Laplace Chi-Sqare Log-Normal Average S.D (LH) T 0 T0 T 1 T1 Nominal Sizes Nominal Sizes Nominal Sizes Nominal Sizes Distribution 10% 5% 1% 10% 5% 1% 10% 5% 1% 10% 5% 1% Normal Laplace Uniform Skew-Laplace Chi-Square Log-Normal Average S.D (BNP) T 0 T0 T 1 T1 Nominal Sizes Nominal Sizes Nominal Sizes Nominal Sizes Distribution 10% 5% 1% 10% 5% 1% 10% 5% 1% 10% 5% 1% Normal Laplace Uniform Skew-Laplace Chi-Square Log-Normal Average S.D Condition for the variance In this subsection, we consider the condition that the cumulants of ε vanish simultaneously in the second terms of asymptotic expansions for both the mean and variance of the test statistic. From the formula (2.4) in Lemma 2.1, we obtain the asymptotic expansion of the second moment of T as E[T 2 ]=cd(cd +2)+ 1 n 3 β j (cd +2j)(cd +2+2j)+o(n 1 ). j=0 By using the relation 3 j=0 β j = 0, we can write a concrete form of the expansion of E[T 2 ]as (3.4) E[T 2 ]=cd(cd +2) + 1 n [m(1) 4 κ(ψ, Ψ)

12 146 HIROKAZU YANAGIHARA 2{2m (1) 3,1 (cd +2c +6)m(1) 1,1 } {γ1(ψ, 2 Ψ, Ψ)+γ2(Ψ, 2 Ψ, Ψ)} 2{4m (1) 3,1 (cd +4c + 10)m(1) 1,1 }γ2 1(Ψ, Ψ, Φ) 2{2m (1) 3,1 (cd +2c +6)m(1) 1,1 } {γ2(ψ, 2 Ψ, Φ)+γ2(Ψ, 2 Φ, Ψ)} +2(cd +2)m (1) 1,1 {3γ1(Ψ, 2 Φ, Φ)+2γ2(Ψ, 2 Φ, Φ)+γ2(Φ, 2 Ψ, Φ)}] + 2 cd[(cd +2){r(c + d +1)+d +1} +(c + d +1)(1+2r)] n +o(n 1 ) ( = cd(cd +2) 1+ η 2 n ) + o(n 1 ). It is easy to see that the coefficient m (1) 3,1 becomes 0 if m(1) 1,1 from these equations, if either the coefficient m (1) 1,1 = 0 or all γ2 j = 0. Therefore, (,, ) = 0 (j =1, 2) holds and also the equation m (1) 4 = 0 holds simultaneously, η 1 and η 2 become independent of the cumulants of ε. This means that the location and dispersion of the null distribution of the test statistic do not easily change due to nonnormality of the true distribution. Notice that the coefficient m (1) 4 signifies the sample kurtosis of u 1,...,u n, where u i is given by (2.7). Therefore, we obtain the condition for robustness to nonnormality in the following theorem. Theorem 3.2. Suppose that a considered test satisfies either Condition 1 or 2 in Theorem 3.1 and furthermore that the following condition also holds: Condition 3: The sample kurtosis of u 1,...,u n is 0. Then, the second terms in asymptotic expansions of the mean and variance of the test statistic T become independent of the cumulants of ε, i.e., E[T ]=cd [1+ 1n ] {r(c + d +1)+d +1} + o(n 1 ), (3.5) Var[T ]=2cd [1+ 1n ] {4r(c + d +1)+c +3(d +1)} + o(n 1 ). Consequently, the location and dispersion of the null distribution of the test statistic do not easily change due to nonnormality of the true distribution. From Wakaki et al. (2002), if the between-individuals design matrix A is type II and the equation C1 k = 0 c holds, then the coefficient m (1) 4 is rewritten as m (1) 4 = k j=1 n/n j k 2 2k + 2. Therefore, we obtain the following corollary. Corollary 3.4. Suppose that the between-individuals design matrix A is type II and equation C1 k = 0 c holds. Let the sample sizes of k-groups be denoted by n 1,...,n k. If the equation k j=1 n/n j = k 2 +2k 2 holds, then Conditions 1 and 3 hold simultaneously.

13 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 147 Table 2. Robust design to nonnormality in testing the equality. (k =4) 1:1:2:4 1:2:2:5 1:2:4:4 n 1 : n 2 : n 3 : n 4 1:3:4:4 2:5:5:10 (k =5) n 1 : n 2 : n 3 : n 4 : n 5 1:3:3:3:5 3:5:5:7:15 1:1:1:3:3 1:1:3:3:3 3:5:8:12:12 3:7:7:10:15 (k =6) n 1 : n 2 : n 3 : n 4 : n 5 : n 6 2:3:4:5:8:8 2:3:5:5:5:10 3:3:3:4:5:12 3:4:4:4:9:12 3:4:6:6:6:15 3:6:6:8:10:15 3:9:9:9:12:14 4:5:5:5:12:15 4:5:12:12:12:15 4:8:10:15:16:16 4 : 10 : 10 : 10 : 15 : 20 5 : 8 : 12 : 15 : 20 : 20 This result means that the testing equality for the means of k-groups becomes robust to nonnormality by adjusting each sample size. In the ANOVA model, this fact was reported by Box and Watson (1962) and Box and Draper (1975). They called such a design a robust design. Table 2 shows the robust design of our test statistic in the cases of k =4, 5 and 6. If the ratio of sample sizes is satisfied as in Table 2, then the test statistic becomes more robust to nonnormality of the true distribution. The modified Bartlett correction proposed by Fujikoshi (2000) does not only bring the mean close to an asymptotic mean but it also brings the variance close to an asymptotic variance. If the mean and variance of the test statistic T are expanded as (3.1) and (3.4), then the modified Bartlett corrections are given by where ( (1) T =(nˆν 1 +ˆν 2 ) log 1+ T ), ˆν 1 > 0, nˆν 1 +ˆν 2 > 0, nˆν 1 (2) T = T + T ( ˆν2 T ), ˆν 1 < 0, nˆν 1 +ˆν 2 < 0, n ˆν 1 2ˆν 1 (3) T =(nˆν 1 +ˆν 2 ) ˆν 1 = { 1 exp ( T nˆν 1 )}, for any n, ˆν 1 and ˆν 2, 2, ˆν 2 = (cd +2)ˆη 2 2(cd +4)ˆη 1, ˆη 2 2ˆη 1 2(ˆη 2 2ˆη 1 ) and ˆη 1 and ˆη 2 are consistent estimators of η 1 and η 2, respectively. If the true distribution is the normal distribution, the modified Bartlett correction is the transformation by a known monotonic increasing function. On the other hand, in general, it is necessary to estimate skewnesses and kurtosis in order to use the modified Bartlett correction. However, from Theorem 3.2, we can see that

14 148 HIROKAZU YANAGIHARA Table 3. Actual test sizes of several test statistics for testing the equality (LR). Nominal Test Distribution Size m (1) 4 Statistic (i) (ii) (iii) (iv) (v) (vi) Average S.D. 10% 6.0 T T T T T T % 6.0 T T T T T T % 6.0 T T T T T T the modified Bartlett correction in the normal case can improve the quality of the chi-square approximation, even if the true distribution is not the normal distribution; this is because η 1 and η 2 are independent of the cumulants of ε. Therefore, we obtain the following corollary. Corollary 3.5. Let T denote the test statistics adjusted by the modified Bartlett corrections in the normal case as cd +2 {n + 12 } { T c + d +1 (c d 1) log 1+ c + d +1 } n(cd +2) T LH = T BNP + 1 ( c d 1+ c + d +1 ). 2n cd +2 T BNP T BNP Suppose that either Condition 1 or 2 holds, and also Condition 3 holds simultaneously, then the mean and variance of T are expanded as E[ T ]=cd + o(n 1 ) and Var[ T ]=2cd + o(n 1 ). This naturally implies that the modified Bartlett correction in the normal case can improve the quality of the chi-square approximation even under nonnormality. From equations (3.1) and (3.4), when either Condition 1 or 2 holds, and also Condition 3 holds, the mean and the second moment of T LR in the normal case are expanded as { E[T LR ]=cd 1 1 } (c d 1) + o(n 1 ), 2n

15 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 149 Table 4. Actual test sizes of several test statistics for testing the equality (LH). Nominal Test Distribution Size m (1) 4 Statistic (i) (ii) (iii) (iv) (v) (vi) Average S.D. 10% 6.0 T T T T T T T T T % 6.0 T T T T T T T T T % 6.0 T T T T T T T T T E[TLR] 2 =cd(cd +2) {1 1n } (c d 1) + o(n 1 ). These expansions make η 2 = 2η 1. Therefore, we can see that the modified Bartlett correction in the LR test statistic cannot be defined in the normal case. However, the Bartlett correction of the LR test statistic does not only bring a mean close to an asymptotic mean, but it also brings the null distribution close to the chi-square distribution, when the true distribution is also the normal distribution (see e.g., Barndorff-Nielsen and Cox (1984), and Barndorff-Nielsen and Hall (1988)). Therefore, if the true distribution is the normal distribution, the Bartlett correction in the normal case can improve the quality of the chi-square approximation to the same extent as the modified Bartlett correction, although the modified Bartlett correction cannot be defined. However, even under nonnormality, T LR adjusted by the Bartlett correction in the normal case also can

16 150 HIROKAZU YANAGIHARA Table 5. Actual test sizes of several test statistics for testing the equality (BNP). Nominal Test Distribution Size m (1) 4 Statistic (i) (ii) (iii) (iv) (v) (vi) Average S.D. 10% 6.0 T T T T T T T T T % 6.0 T T T T T T T T T % 6.0 T T T T T T T T T improve the quality of the chi-square approximation as in the following corollary (the proof is given in Appendix A.4). Corollary 3.6. Let T LR be the version of T LR adjusted by the Bartlett correction in the normal case. Suppose that either Condition 1 or 2 holds, and also Condition 3 holds simultaneously. Then, T LR can improve the quality of the chi-square approximation even under nonnormality as in the modified Bartlett correction, i.e., E[ T LR ]=cd + o(n 1 ) and Var[ T LR ]=2cd + o(n 1 ). Before concluding this subsection, we verify the validity of our theorem through a numerical experiment. The error distributions used are described in Appendix A.1. We dealt with the test for equality of Ξ in the case of p =2, k = 4 and n = 32. The p (p 1) within-individuals design matrix X was generated from the uniform distribution on ( 1, 1) and the coefficient matrices

17 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 151 for hypothesis were C =(I k 1 1 k 1 ) and D = I q. We prepared the three test statistics specified by the type II design matrix A in (3.3), as follows: Case 1: n 1 = n 2 = n 3 = n 4 =8(m (1) 4 = 6.0), Case 2: n 1 = n 2 =4,n 3 = 8 and n 4 =16(m (1) 4 =0.0), Case 3: n 1 =3,n 2 = n 3 = 4 and n 4 =21(m (1) 4 =6.2). Moreover, we adjusted each test statistic by the Bartlett correction and the modified Bartlett correction. It is easy to see that the three test statistics satisfy Condition 1, and also Condition 3 when A is Case 2. Tables 3, 4 and 5 show actual test sizes of the LR test statistic, HL test statistic and BNP test statistic, respectively. The average and S.D. of actual test sizes for every true distribution are also given in these tables. From these tables, we can see that the S.D. becomes small when Condition 3 holds. Therefore, we can say that a variation of the actual test size by the difference in distributions is small in this case. Then, the modified Bartlett correction in the normal case improves the quality of the chi-square approximation. 4. Discussion and conclusion In this paper, we presented the conditions for robustness to nonnormality. Our conclusion is as follows. First, if either Condition 1 or 2 holds, the cumulants of the true distribution vanish in the second term of the asymptotic expansion for the mean of the test statistic T. This results in the fact that the location of the null distribution of T does not easily change due to nonnormality of the true distribution. In this case, the Bartlett correction without estimating several cumulants improves the quality of the chi-square approximation even under nonnormality. Furthermore, since Condition 1 does not depend on the unknown true distribution, only with Condition 1, it becomes possible to see if the test statistic used is robust to nonnormality of the true distribution. This is an important and useful point of Condition 1. Second, if Condition 3 holds with either Condition 1 or 2, the cumulants of the true distribution vanish in the second terms of the asymptotic expansions for both the mean and variance of the test statistic T. The location and dispersion of the null distribution of T, thus do not easily change due to nonnormality of the true distribution. We use the modified Bartlett correction without estimating several cumulants in this case. By adjusting the ratio of sample sizes in the k- groups, the test statistic for the equality of several means becomes more robust to nonnormality of the true distribution. However, as Box and Watson (1962) suggested, this does not mean that one needs to seek unequal sample sizes to reduce the effect of nonnormality on the test statistic for comparison of means in the k-groups. This is mainly because unequal sample sizes would result in the increase in sensitivity of variance inequalities. Moreover, unequal sample sizes do not result in an optimal design for estimating variance (see Guiard et al. (2000)). All in all, if Condition 1 holds, the test statistic becomes robust to nonnormality of the true distribution. Therefore, it is encouraged to use the betweenindividuals design matrix with the segment and adjust its test statistic by the

18 152 HIROKAZU YANAGIHARA Bartlett correction in the normal case. It is also true that our conditions depend only on the between-individuals design matrix A and the coefficient matrix for hypothesis C. Therefore, the robustness to nonnormality on the test in a GMANOVA model is independent of both the within-individuals design matrix X and the coefficient matrix for hypothesis D. As a result, our conditions essentially coincide with the conditions of testing H 0 : CΞ = O c p in a multivariate linear model as in Wakaki et al. (2002). Appendix Used distributions as true distributions In this subsection, we describe the used distributions as the true distributions for the numerical studies in our paper. The elements of ε are independently and identically distributed and generated from the following six distributions: (i) Normal Distribution: ε j N(0, 1), (κ (1) 3,3 = κ(2) 3,3 = 0 and κ(1) 4 = 0). (ii) Laplace Distribution: ε j is generated from the Laplace distribution with mean 0 and standard deviation 1 (κ (1) 3,3 = κ(2) 3,3 = 0 and κ(1) 4 =3p). (iii) Uniform Distribution: ε j is generated from the uniform distribution on ( 5, 5) divided by the standard deviation 5/ 3(κ (1) 3,3 = κ(2) 3,3 = 0 and κ(1) 4 = 1.2p). (iv) Skew-Laplace Distribution: ε j is generated from the skew-laplace distribution with location parameter 0, dispersion parameter 1 and skew parameter 1 standardized by mean 3/4 and standard deviation 23/4 (κ (1) 3,3 = κ(2) 3,3 1.12p and κ (1) p). (v) Chi-Square Distribution: ε j is generated from the chi-square distribution with 2 degrees of freedom standardized by mean 2 and standard deviation 2(κ (1) 3,3 = κ(2) 3,3 =2p and κ(1) 4 =6p). (vi) Log-Normal Distribution: ε j is generated from the lognormal distribution such that log ε j N(0, 1), standardized by mean e 1/2 and standard deviation e(e 1) (κ (1) 3,3 = κ(2) 3,3 6.18p and κ(1) p). The skew-laplace distribution was proposed by Balakrishnan and Ambagaspitiya (1994) (for the probability density function, see e.g., Yanagihara and Yuan (2005)). It is easy to see that distributions 1, 2 and 3 are symmetric distributions, and distributions 4, 5 and 6 are skewed distributions. Proof of Corollary 3.1. In this subsection, we give the proof of Corollary 3.1. If the distribution of ε possesses orthant symmetry, the third multivariate moment of ε is given by µ abc = E[ε a ε b ε c ]= ε a ε b ε c f(ε 1,...,ε p )dε, where f( ) is a density function of ε. Since the distribution of ε possesses orthant symmetry, the equation f(ε 1,...,ε l,...,ε p )=f(ε 1,..., ε l,...,ε p ) holds for

19 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 153 any ε l s. Let g(ε a,ε b,ε c ) be the marginal distribution of ε a, ε b and ε c. Therefore, µ abc = 0 ε a ε b (ε c ε c )g(ε a,ε b,ε c )dε a dε b dε c =0. Notice that µ abc = κ abc. We obtain κ abc = 0 under an orthant symmetric distribution. Since all γj 2(,, ) consist of κ abc, it implies that all γj 2 (,, ) =0 under an orthant symmetric distribution (j = 1, 2). Proof of Corollary 3.2. In this subsection, we give the proof of Corollary 3.2. For the proof of Corollary 3.2, it is sufficient to confirm C(A A) 1 A 1 n = 0 c. When the between-individuals design matrix A is type I, from the general formula of an inverse matrix (see e.g., Siotani et al. (1985, p. 592)), (A A) 1 is given by ( ) (A A) 1 = na (A A ) 1 w (A A ) 1 A, 1 n W where w = n 1 np A 1 n and W = w(a A ) 1 +(A A ) 1 A 1 n 1 na (A A ) 1. If C =(0 c C ), where C is any c (k 1) matrix with the rank c, then, C(A A) 1 A 1 n = 1 w C {(A A ) 1 A 1 n 1 n WA }1 n = 1 w (n w 1 np A 1 n )C (A A ) 1 A 1 n = 0 c. On the other hand, when the between-individuals design matrix A is type II, (A A) 1 = diag(n 1 1,...,n 1 k ) and A 1 n =(n 1,...,n k ). Notice that C1 k = 0 c. Therefore, C(A A) 1 A 1 n = C1 k = 0 c. From these equations, we obtain the proof of Corollary 3.2. Proof of Corollary 3.6. In this subsection, we give the proof of Corollary 3.6. From the equations (3.5), if either Condition 1 or 2 holds, and also Condition 3 holds, the variance of T LR is expanded as Var[T LR ]=2cd {1 1n } (c d 1) + o(n 1 ). It is known that Var[ T LR ]= ( 1+ c d 1 ) 2 Var[T LR] =2cd + o(n 1 ). 2n Therefore, we obtain the proof of Corollary 3.6.

20 154 HIROKAZU YANAGIHARA Acknowledgements The author would like to thank Professor Y. Fujikoshi, Chuo University; Professor H. Wakaki, Hiroshima University; Dr. H. Fujisawa, The Institute of Statistical Mathematics; Dr. K.-H. Yuan, University of Notre Dame; Ms. M. Yamamura, University of Tsukuba; the referee and the associate editor for their valuable advice and encouragement. The author also would like to thank Dr. A. Yoshimoto, Tohoku University, for comments leading to improvements of the original manuscript. This research was supported by the Ministry of Education, Science, Sports and Culture Grant-in-Aid for Young Scientists (B), # , References Balakrishnan, N. and Ambagaspitiya, R. S. (1994). On skew Laplace distribution, Technical Report, Department of Mathematics & Statistics, McMaster University, Hamilton, Ontario, Canada. Barndorff-Nielsen, O. E. and Cox, D. R. (1984). Bartlett adjustments to the likelihood ratio statistic and the distribution of the maximum likelihood estimator, J. Roy. Statist. Soc. Ser. B, 46, Barndorff-Nielsen, O. E. and Hall, P. (1988). On the level-error after Bartlett adjustment of the likelihood ratio statistic, Biometrika, 75, Bartlett, M. S. (1937). Properties of sufficiency and statistical tests, Proc. Roy. Soc. London Ser. A, 160, Box, G. E. P. and Draper, N. R. (1975). Robust designs, Biometrika, 62, Box, G. E. P. and Watson, G. S. (1962). Robustness to non-normality of regression tests, Biometrika, 49, Chase, G. R. and Bulgren, W. G. (1971). A Monte Carlo investigations of the robustness of T 2, J. Amer. Statist. Assoc., 66, Dawid, A. P. (1977). Spherical matrix distributions and a multivariate model, J. Roy. Statist. Soc. Ser. B, 39, Eaton, M. L. and Efron, B. (1970). Hotelling s T 2 test under symmetry conditions, J. Amer. Statist. Assoc., 65, Efron, B. (1969). Student s t-test under symmetry conditions, J. Amer. Statist. Assoc., 64, Everitt, B. S. (1979). A Monte Carlo investigation of the robustness of Hotelling s one- and two-sample T 2 tests, J. Amer. Statist. Assoc., 74, Fujikoshi, Y. (1993). The growth curve model theory and its application, Math. Sci., 358, (in Japanese). Fujikoshi, Y. (2000). Transformations with improved chi-squared approximations, J. Multivariate Anal., 72, Guiard, V., Herrendörfer, G., Sumpf, D. and Nürnberg, G. (2000). About optimal designs for estimating variance components with ANOVA in one-way-classification under non-normality, J. Statist. Plann. Inference, 89, Isogai, T. (1983). On measures of multivariate skewness and kurtosis, Math. Japon., 28, Kariya, T. (1978). The general MANOVA problem, Ann. Statist., 6, Kariya, T. (1981a). A robustness property of Hotelling s T 2 -test, Ann. Statist., 9, Kariya, T. (1981b). Robustness of multivariate tests, Ann. Statist., 9, Khatri, C. G. (1988). Robustness study for a linear growth model, J. Multivariate Anal., 24, Kshirsagar, A. M. and Smith, W. B. (1995). Growth Curves, Marcel Dekker, Inc., New York, Basel, Hong Kong.

21 ROBUSTNESS TO NONNORMALITY ON GMANOVA TESTS 155 Mardia, K. V. (1970). Measures of multivariate skewness and kurtosis with applications, Biometrika, 57, Potthoff, R. F. and Roy, S. N. (1964). A generalized multivariate analysis of variance model useful especially for growth curve problems, Biometrika, 51, Siotani, M., Hayakawa, T. and Fujikoshi, Y. (1985), Modern Multivariate Statistical Analysis: A Graduate Course and Handbook, American Sciences Press, Columbus, Ohio. Srivastava, M. S. and von Rosen, D. (1999). Growth curve models, Multivariate Analysis, Design of Experiments, and Survey Sampling (ed. S. Ghosh), , Marcel Dekker, New York. von Rosen, D. (1991). The growth curve model: A review, Comm. Statist. Theory Methods, 20, Wakaki, H., Yanagihara, H. and Fujikoshi, Y. (2002). Asymptotic expansion of the null distributions of test statistics for multivariate linear hypothesis under nonnormality, Hiroshima Math. J., 32, Yanagihara, H. (2001). Asymptotic expansions of the null distributions of three test statistics in a nonnormal GMANOVA model, Hiroshima Math. J., 31, Yanagihara, H. (2007). A family of estimators for multivariate kurtosis in a nonnormal linear regression model, J. Multivariate Anal., 98, Yanagihara, H. and Yuan, K.-H. (2005). Four improved statistics for contrasting means by correcting skewness and kurtosis, British J. Math. Statist. Psych., 58,

Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population

Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population Edgeworth Expansions of Functions of the Sample Covariance Matrix with an Unknown Population (Last Modified: April 24, 2008) Hirokazu Yanagihara 1 and Ke-Hai Yuan 2 1 Department of Mathematics, Graduate

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

An Unbiased C p Criterion for Multivariate Ridge Regression

An Unbiased C p Criterion for Multivariate Ridge Regression An Unbiased C p Criterion for Multivariate Ridge Regression (Last Modified: March 7, 2008) Hirokazu Yanagihara 1 and Kenichi Satoh 2 1 Department of Mathematics, Graduate School of Science, Hiroshima University

More information

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems

More information

An Introduction to Multivariate Statistical Analysis

An Introduction to Multivariate Statistical Analysis An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents

More information

High-dimensional asymptotic expansions for the distributions of canonical correlations

High-dimensional asymptotic expansions for the distributions of canonical correlations Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic

More information

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large Tetsuji Tonda 1 Tomoyuki Nakagawa and Hirofumi Wakaki Last modified: March 30 016 1 Faculty of Management and Information

More information

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control Takahiro Nishiyama a a Department of Mathematical Information Science, Tokyo University of Science,

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi Faculty of Science and Engineering, Chuo University, Kasuga, Bunkyo-ku,

More information

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors Takashi Seo a, Takahiro Nishiyama b a Department of Mathematical Information Science, Tokyo University

More information

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples

Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Nobumichi Shutoh, Masashi Hyodo and Takashi Seo 2 Department of Mathematics, Graduate

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

Likelihood Ratio Tests in Multivariate Linear Model

Likelihood Ratio Tests in Multivariate Linear Model Likelihood Ratio Tests in Multivariate Linear Model Yasunori Fujikoshi Department of Mathematics, Graduate School of Science, Hiroshima University, 1-3-1 Kagamiyama, Higashi Hiroshima, Hiroshima 739-8626,

More information

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives TR-No. 14-06, Hiroshima Statistical Research Group, 1 11 Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives Mariko Yamamura 1, Keisuke Fukui

More information

A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification

A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification TR-No. 13-08, Hiroshima Statistical Research Group, 1 23 A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification

More information

Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions

Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions TR-No. 14-07, Hiroshima Statistical Research Group, 1 12 Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions Keisuke Fukui 2, Mariko Yamamura 1 and Hirokazu Yanagihara 2 1

More information

Testing for a unit root in an ar(1) model using three and four moment approximations: symmetric distributions

Testing for a unit root in an ar(1) model using three and four moment approximations: symmetric distributions Hong Kong Baptist University HKBU Institutional Repository Department of Economics Journal Articles Department of Economics 1998 Testing for a unit root in an ar(1) model using three and four moment approximations:

More information

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA

More information

Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification

Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification (Last Modified: May 22, 2012) Yusuke Hashiyama, Hirokazu Yanagihara 1 and Yasunori

More information

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA Kazuyuki Koizumi Department of Mathematics, Graduate School of Science Tokyo University of Science 1-3, Kagurazaka,

More information

arxiv: v1 [math.st] 2 Apr 2016

arxiv: v1 [math.st] 2 Apr 2016 NON-ASYMPTOTIC RESULTS FOR CORNISH-FISHER EXPANSIONS V.V. ULYANOV, M. AOSHIMA, AND Y. FUJIKOSHI arxiv:1604.00539v1 [math.st] 2 Apr 2016 Abstract. We get the computable error bounds for generalized Cornish-Fisher

More information

Bias-corrected AIC for selecting variables in Poisson regression models

Bias-corrected AIC for selecting variables in Poisson regression models Bias-corrected AIC for selecting variables in Poisson regression models Ken-ichi Kamo (a), Hirokazu Yanagihara (b) and Kenichi Satoh (c) (a) Corresponding author: Department of Liberal Arts and Sciences,

More information

Analysis of variance and linear contrasts in experimental design with generalized secant hyperbolic distribution

Analysis of variance and linear contrasts in experimental design with generalized secant hyperbolic distribution Journal of Computational and Applied Mathematics 216 (2008) 545 553 www.elsevier.com/locate/cam Analysis of variance and linear contrasts in experimental design with generalized secant hyperbolic distribution

More information

Selection of Model Selection Criteria for Multivariate Ridge Regression

Selection of Model Selection Criteria for Multivariate Ridge Regression Selection of Model Selection Criteria for Multivariate Ridge Regression (Last Modified: January, 24) Isamu Nagai, Department of Mathematics, Graduate School of Science, Hiroshima University -3- Kagamiyama,

More information

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE J Jaan Statist Soc Vol 34 No 2004 9 26 ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE Yasunori Fujikoshi*, Tetsuto Himeno

More information

Analysis of variance, multivariate (MANOVA)

Analysis of variance, multivariate (MANOVA) Analysis of variance, multivariate (MANOVA) Abstract: A designed experiment is set up in which the system studied is under the control of an investigator. The individuals, the treatments, the variables

More information

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Tetsuro Sakurai and Yasunori Fujikoshi Center of General Education, Tokyo University

More information

T 2 Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem

T 2 Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem T Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem Toshiki aito a, Tamae Kawasaki b and Takashi Seo b a Department of Applied Mathematics, Graduate School

More information

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order

More information

On testing the equality of mean vectors in high dimension

On testing the equality of mean vectors in high dimension ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.

More information

High-dimensional two-sample tests under strongly spiked eigenvalue models

High-dimensional two-sample tests under strongly spiked eigenvalue models 1 High-dimensional two-sample tests under strongly spiked eigenvalue models Makoto Aoshima and Kazuyoshi Yata University of Tsukuba Abstract: We consider a new two-sample test for high-dimensional data

More information

YUN WU. B.S., Gui Zhou University of Finance and Economics, 1996 A REPORT. Submitted in partial fulfillment of the requirements for the degree

YUN WU. B.S., Gui Zhou University of Finance and Economics, 1996 A REPORT. Submitted in partial fulfillment of the requirements for the degree A SIMULATION STUDY OF THE ROBUSTNESS OF HOTELLING S T TEST FOR THE MEAN OF A MULTIVARIATE DISTRIBUTION WHEN SAMPLING FROM A MULTIVARIATE SKEW-NORMAL DISTRIBUTION by YUN WU B.S., Gui Zhou University of

More information

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Takahiro Nishiyama a,, Masashi Hyodo b, Takashi Seo a, Tatjana Pavlenko c a Department of Mathematical

More information

Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions

Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions c 215 FORMATH Research Group FORMATH Vol. 14 (215): 27 39, DOI:1.15684/formath.14.4 Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions Keisuke Fukui 1,

More information

Ken-ichi Kamo Department of Liberal Arts and Sciences, Sapporo Medical University, Hokkaido, Japan

Ken-ichi Kamo Department of Liberal Arts and Sciences, Sapporo Medical University, Hokkaido, Japan A STUDY ON THE BIAS-CORRECTION EFFECT OF THE AIC FOR SELECTING VARIABLES IN NOR- MAL MULTIVARIATE LINEAR REGRESSION MOD- ELS UNDER MODEL MISSPECIFICATION Authors: Hirokazu Yanagihara Department of Mathematics,

More information

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Econometrics Working Paper EWP0402 ISSN 1485-6441 Department of Economics TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Lauren Bin Dong & David E. A. Giles Department

More information

MULTIVARIATE ANALYSIS OF VARIANCE UNDER MULTIPLICITY José A. Díaz-García. Comunicación Técnica No I-07-13/ (PE/CIMAT)

MULTIVARIATE ANALYSIS OF VARIANCE UNDER MULTIPLICITY José A. Díaz-García. Comunicación Técnica No I-07-13/ (PE/CIMAT) MULTIVARIATE ANALYSIS OF VARIANCE UNDER MULTIPLICITY José A. Díaz-García Comunicación Técnica No I-07-13/11-09-2007 (PE/CIMAT) Multivariate analysis of variance under multiplicity José A. Díaz-García Universidad

More information

Point and Interval Estimation for Gaussian Distribution, Based on Progressively Type-II Censored Samples

Point and Interval Estimation for Gaussian Distribution, Based on Progressively Type-II Censored Samples 90 IEEE TRANSACTIONS ON RELIABILITY, VOL. 52, NO. 1, MARCH 2003 Point and Interval Estimation for Gaussian Distribution, Based on Progressively Type-II Censored Samples N. Balakrishnan, N. Kannan, C. T.

More information

Estimation of parametric functions in Downton s bivariate exponential distribution

Estimation of parametric functions in Downton s bivariate exponential distribution Estimation of parametric functions in Downton s bivariate exponential distribution George Iliopoulos Department of Mathematics University of the Aegean 83200 Karlovasi, Samos, Greece e-mail: geh@aegean.gr

More information

Chapter 7, continued: MANOVA

Chapter 7, continued: MANOVA Chapter 7, continued: MANOVA The Multivariate Analysis of Variance (MANOVA) technique extends Hotelling T 2 test that compares two mean vectors to the setting in which there are m 2 groups. We wish to

More information

Application of Parametric Homogeneity of Variances Tests under Violation of Classical Assumption

Application of Parametric Homogeneity of Variances Tests under Violation of Classical Assumption Application of Parametric Homogeneity of Variances Tests under Violation of Classical Assumption Alisa A. Gorbunova and Boris Yu. Lemeshko Novosibirsk State Technical University Department of Applied Mathematics,

More information

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work Hiroshima Statistical Research Group: Technical Report Approximate interval estimation for PMC for improved linear discriminant rule under high dimensional frame work Masashi Hyodo, Tomohiro Mitani, Tetsuto

More information

ACCURATE ASYMPTOTIC ANALYSIS FOR JOHN S TEST IN MULTICHANNEL SIGNAL DETECTION

ACCURATE ASYMPTOTIC ANALYSIS FOR JOHN S TEST IN MULTICHANNEL SIGNAL DETECTION ACCURATE ASYMPTOTIC ANALYSIS FOR JOHN S TEST IN MULTICHANNEL SIGNAL DETECTION Yu-Hang Xiao, Lei Huang, Junhao Xie and H.C. So Department of Electronic and Information Engineering, Harbin Institute of Technology,

More information

On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly Unrelated Regression Model and a Heteroscedastic Model

On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly Unrelated Regression Model and a Heteroscedastic Model Journal of Multivariate Analysis 70, 8694 (1999) Article ID jmva.1999.1817, available online at http:www.idealibrary.com on On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly

More information

STAT 730 Chapter 5: Hypothesis Testing

STAT 730 Chapter 5: Hypothesis Testing STAT 730 Chapter 5: Hypothesis Testing Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 28 Likelihood ratio test def n: Data X depend on θ. The

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

The Third International Workshop in Sequential Methodologies

The Third International Workshop in Sequential Methodologies Area C.6.1: Wednesday, June 15, 4:00pm Kazuyoshi Yata Institute of Mathematics, University of Tsukuba, Japan Effective PCA for large p, small n context with sample size determination In recent years, substantial

More information

Lecture 5: Hypothesis tests for more than one sample

Lecture 5: Hypothesis tests for more than one sample 1/23 Lecture 5: Hypothesis tests for more than one sample Måns Thulin Department of Mathematics, Uppsala University thulin@math.uu.se Multivariate Methods 8/4 2011 2/23 Outline Paired comparisons Repeated

More information

POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE

POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE Supported by Patrick Adebayo 1 and Ahmed Ibrahim 1 Department of Statistics, University of Ilorin, Kwara State, Nigeria Department

More information

1 Appendix A: Matrix Algebra

1 Appendix A: Matrix Algebra Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix

More information

Analysis of Microtubules using. for Growth Curve modeling.

Analysis of Microtubules using. for Growth Curve modeling. Analysis of Microtubules using Growth Curve Modeling Md. Aleemuddin Siddiqi S. Rao Jammalamadaka Statistics and Applied Probability, University of California, Santa Barbara March 1, 2006 1 Introduction

More information

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC A Simple Approach to Inference in Random Coefficient Models March 8, 1988 Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC 27695-8203 Key Words

More information

Specification Test for Instrumental Variables Regression with Many Instruments

Specification Test for Instrumental Variables Regression with Many Instruments Specification Test for Instrumental Variables Regression with Many Instruments Yoonseok Lee and Ryo Okui April 009 Preliminary; comments are welcome Abstract This paper considers specification testing

More information

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data Sri Lankan Journal of Applied Statistics (Special Issue) Modern Statistical Methodologies in the Cutting Edge of Science Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations

More information

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS Communications in Statistics - Simulation and Computation 33 (2004) 431-446 COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS K. Krishnamoorthy and Yong Lu Department

More information

A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common

A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common Journal of Statistical Theory and Applications Volume 11, Number 1, 2012, pp. 23-45 ISSN 1538-7887 A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when

More information

On Bayesian Inference with Conjugate Priors for Scale Mixtures of Normal Distributions

On Bayesian Inference with Conjugate Priors for Scale Mixtures of Normal Distributions Journal of Applied Probability & Statistics Vol. 5, No. 1, xxx xxx c 2010 Dixie W Publishing Corporation, U. S. A. On Bayesian Inference with Conjugate Priors for Scale Mixtures of Normal Distributions

More information

CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES. D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3

CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES. D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3 CONVERTING OBSERVED LIKELIHOOD FUNCTIONS TO TAIL PROBABILITIES D.A.S. Fraser Mathematics Department York University North York, Ontario M3J 1P3 N. Reid Department of Statistics University of Toronto Toronto,

More information

EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS

EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS Statistica Sinica 18(2008), 135-154 EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS Holger Dette, Joachim Kunert and Andrey Pepelyshev Ruhr-Universität Bochum, Universität

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

HANDBOOK OF APPLICABLE MATHEMATICS

HANDBOOK OF APPLICABLE MATHEMATICS HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester

More information

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error

More information

Multivariate Normal-Laplace Distribution and Processes

Multivariate Normal-Laplace Distribution and Processes CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by

More information

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function

More information

Testing Structural Equation Models: The Effect of Kurtosis

Testing Structural Equation Models: The Effect of Kurtosis Testing Structural Equation Models: The Effect of Kurtosis Tron Foss, Karl G Jöreskog & Ulf H Olsson Norwegian School of Management October 18, 2006 Abstract Various chi-square statistics are used for

More information

A fast and consistent variable selection method for high-dimensional multivariate linear regression with a large number of explanatory variables

A fast and consistent variable selection method for high-dimensional multivariate linear regression with a large number of explanatory variables A fast and consistent variable selection method for high-dimensional multivariate linear regression with a large number of explanatory variables Ryoya Oda and Hirokazu Yanagihara Department of Mathematics,

More information

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

Application of Variance Homogeneity Tests Under Violation of Normality Assumption Application of Variance Homogeneity Tests Under Violation of Normality Assumption Alisa A. Gorbunova, Boris Yu. Lemeshko Novosibirsk State Technical University Novosibirsk, Russia e-mail: gorbunova.alisa@gmail.com

More information

Multivariate Linear Models

Multivariate Linear Models Multivariate Linear Models Stanley Sawyer Washington University November 7, 2001 1. Introduction. Suppose that we have n observations, each of which has d components. For example, we may have d measurements

More information

MARGINAL HOMOGENEITY MODEL FOR ORDERED CATEGORIES WITH OPEN ENDS IN SQUARE CONTINGENCY TABLES

MARGINAL HOMOGENEITY MODEL FOR ORDERED CATEGORIES WITH OPEN ENDS IN SQUARE CONTINGENCY TABLES REVSTAT Statistical Journal Volume 13, Number 3, November 2015, 233 243 MARGINAL HOMOGENEITY MODEL FOR ORDERED CATEGORIES WITH OPEN ENDS IN SQUARE CONTINGENCY TABLES Authors: Serpil Aktas Department of

More information

PRE-TEST ESTIMATION OF THE REGRESSION SCALE PARAMETER WITH MULTIVARIATE STUDENT-t ERRORS AND INDEPENDENT SUB-SAMPLES

PRE-TEST ESTIMATION OF THE REGRESSION SCALE PARAMETER WITH MULTIVARIATE STUDENT-t ERRORS AND INDEPENDENT SUB-SAMPLES Sankhyā : The Indian Journal of Statistics 1994, Volume, Series B, Pt.3, pp. 334 343 PRE-TEST ESTIMATION OF THE REGRESSION SCALE PARAMETER WITH MULTIVARIATE STUDENT-t ERRORS AND INDEPENDENT SUB-SAMPLES

More information

Quadratic forms in skew normal variates

Quadratic forms in skew normal variates J. Math. Anal. Appl. 73 (00) 558 564 www.academicpress.com Quadratic forms in skew normal variates Arjun K. Gupta a,,1 and Wen-Jang Huang b a Department of Mathematics and Statistics, Bowling Green State

More information

Mean squared error matrix comparison of least aquares and Stein-rule estimators for regression coefficients under non-normal disturbances

Mean squared error matrix comparison of least aquares and Stein-rule estimators for regression coefficients under non-normal disturbances METRON - International Journal of Statistics 2008, vol. LXVI, n. 3, pp. 285-298 SHALABH HELGE TOUTENBURG CHRISTIAN HEUMANN Mean squared error matrix comparison of least aquares and Stein-rule estimators

More information

Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension

Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension (Last Modified: November 3 2008) Mariko Yamamura 1, Hirokazu Yanagihara 2 and Muni S. Srivastava 3

More information

M M Cross-Over Designs

M M Cross-Over Designs Chapter 568 Cross-Over Designs Introduction This module calculates the power for an x cross-over design in which each subject receives a sequence of treatments and is measured at periods (or time points).

More information

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION J. Japan Statist. Soc. Vol. 32 No. 2 2002 165 181 A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION Hisayuki Tsukuma* The problem of estimation in multivariate linear calibration

More information

STAT 501 EXAM I NAME Spring 1999

STAT 501 EXAM I NAME Spring 1999 STAT 501 EXAM I NAME Spring 1999 Instructions: You may use only your calculator and the attached tables and formula sheet. You can detach the tables and formula sheet from the rest of this exam. Show your

More information

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data Yujun Wu, Marc G. Genton, 1 and Leonard A. Stefanski 2 Department of Biostatistics, School of Public Health, University of Medicine

More information

ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE

ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE Sankhyā : The Indian Journal of Statistics 999, Volume 6, Series B, Pt. 3, pp. 488 495 ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE By S. HUDA and A.A. AL-SHIHA King Saud University, Riyadh, Saudi Arabia

More information

Sample size calculations for logistic and Poisson regression models

Sample size calculations for logistic and Poisson regression models Biometrika (2), 88, 4, pp. 93 99 2 Biometrika Trust Printed in Great Britain Sample size calculations for logistic and Poisson regression models BY GWOWEN SHIEH Department of Management Science, National

More information

Testing equality of two mean vectors with unequal sample sizes for populations with correlation

Testing equality of two mean vectors with unequal sample sizes for populations with correlation Testing equality of two mean vectors with unequal sample sizes for populations with correlation Aya Shinozaki Naoya Okamoto 2 and Takashi Seo Department of Mathematical Information Science Tokyo University

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

MANOVA MANOVA,$/,,# ANOVA ##$%'*!# 1. $!;' *$,$!;' (''

MANOVA MANOVA,$/,,# ANOVA ##$%'*!# 1. $!;' *$,$!;' ('' 14 3! "#!$%# $# $&'('$)!! (Analysis of Variance : ANOVA) *& & "#!# +, ANOVA -& $ $ (+,$ ''$) *$#'$)!!#! (Multivariate Analysis of Variance : MANOVA).*& ANOVA *+,'$)$/*! $#/#-, $(,!0'%1)!', #($!#$ # *&,

More information

Multivariate Regression (Chapter 10)

Multivariate Regression (Chapter 10) Multivariate Regression (Chapter 10) This week we ll cover multivariate regression and maybe a bit of canonical correlation. Today we ll mostly review univariate multivariate regression. With multivariate

More information

MULTIVARIATE ANALYSIS OF VARIANCE

MULTIVARIATE ANALYSIS OF VARIANCE MULTIVARIATE ANALYSIS OF VARIANCE RAJENDER PARSAD AND L.M. BHAR Indian Agricultural Statistics Research Institute Library Avenue, New Delhi - 0 0 lmb@iasri.res.in. Introduction In many agricultural experiments,

More information

Bootstrap prediction and Bayesian prediction under misspecified models

Bootstrap prediction and Bayesian prediction under misspecified models Bernoulli 11(4), 2005, 747 758 Bootstrap prediction and Bayesian prediction under misspecified models TADAYOSHI FUSHIKI Institute of Statistical Mathematics, 4-6-7 Minami-Azabu, Minato-ku, Tokyo 106-8569,

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

GARCH Models Estimation and Inference

GARCH Models Estimation and Inference GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in

More information

HIGHER ORDER CUMULANTS OF RANDOM VECTORS, DIFFERENTIAL OPERATORS, AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES

HIGHER ORDER CUMULANTS OF RANDOM VECTORS, DIFFERENTIAL OPERATORS, AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES HIGHER ORDER CUMULANTS OF RANDOM VECTORS DIFFERENTIAL OPERATORS AND APPLICATIONS TO STATISTICAL INFERENCE AND TIME SERIES S RAO JAMMALAMADAKA T SUBBA RAO AND GYÖRGY TERDIK Abstract This paper provides

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

TOLERANCE INTERVALS FOR DISCRETE DISTRIBUTIONS IN EXPONENTIAL FAMILIES

TOLERANCE INTERVALS FOR DISCRETE DISTRIBUTIONS IN EXPONENTIAL FAMILIES Statistica Sinica 19 (2009), 905-923 TOLERANCE INTERVALS FOR DISCRETE DISTRIBUTIONS IN EXPONENTIAL FAMILIES Tianwen Tony Cai and Hsiuying Wang University of Pennsylvania and National Chiao Tung University

More information

RELATION BETWEEN SMALL FUNCTIONS WITH DIFFERENTIAL POLYNOMIALS GENERATED BY MEROMORPHIC SOLUTIONS OF HIGHER ORDER LINEAR DIFFERENTIAL EQUATIONS

RELATION BETWEEN SMALL FUNCTIONS WITH DIFFERENTIAL POLYNOMIALS GENERATED BY MEROMORPHIC SOLUTIONS OF HIGHER ORDER LINEAR DIFFERENTIAL EQUATIONS Kragujevac Journal of Mathematics Volume 38(1) (2014), Pages 147 161. RELATION BETWEEN SMALL FUNCTIONS WITH DIFFERENTIAL POLYNOMIALS GENERATED BY MEROMORPHIC SOLUTIONS OF HIGHER ORDER LINEAR DIFFERENTIAL

More information

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω)

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω) Online Appendix Proof of Lemma A.. he proof uses similar arguments as in Dunsmuir 979), but allowing for weak identification and selecting a subset of frequencies using W ω). It consists of two steps.

More information

Tests for Covariance Matrices, particularly for High-dimensional Data

Tests for Covariance Matrices, particularly for High-dimensional Data M. Rauf Ahmad Tests for Covariance Matrices, particularly for High-dimensional Data Technical Report Number 09, 200 Department of Statistics University of Munich http://www.stat.uni-muenchen.de Tests for

More information

On Selecting Tests for Equality of Two Normal Mean Vectors

On Selecting Tests for Equality of Two Normal Mean Vectors MULTIVARIATE BEHAVIORAL RESEARCH, 41(4), 533 548 Copyright 006, Lawrence Erlbaum Associates, Inc. On Selecting Tests for Equality of Two Normal Mean Vectors K. Krishnamoorthy and Yanping Xia Department

More information

Application of Ghosh, Grizzle and Sen s Nonparametric Methods in. Longitudinal Studies Using SAS PROC GLM

Application of Ghosh, Grizzle and Sen s Nonparametric Methods in. Longitudinal Studies Using SAS PROC GLM Application of Ghosh, Grizzle and Sen s Nonparametric Methods in Longitudinal Studies Using SAS PROC GLM Chan Zeng and Gary O. Zerbe Department of Preventive Medicine and Biometrics University of Colorado

More information