Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means

Size: px
Start display at page:

Download "Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means"

Transcription

1 Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means. One popular test is the so-called Hotelling 2 test. However, as the dimension diverges, the Hotelling 2 test performs poorly due to the possible inconsistency of the sample covariance estimation. o overcome this issue and allow the dimension to diverge as fast as possible, Bai and Saranadasa (996) and Chen and Qin (200) proposed tests without the sample covariance involved, and derived the asymptotic limits which depend on whether the dimension is fixed or diverges under a specific multivariate model. In this paper, we propose a jackknife empirical likelihood test which has a chi-square limit independent of the dimension, and the conditions are much weaker than those in the existing methods. A simulation study shows that the proposed new test has a very robust size with respect to the dimension, and is powerful too. Keywords: Jackknife empirical likelihood, high dimensional mean, hypothesis test. Introduction Suppose X = (X,,, X,d ),, X n = (X n,,, X n,d) and Y = (Y,,, Y,d ),, Y n2 = (Y n2,,, Y n2,d) are two independent random samples with means µ and µ 2, respectively. It has been a long history to test H 0 : µ = µ 2 against H a : µ µ 2 for a fixed dimension d. When both X and Y have a multivariate normal distribution with equal covariance, the well-known test is the School of Mathematics, Georgia Institute of echnology, Atlanta, GA , USA. ruodu.wang@math.gatech.edu, peng@math.gatech.edu 2 Department of Mathematics and Statistics, University of Minnesota Duluth, 7 University Drive, Duluth, MN 5582, USA. yqi@d.umn.edu

2 so-called Hotelling 2 test defined as 2 = η( X Ȳ) A n ( X Ȳ), () where η = (n +n 2 2)n n 2 n +n 2, X = n n X i, Ȳ = n 2 n2 Y i and A n = n (X i X)(X i X) + n2 (Y i Ȳ)(Y i Ȳ). However, when d = d(n, n 2 ), the Hotelling 2 test performs poorly due to the possible inconsistency of the sample covariance estimation. When d/(n + n 2 ) c (0, ), Bai and Saranadasa (996) derived the asymptotic power of 2. o overcome the restriction c <, Bai and Saranadasa (996) proposed to employ M n = ( X Ȳ) ( X Ȳ) η tr(a n ) instead of 2 under a special multivariate model without assuming multivariate normality while keeping the condition of equal covariance, and derived the asymptotic limit when d/n c > 0. Recently Chen and Qin (200) proposed to use the following test statistic CQ = n i j X i X n2 j n (n ) + i j Y i Y j n 2 (n 2 ) 2 n n2 X i Y j (2) n n 2 in order to allow d to be a possible larger order than that in Bai and Saranadasa (996). Again, the asymptotic limit of the proposed test statistic CQ depends on whether the dimension is fixed or diverges, which results in either a normal limit or a chi-square limit, and special models for {X i } and {Y i } are employed. Another modification of Hotelling 2 test is proposed by Srivastava and Du (2008) and Srivastava (2009) with the covariance matrix replaced by a diagonal matrix. Rates of convergence for high dimensional means are studied by Kuelbs and Vidyashankar (200). For nonasymptotic studies of high dimensional means, we refer to Arlot, Blanchard and Roquain (200a,b). Here, we are interested in seeking a test which does not need to distinguish whether the dimension is fixed or diverges. By noting that µ = µ 2 is equivalent to (µ µ 2 ) (µ µ 2 ) = 0, one may think of applying an empirical likelihood test to the estimating equation E{(X i Y j ) (X i2 Y j2 )} = 0 for i i 2 2

3 and j j 2. If one directly applies the empirical likelihood method based on estimating equations proposed in Qin and Lawless (994) by using the samples X,, X n and Y,, Y n2, then one may define the empirical likelihood function as sup{{ n (n p i )}{ n 2 (n 2q j )} : p 0,, p n 0, q 0,, q n2 0, n p i =, n2 q j =, n n2 i = i 2 i j = j 2 j (p i X i q j Y j ) (p i2 X i2 q j2 Y j2 ) = 0}, which makes the minimization unsolvable. he reason is that the estimating equation defines a nonlinear functional, and in general one has to linearize the nonlinear functional before applying the empirical likelihood method. For more details on empirical likelihood methods, we refer to Owen (200) and the review paper of Chen and Van Keilegom (2009). Recently, Jing, Yuan and Zhou (2009) proposed a so-called jackknife empirical likelihood method to construct confidence regions for nonlinear functionals with a particular focus on U-statistics. Using this idea, one needs to construct a jackknife sample based on the following estimator n (n ) n 2 (n 2 ) n i i 2 n2 j j 2 (X i Y j ) (X i2 Y j2 ), which equals the statistic CQ given in (2). However, in order to have the jackknife empirical likelihood method work, one has to show that n n 2 n (n ) n 2 (n 2 ) n i i 2 n2 j j 2 (X i Y j ) (X i2 Y j2 ) has a normal limit when µ = µ 2. Consider n = n 2 = n, d = and µ = µ 2. hen it is easy to see that = n (n ) 2 n i i 2 n n n { (X i Y i )} 2 n d (χ 2 )E(X Y ) 2 j j 2 (X i Y j ) (X i2 Y j2 ) n (X i Y i ) n(n ) n X i n Y j 2 (n ) n X i Y i as n, where χ 2 denotes a random variable having a chi-square distribution with degree of freedom. Obviously, the above limit is not normally distributed. Hence a direct application of the jackknife empirical likelihood method to the statistic CQ will not lead to a chi-square limit. In this paper, we propose a novel way to formulate a jackknife empirical likelihood test for testing H 0 : µ = µ 2 against H a : µ µ 2 by dividing the samples into two parts. he proposed new test 3

4 has no need to distinguish whether the dimension is fixed or goes to infinity. It turns out that the asymptotic limit of the new test under H 0 is a chi-square limit independent of the dimension, the conditions on p and random vectors {X i } and {Y j } are weaker too. A simulation study shows that the size of the new test is quite stable with respect to the dimension and the proposed test is powerful as well. We organize this paper as follows. In Section 2, the new methodology and main results are given. Section 3 presents a simulation study and a real data analysis. All proofs are put in Section 4. 2 Methodology hroughout assume X i = (X i,,, X i,d ) for i =,, n and Y j = (Y j,,, Y j,d ) for j =,, n 2 are two independent random samples with means µ and µ 2, respectively. Assume min{n, n 2 } goes to infinity. he question is to test H 0 : µ = µ 2 against H a : µ µ 2. Since µ = µ 2 is equivalent to (µ µ 2 ) (µ µ 2 ) = 0 and E(X i Y j ) (X i2 Y j2 ) = (µ µ 2 ) (µ µ 2 ) for i i 2 and j j 2, we propose to apply the jackknife empirical likelihood method to the above estimating equation. As explained in the introduction, a direct application fails to have a chi-square limit. Here we propose to split the samples into two groups as follows. Put m = [n /2], m 2 = [n 2 /2], m = m + m 2, Xi = X i+m for i =,, m, and Ȳj = Y j+m2 for j =,, m 2. First we propose to estimate (µ µ 2 ) (µ µ 2 ) by m m 2 (X i Y j ) ( m m X i Ȳj), (3) 2 which is less efficient than the statistic CQ. However, it allows us to add more estimating equations and to employ the empirical likelihood method without estimating the asymptotic covariance. By noting that E{(X i Y j ) ( X i Ȳj)} = (µ µ 2 ) (µ µ 2 ) = µ µ 2 2 instead of O( µ µ 2 ), one may expect that a test based on (3) will not be powerful for a small value of µ µ 2, confirmed by a brief simulation study. In order to improve the power, we propose to apply the jackknife empirical 4

5 likelihood method in Jing, Yuan and Zhou (2009) to both (3) and a linear functional such as m m 2 m m 2 {α (X i Y j ) + α ( X i Ȳj)} (4) rather than only (3), where α R d is a vector chosen based on prior information. heoretically, when no additional information is available, any linear functional is a possible choice and α = (,, ) R d is a convenient one. Note that more linear functionals in equation (4) with different choices of α can be added to further improve the power. In the simulation study we apply the jackknife empirical likelihood to (3) and (4) with α = (,, ) R d, which results in a test with good power and quite robust size with respect to the dimension. As in Jing, Yuan and Zhou (2009), based on (3) and (4), we formulate the jackknife sample as Z k = (Z k,, Z k,2 ) for k =,, m, where Z k, = m +m 2 m m 2 Z k,2 = m +m 2 m m 2 for k =,, m, and Z k, = m +m 2 m m 2 m m2 m +m 2 (m )m 2 m i k, (X i Y j ) ( X i Ȳj) m2 (X i Y j ) ( X i Ȳj) m m2 {α (X i Y j ) + α ( X i Ȳj)} m +m 2 (m )m 2 m i k, m +m 2 m (m 2 ) Z k,2 = m +m 2 m m 2 m m2 m m +m 2 m (m 2 ) m m2 m m2 {α (X i Y j ) + α ( X i Ȳj)} (X i Y j ) ( X i Ȳj) m2 j k m, (X i Y j ) ( X i Ȳj) {α (X i Y j ) + α ( X i Ȳj)} m2 j k m, {α (X i Y j ) + α ( X i Ȳj)} for k = m +,, m. Based on this jackknife sample, the jackknife empirical likelihood ratio function for testing H 0 : µ = µ 2 is defined as m m m L m = sup{ (mp i ) : p 0,, p m 0, p i =, p i Z i = (0, 0) }. By the Lagrange multiplier technique, we have p i = m { + β Z i } for i =,, m and l m := 2 log L m = 2 5 m log{ + β Z i },

6 where β satisfies m m Z i + β Z i = (0, 0). (5) Write Σ = (σ i,j ) i d, j d = E{(X µ )(X µ ) }, the covariance matrix of X, and use λ λ d to denote the d eigenvalues of the matrix Σ. Similarly, write Σ = ( σ i,j ) i d, j d = E{(Y µ 2 )(Y µ 2 ) } and use λ λ d to denote the d eigenvalues of the matrix Σ. Also write d d ρ = σi,j 2 = tr(σ 2 ), ρ 2 = σ i,j 2 = tr( Σ 2 ), τ = 2α Σα, τ 2 = 2α Σα. (6) i, i, Here tr means trace for a matrix. Note that ρ = E[(X µ ) (X 2 µ )] 2, ρ 2 = E[(Y µ 2 ) (Y 2 µ 2 )] 2, τ = 2E[α (X µ )] 2 and τ 2 = 2E[α (Y µ 2 )] 2, and these quantities may depend on n, n 2 since d may depend on n, n 2. heorem. Assume min{n, n 2 }, τ and τ 2 in (6) are positive, and for some δ > 0, E (X µ ) ( X µ ) 2+δ ρ (2+δ)/2 E (Y µ 2 ) (Ȳ µ 2 ) 2+δ ρ (2+δ)/2 2 E α (X + X 2µ ) 2+δ τ (2+δ)/2 = o(m δ+min(δ,2) 4 ), (7) = o(m δ+min(δ,2) 4 2 ), (8) = o(m δ+min(δ,2) 4 ), (9) and E α (Y + Ȳ 2µ 2 ) 2+δ τ (2+δ)/2 2 = o(m δ+min(δ,2) 4 2 ). (0) hen, under H 0 : µ = µ 2, l m converges in distribution to a chi-square distribution with two degrees of freedom as min{n, n 2 }. Based on the above theorem, one can test H 0 : µ = µ 2 against H a : µ µ 2 by rejecting H 0 when l m χ 2 2,γ, where χ2 2,γ denotes the ( γ) quantile of a chi-square distribution with two degrees of freedom and γ is the significant level. 6

7 Remark. In equations (7) (0), the restrictions are put on E W 2+δ /(EW 2 ) (2+δ)/2 for some random variables W, which are necessary for the CL to hold for random arrays. Later we will see those conditions are satisfied by imposing some conditions on the higher-order moments or special dependence structure. Remark 2. he proposed test has the following merits:. he limiting distribution is always chi-square without estimating the asymptotic covariance. 2. It does not require any specific structure such as the one used in Bai and Saranadasa (996) and Chen and Qin (200), which will be discussed later. 3. With higher-order moment condition or special dependence structure of {X i } and {Y i }, d can be very large. 4. here is no restriction imposed on the relation between n and n 2 except that min{n, n 2 }. hat is, no need to assume a limit or bound on the ratio n /n 2. Moreover, no assumptions are needed on ρ /ρ 2 or τ /τ 2. Hence the covariance matrices Σ and Σ 2 can be arbitrary as long as τ,τ 2 > 0, which are simply equivalent to saying that both α X and α Y are non-degenerate random variables. Next we verify heorem by imposing conditions on the moments and dimension of the random vectors. A: 0 < lim inf λ lim sup n n λ d < and 0 < lim inf n 2 λ lim sup λ d < ; n 2 A2: For some δ > 0, d d E{ X,i µ,i 2+δ + Y,i µ 2,i 2+δ } = O(); A3: d = o(m δ+min(δ,2) 2(2+δ) ). Corollary. Assume min{n, n 2 } and conditions A A3 hold. hen, under H 0 : µ = µ 2, conditions (7) (0) are satisfied, i.e., heorem holds. 7

8 Condition A3 is a somewhat restrictive condition for the dimension d. Note that conditions A and A2 are related only to the covariance matrices and some higher moments on the components of the random vectors. he higher moments we have, the less restriction is imposed on d. Condition A3 can be removed for models with some special dependence structures. For comparisons, we prove the Wilks theorem for the proposed jackknife empirical likelihood test under the following model B considered by Bai and Saranadasa (996), Chen, Peng and Qin (2009) and Chen and Qin (200): B (Factor model). X i = Γ B i +µ for i =,, n, Y j = Γ 2 Bj +µ 2 for j =,, n 2, where Γ, Γ 2 are d k matrices with Γ Γ = Σ, Γ 2Γ 2 = Σ, {B i = (B i,,, B i,k ) } n and { B j = ( B j,,, B j,k ) } n 2 are two independent random samples satisfying that EB i = E B i = 0, Var(B i ) = Var( B i ) = I k k, EB 4 i,j = 3 + ξ <, E B 4 i,j = 3 + ξ 2 <, E k l= Bν l i,l = k l= EBν l i,l and E k l= B ν l i,l = k l= E B ν l i,l whenever ν + + ν k = 4 for distinct nonnegative integers ν l s. heorem 2. Assume min{n, n 2 }, and τ and τ 2 in (6) are positive. Under model B and H 0 : µ = µ 2, l m converges in distribution to a chi-square distribution with two degrees of freedom as min{n, n 2 }. Remark 3. It can be seen from the proof of heorem 2 that assumptions EB 4 i,j = 3 + ξ < and E B 4 i,j = 3+ξ 2 < in model B can be replaced by the much weaker conditions max j k EB 4,j = o(m) and max j k E B 4,j = o(m). Unlike Bai and Saranadasa (996) and Chen and Qin (200), there is no restriction on d and k for our proposed method. he only constraint imposed on matrices Γ and Γ 2 is that both α Σα and α Σα are positive, which is very weak. 3 Simulation study and data analysis 3. Simulation study We investigate the finite sample behavior of the proposed jackknife empirical likelihood test (JEL) and compare it with the test statistic in (2) proposed by Chen and Qin (200) in terms of both size 8

9 and power. Let W,, W d be iid random variables with the standard normal distribution function, and let W,, W d, independent of W i s be iid random variables with distribution function t(8). Put X, = W, X,2 = W +W 2,, X,d = W d +W d, Y, = W +µ 2,, Y,2 = W + W 2 +µ 2,2,, Y,d = W d + W d + µ 2,d, where µ 2,i = c if i [c 2 d], and µ 2,i = 0 if i > [c 2 d]. hat is, 00c 2 % of the components of Y have a shifted mean compared to that of X. Since we test H 0 : EX = EY against H a : EX EY, the case of c = 0 denotes the size of tests. By drawing, 000 random samples of sizes n = 30, 00, 50 from X = (X,,, X,d ) and independently drawing, 000 random samples of sizes n 2 = 30, 00, 200 from Y = (Y,,, Y,d ) with d = 0, 20,, 00, 300, 500, c = 0, 0. and c 2 = 0.25, 0.75, we calculate the powers of the two tests mentioned above. In ables 3, we report the empirical sizes and powers for the proposed jackknife empirical likelihood test with α = (,, ) R d and the test in Chen and Qin (200) at level 5%. Results for level 0% are similar. From these three tables, we observe that (i) the sizes of both tests, i.e., results for c = 0 are comparable and quite stable with respect to the dimension d; (ii) the proposed jackknife empirical likelihood test is more powerful than the test in Chen and Qin (200) for the case of c 2 = 0.75 and the case when the data is sparse, but d is large (i.e., the case of c = 0., c 2 = 0.25). Since equation (4) has nothing to do with sparsity, it is expected that the proposed jackknife empirical likelihood method is not powerful when the data is sparse. Hence, it would be of interest to connect sparsity with some estimating equations so as to improve the power of the proposed jackknife empirical likelihood test. In conclusion, the proposed jackknife empirical likelihood test has a very stable size with respect to the dimension and is powerful as well. Moreover, the new test is easy to compute, flexible to take other information into account, and works for both fixed dimension and divergent dimension. In comparison with the test in Chen and Qin (200), the new test has a comparable size, but is more powerful when 9

10 the data is not very sparse. Some further research on formulating sparsity into estimating equations will be pursued in future. 3.2 Data analysis In this subsection we employ the proposed method to the Colon data with 2000 gene expression levels on 22 (n ) normal colon tissues and 40 (n 2 ) tumor colon tissues. his data set is available from the R package plsgenomics, and has been anlayzed by Alon et al. (999) and Srivastava and Du (2009). he p-values of three tests proposed by Srivastava and Du (2009) equal to.38e-06, 4.48e- and , which clearly reject the null hypothesis that the tumor group has the same gene expression levels as the normal group. A direct application of the proposed jackknife empirical likelihood method and the CQ test for testing the equality of means gives p-values.36e-0 and 5.06e-09, respectively, which clearly result in a contradiction. Although the test in Chen and Qin (200) and the three tests in Srivastava and Du (2009) clearly reject the null hypothesis, the p-values are significantly different. A closer look at the difference of sample means (see Figure ) shows that some genes have a significant difference of sample means and a high variability, which may play an important role in the CQ test and the equation (4) with α = (,, ) R d of the proposed jackknife empirical likelihood method. o examine this effect, we apply the methods to those genes without the significant difference in sample means and the logarithms of the 2000 gene expression levels. First we apply the proposed jackknife empirical likelihood method and the CQ test to those genes satisfying n n X i,j n 2 n2 Y i,j c 3 for some given threshold c 3 > 0. In able 4, we report the p-values for these two methods for different c 3, which confirm that some genes play an important role in rejecting the equality of means in the CQ test. Figure clearly shows that some genes have a large positive mean and some genes have a large negative mean, and the equation (4) with the simple a = (,,, ) can not catch this characteristic. 0

11 Here, we propose to replace (4) by m m 2 (X i Y j ){I( m m X i Ȳj > 0) I( X i Ȳj < 0)}, () 2 which results in the P-value 2.63e-03, and so we reject the null hypothesis that the tumor group has the same gene expression levels as the normal group. Although the derived theorems based on (3) and (4) can not be applied to (3) and () directly, results hold under some similar conditions by noting the boundedness of I( X i Ȳj > 0) I( X i Ȳj < 0). It is well known that gene expression data are quite noisy and some transformation and normalization are needed before doing statistical analysis; see Chapter 6 of Lee (2004). Here we apply the CQ test and the proposed empirical likelihood methods based on (3) and (4), and (3) and () to testing the equality of means of the logarithms of the 2000 gene expression levels on normal colon tissues and tumor colon tissues, which give p-values 0.84, and 0.48, respectively. hat is, one can not reject the equality of means of the logarithms of the 2000 gene expression levels. We plot the differences of sample means of the logarithms in Figure 2, which are much less volatile than the differences of sample means in Figure. In summary, carefully choosing α in the empirical likelihood method is needed when it is applied to the colon data, which has a small sample size and a large variation. However simply choosing α = (,, ) in the empirical likelihood method gives a similar result as the test in Chen and Qin (200) for testing the equal means of the logarithms of Colon data. 4 Proofs In the proofs we use to denote the L 2 norm of a vector or matrix. Since µ µ 2 is our target and under the null hypothesis µ µ 2 = 0, without loss of generality we assume µ = µ 2 = 0. Write u ij = (X i Y j ) ( X i Ȳj) and v ij = α (X i Y j ) + α ( X i Ȳj) for i m, j m 2. hen it

12 is easily verified that for i, k m, j, l m 2, E(u ij ) = E(v kl ) = E(u ij v kl ) = 0, and d Var(u kl ) = (σi,j 2 + σ i,j) 2 = ρ + ρ 2, i, Var(v kl ) = 2α (Σ + Σ)α = τ + τ 2. Lemma. Under conditions of heorem, we have as min{n, n 2 } m m X i d N(0, ), (2) ρ X i and Proof. Since Var(X i m 2 m2 m m m 2 m2 Yj Ȳ j ρ2 d N(0, ), (3) α (X i + X i ) τ d N(0, ), (4) α (Y j + Ȳj) τ2 d N(0, ). (5) X i ) = ρ and X X,, X m Xm are i.i.d. for fixed m, equation (2) follows from (7) and the Lyapunov central limit theorem. he rest can be shown in the same way. From now on we denote ρ = m m ρ + m m 2 ρ 2 and τ = m m τ + m m 2 τ 2. Lemma 2. Under conditions of heorem, we have as min{n, n 2 } m m m m 2 ρ X i m 2 Ȳ j p 0, (6) m α p X i 0, (7) m τ 2

13 and m m 2 m m2 α p Y j 0, (8) m 2 τ m m m 2 m 2 m m 2 m 2 m m 2 X i (X i X i ) 2 ρ p, (9) (Yj Ȳ j ) 2 p, (20) ρ 2 [α (X i + X i )] 2 τ p, (2) [α (Y j + Ȳj)] 2 τ 2 p, (22) X i [α (X i + X i )] ρ τ p 0, (23) Yj Ȳ j [α (Y j + Ȳj)] p 0. (24) ρ2 τ 2 Proof. Note that µ = µ 2 = 0 is assumed in Section 4. hen (6) follows from the fact that m m Var( m m 2 ρ X i m 2 2 Y j ) = E m m m 2 m 2 m2 2 ρ Xi Y j = E m m m 2 m 2 m2 2 ρ (Xi Y j ) 2 [ ] m = E m m 2 ρ (X Y ) 2 = m m m 2 ρ d σ ij σ ij i, m ρ + ρ 2 m + m 2 2ρ 2 ( m + m 2 ) = o(). In the same way, we can show (7) and (8). 3

14 o show (9), write u i = X i X i. We need to estimate E m u2 i m ρ (2+δ)/2. Note that ρ = Eu 2. When 0 < δ 2, it follows from von Bahr and Esseen (965) that m E u 2 i m ρ (2+δ)/2 2m E u 2 E(u 2 ) (2+δ)/2 = O(m E u 2+δ ). (25) When δ > 2, it follows from Dharmadhikari and Jogdeo (969) that m E u 2 i m ρ (2+δ)/2 = O(m (2+δ)/4 E u 2 E(u 2 ) (2+δ)/2 ) = O(m (2+δ)/4 E u 2+δ ). (26) herefore, by (25), (26) and (7) we have for any ε > 0 m P( u2 i > ε) m ρ ε (2+δ)/2 E m u2 i mρ (2+δ)/2 (m ρ ) (2+δ)/2 = O(m (δ+min(δ,2))/4 E u ρ 2+δ ) = o(), which implies (9). he rest can be shown in the same way. Lemma 3. Under conditions of heorem, we have as min{n, n 2 } m m 2 m u ij ρ m m 2 m m m 2 m 2 m2 2 ρ m2 m m 2 m2 2 ρ k= ( m k= m m m 2 m 2 m2 2 τ m2 m m 2 m2 2 τ k= ( m k= v ij τ u kj d N(0, I 2 ), (27) 2 mρ m ρ u ik ) 2 mρ 2 m 2 ρ v kj 2 mτ m τ v ik ) 2 mτ 2 m τ p 0, (28) p 0, (29) p 0, (30) p 0, (3) 4

15 m 2 m2 2 m 2 m2 2 where I 2 is the 2 2 identity matrix. m m m 2 ρτ k= m m 2 m ρτ m 2 u ki v kj m u ik v jk k= p 0, (32) p 0, (33) Proof. It follows from Lemma 2 that m m m 2 m 2 m u ij ρ = = = m2 m m m 2 ρ m m X m ρ i mρ m ρ m m m (Xi X i + X i = a m A m + b m B m + o p (), X i + Yj Ȳ j Xi m2 m m 2 ρ X i mρ2 + ρ m2 ρ Yj Ȳ j m2 m 2 Ȳ j Y j X i ) m2 m m m 2 ρ Yj Ȳ j m (Xi ρ2 + o p () Ȳ j + Y j X i ) where a m = mρ m ρ, b m = mρ2 m2 ρ, A m = m m X i X i ρ d N(0, ) and Bm = m2 Yj Ȳj m2 ρ2 d N(0, ). Obviously a 2 m + b 2 m = and A m, B m are independent. Denote the characteristic functions of A m and B m by Φ m and Ψ m, respectively. hen, E exp(it(a m A m + b m B m )) = E exp(ita m A m )E exp(itb m B m ) = Φ m (ta m )Ψ m (tb m ) = [exp( (ta m) 2 ) + o()][exp( (tb m) 2 ) + o()] 2 2 = exp( t2 2 ) + o(), i.e., Similarly, we have m m m 2 m 2 m u ij ρ d N(0, ). (34) m m m 2 m 2 m v ij τ = mτ m τ m m α (X i + X i ) τ mτ2 m2 τ m 2 m2 α (Y j + Ȳj) τ2 d N(0, ). 5

16 Let a and b be two real numbers with a 2 + b 2 0. Note that m m m 2 m 2 m (a u ij + b v ij ) ρ τ mρ = a m ρ = +b mτ m τ ( a mρ m ρ + a mρ 2 m2 ρ m m m m m m = I + I 2 + o p (). Xi ρ X i m 2 m2 X i mρ2 m2 ρ m2 m 2 α (X i + X i ) mτ2 + τ m2 τ X i + b mτ ρ m τ Yj Ȳ j b mτ 2 ρ2 m2 τ m m m 2 m2 Yj Ȳ j ρ2 m 2 m2 α (X i + X ) i ) τ α (Y j + Ȳj) + o p () τ2 α (Y j + Ȳj) + o p () τ2 Since mρ m ρ, mρ2 m2 ρ, mτ m τ, mτ2 m2 τ are all bounded by one, it is easy to check that I and I 2 satisfy the Lyapunov condition by (7) - (0). herefore and Since X i {a 2 mρ m ρ + b2 mτ m τ } /2 I d N(0, ) {a 2 mρ 2 m 2 ρ + b2 mτ 2 m 2 τ } /2 I 2 d N(0, ). s are independent of Y i s, it follows from the same arguments in proving (34) that I + I 2 d N(0, a 2 + b 2 ), i.e., (27) holds. o prove (28), we write m m m 2 m2 2 ρ k= = m m 2 m2 2 ρ m = m m 2 ρ m k= k= ( m2 ) 2 u kj ( m2 ( X k (X X k k + Yj X k + m 2 m2 Y j Ȳ j Y j ) 2 X k Xk Ȳj) Ȳ j m2 m 2 Y j X k X k (35) ) 2 m2 m 2 Ȳj. 6

17 Since mρ /m ρ, it follows from Lemma 2 that m m m 2 ρ (Xk X k ) 2 mρ m ρ k= p 0. (36) By Lemma, we have m m m 2 ρ m 2 m 2 k= Y j Ȳ j 2 = O p ( mρ 2 m m 2 ρ ) = o p(). (37) A direct calculation shows that which implies that Similarly we have It follows from (36) and (38) that E{ m2 m 2 Y j X k } 2 = E{( m 2 m2 Y j ) X k X k ( m 2 m2 Y j)} = Etr{ m2 m 2 Y j X k X k ( m 2 m2 Y j)} = Etr{ X k X k ( m 2 m2 Y j)( m 2 m2 Y i )} = tre{ X k X k ( m 2 m2 Y j)( m 2 m2 Y i )} = tr{σ m 2 Σ} = O( ρ +ρ 2 m 2 ) = O( m ρ m 2 m ) + O( ρ 2 m 2 ) = o( m ρ m ), m m m 2 ρ { m 2 m 2 k= m m m 2 ρ Y j {X m 2 k m 2 k= X k } 2 = o p (). (38) Ȳ j } 2 = o p (). (39) m m m 2 ρ k= (X X k k )( m2 m 2 Y j X k ) { m m m 2 ρ k= (X X k k ) 2 } /2 { m m m 2 ρ k= ( m2 m 2 Y j X k ) 2 } /2 (40) = O p ()o p () = o p (). 7

18 Similarly we can show that m m m 2 ρ k= (X X k k )( m2 m 2 Y j Ȳ j ) = o p () m m 2 ρ m k= (X k X k )(X k m 2 m2 Ȳ j ) = o p() m m 2 ρ m k= ( m 2 m2 Y j Ȳ j )( m2 m 2 Y i X k ) = o p () (4) m m 2 ρ m k= ( m 2 m2 Y j m m m 2 ρ k= ( m2 m 2 Y j Ȳ j )(X k X k )(X k m 2 m2 Ȳi) = o p () m 2 m2 Y i) = o p (). Hence (28) follows from (35) (4). he rest can be shown in the same way as proving (28). Lemma 4. Under conditions of heorem, we have as min{n, n 2 } m m k= Z k, ρ Z k,2 τ d N(0, I 2 ), (42) Moreover, we have mρ mτ m ρτ m Zk, 2 p 0, (43) k= m Zk,2 2 p 0, (44) k= m p Z k, Z k,2 0. (45) k= max Z k, = o p (m /2 ) and k m ρ max Z k,2 = o p (m /2 ). (46) k m τ Proof. Note that for k m, and for m + k m, Z k, = Z k,2 = (m )m (m )m m 2 m m 2 m u ij + m + m 2 (m )m 2 v ij + m + m 2 (m )m 2 m 2 m 2 u kj, v kj, Z k, = (m 2 )m 2 m 2 m u ij + m + m 2 (m 2 )m m u i,k m, 8

19 hus m 2 m Z k,2 = (m 2 )m 2 u ij + m + m 2 (m 2 )m m v i,k m. m m k= Z k, ρ = = ( m m 2 + m + m + m 2 + m + m 2 m 2 ) (m )m 2 (m 2 )m m m m 2 m 2 m u ij ρ m u ij ρ and which imply (42) by using Lemma 3. It follows from Lemma 3 that mρ = mρ + mρ m k= m k= Z 2 k, m m k= m 2 m (m )m m 2 m 2 m (m 2 )m 2 k= m 2 m = { (m ) u ij } 2 + m mρ Z k,2 τ = m m m 2 u ij + m + m 2 (m )m 2 m 2 m u ij + m + m 2 (m 2 )m v ij τ, m 2 u kj m u ik (m ) 2 m (m ) 2 m 2 2 mρ k= 2 2 m 2 ( u kj ) 2 m m 2 m 2{( mρ(m ) 2 ) /2 u ij } 2 m 2 m + { m m 2 (m 2 ) u ij } 2 m 2 mρ m 2 m (m ) 2 m 2 m + mρ(m 2 ) 2 m 2 ( u ik ) 2 m 2{( mρ(m 2 ) 2 ) /2 u ij } 2 m 2 m = {O p ( m m m +{O p ( m 2 m2 m k= = mρ m ρ + mρ 2 m 2 ρ + o p() = + o p (), m m 2 m )} 2 + (m )2 m 2 (m ) 2 m 2 {mρ m m 2 m )} 2 + (m )2 m 2 2 m 2 (m 2 ) 2 {mρ 2 m ρ + o m m p()} + {O p ( 2 )} 2 m m m 2 m m 2 ρ + o m m p()} + {O p ( 2 )} 2 m 2 m2 m m i.e., (43) holds. Similarly we can show (44) and (45). 9

20 Since E(( m u ij) 2 ) = Var( m u ij) = m (ρ + ρ 2 ), we have which implies that Similarly we have m m 2 m E( max ( u ij ) 2 ) E(( u ij ) 2 ) = m m 2 (ρ + ρ 2 ) j m 2 m max u ij = O p ( m 2 m (ρ + ρ 2 )). j m 2 m 2 max u ij = O p ( m 2 m (ρ + ρ 2 )). m Hence by Lemma 3 and the expression for Z k,, we have max k m Z k, ρ (m )m m m2 u ij ρ + max k m Similarly we can show that + (m 2 )m 2 m m2 u ij ρ + max k m2 = o p () + O p ( m (m )m 2 ρ {m m 2 (ρ + ρ 2 )} /2 ) +o p () + O p ( m (m 2 )m ρ {m m 2 (ρ + ρ 2 )} /2 ) = o p () + O p ( m /2 (min(m,m 2 )) /2 ) = o p (m /2 ). max Z k,2 = o p (m /2 ). k m τ m m2 (m )m 2 m m (m 2 )m u kj ρ u jk ρ Proof of heorem. It follows from Lemma 4 and the standard arguments in empirical likelihood method (see Owen (990)). o show Corollary and heorem 2, we first prove the following lemmas. Lemma 5. tr(σ 4 ) = O ( (tr(σ 2 )) 2), ρ = d λ2 j, and 2 α 2 λ τ 2 α 2 λ d. Proof. Since tr(σ j ) = d λj i for any positive integer j, the first equality follows immediately. he second equality follows since ρ = tr(σ 2 ). he third pair of inequalities on τ come from the definition of τ. 20

21 Lemma 6. For any δ > 0 E X X 2+δ d δ ( d E X,i 2+δ ) 2 and E α (X + X d ) 2+δ 2 4+δ α 2+δ d δ/2 E X,i 2+δ. Proof. It follows from the Cauchy-Schwarz inequality that X X 2 X 2 X 2. hen by using the C r inequality we conclude that ( d (2+δ)/2 ( d E X X 2+δ E X,i) 2 E ( d ) (2+δ)/2 2 = E ( d δ/2 X 2,i ) 2 d E X,i 2+δ X 2,i ) (2+δ)/2 = d δ ( d E X,i 2+δ ) 2. Similarly, from the C r inequality we have E α (X + X ) 2+δ 2 4+δ E α X 2+δ 2 4+δ α 2+δ E( X 2+δ ) = 2 4+δ α 2+δ d +δ/2 E X,i 2 d 2 4+δ α 2+δ d δ/2 E X,i 2+δ. his completes the proof. Proof of Corollary. Equations (7) and (9) follow from conditions A A3 by using Lemmas 5 and 6. So do equations (8) and (0), since we have the same assumptions on {X i } and {Y j }. 2

22 Proof of heorem 2. If suffices to verify conditions (7) and (9) with δ = 2 in heorem. Recall we assume that µ = µ 2 = 0. Note that Var(X ) = Σ = Γ Γ. Denote α Γ = (a,, a k ) and Σ = Γ Γ = (σ j,l ) j,l k. hen and X X = k α (X + X ) = l= k σ j,l B,jB +m,l, k a j (B,j + B +m,j). Set δ j,j 2,j 3,j 4 = E(B,j B,j2 B,j3 B,j4 ). hen δ j,j 2,j 3,j 4 equals 3 + ξ if j = j 2 = j 3 = j 4, equals if j, j 2, j 3 and j 4 form two different pairs of integers, and is zero otherwise. By Lemma 5, we have E(X X ) 4 = = O = O i.e., (7) holds with δ = 2. k k j,j 2,j 3,j 4 = l,l 2,l 3,l 4 = σ j,l σ j 2,l 2 σ j 3,l 3 σ j 4,l 4 δ j,j 2,j 3,j 4 δ l,l 2,l 3,l 4 ( j j2 l l 2 σ j,l σ j,l 2 σ j 2,l σ j 2,l 2 ( k +O l l 2 σ 2 j,l σ 2 j,l 2 ) + O ( k ) ( + O k j j 2 l= σ 2 ) k l= σ 4 j,l ( k k k ) k j = j 2 = l = l 2 = σ j,l σ j,l 2 σ j 2,l σ j 2,l 2 ( k +O k ) ( k k j = j 2 = l= σ 2 j,l σ 2 j 2,l + O k l= σ 4 j,l = O ( tr(σ 4 ) ) ( k ) + O k l= σ 2 j,l )2 = O ( tr(σ 4 ) ) + O ( (tr(σ 2 )) 2) = O ( tr(σ 4 ) ) + O ( (tr(σ 2 )) 2) = O(ρ 2 ), ) j,l σ 2 j 2,l ) 22

23 Similarly we have E(α (X + X ( k ) 4 )) E a j B,j = O k j,j 2 = = O ( k k a 2 j a 2 j 2 + O a 2 j ) 2 ( ( ) ) 2 = O α Γ Γ α = O ( τ 2 ), a 4 j which yields (9) with δ = 2. Equations (8) and (0) can be shown in the same way. Hence heorem 2 follows from heorem. Acknowledgment. We thank an associate editor and two reviewers for helpful comments. Wang s research was supported by the Bob Price fellowship in the School of Mathematics at Georgia ech. Peng s research was supported by NSF Grant DMS and Qi s research was supported by NSF grant DMS References [] Alon, U., Barkai, N., Notterman, D.A., Gish, K., Ybarra, S. Mack, D. and Levin A.J. (999). Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. Proc. Natl. Acad. Sci. U.S.A. 96, [2] Arlot, S., Blanchard, G. and Roquain, E. (200a). Some nonasymptotic results on resampling in high dimension I: confidence regions. Ann. Statist. 38, [3] Arlot, S., Blanchard, G. and Roquain, E. (200b). Some nonasymptotic results on resampling in high dimension II: multiple tests. Ann. Statist. 38,

24 [4] Bai, Z. and Saranadasa, H. (996). Effect of high dimension: by an example of a two sample problem. Statist. Sinica 6, [5] Chen, S.X., Peng, L. and Qin, Y. (2009). Empirical likelihood methods for high dimension. Biometrika 96, [6] Chen, S. and Qin, Y. (200). A two-sample test for high-dimensional data with applications to gene-set testing. Ann. Statist. 38, [7] Chen, S. and Van Keilegom, I. (2009). A review on empirical likelihood methods for regression. est 8, [8] Dharmadhikari, S.W. and Jogdeo, K. (969). Bounds on moments of certain random variables. Ann. Math. Statist. 40, [9] Jing, B.Y., Yuan, J.Q. and Zhou, W. (2009). Jackknife empirical likelihood. J. Amer. Statist. Assoc. 04, [0] Lee, M.L. (2004). Analysis of Microarray Gene Expression Data. Kluwer. [] Kuelbs, J. and Vidyashankar, A.N. (200). Asymptotic inference for high dimensional data. Ann. Statist. 38, [2] Owen, A. (990). Empirical likelihood ratio confidence regions. Ann. Statist. 8, [3] Owen, A. (200). Empirical Likelihood. Chapman & Hall/CRC. [4] Qin, J. and Lawless, J. (994). Empirical likelihood and general estimating equations. Ann. Statist. 22, [5] Srivastava, M.S. (2009). A test for the mean vector with fewer observations than the dimension under non-normality. J. Multi. Anan. 00,

25 [6] Srivastava, M.S. and Du, M. (2008). A test for the mean vector with fewer observations than the dimension. J. Multi. Anan. 99, [7] von Bahr, B. and Esseen, C.G. (965). Inequality for the rth absolute moment of a sum of random variables, r 2. Ann. Math. Statist. 36, difference of sample means genes Figure : Colon data: differences of the sample means are plotted against each gene. 25

26 able : Sizes and powers of the proposed jackknife empirical likelihood test (JEL) and the test in Chen and Qin (200) (CQ) are reported for the case of (n, n 2 ) = (30, 30) at level 5%. d JEL CQ JEL CQ JEL CQ c = 0 c = 0 c = 0. c = 0. c = 0. c = 0. c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.75 c 2 =

27 able 2: Sizes and powers of the proposed jackknife empirical likelihood test (JEL) and the test in Chen and Qin (200) (CQ) are reported for the case of (n, n 2 ) = (00, 00) at level 5%. d JEL CQ JEL CQ JEL CQ c = 0 c = 0 c = 0. c = 0. c = 0. c = 0. c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.75 c 2 =

28 able 3: Sizes and powers of the proposed jackknife empirical likelihood test (JEL) and the test in Chen and Qin (200) (CQ) are reported for the case of (n, n 2 ) = (50, 200) at level 5%. d JEL CQ JEL CQ JEL CQ c = 0 c = 0 c = 0. c = 0. c = 0. c = 0. c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.25 c 2 = 0.75 c 2 =

29 genes difference of sample means of log Figure 2: Colon data: differences of the sample means of logarithms of gene expression levels are plotted against each gene. 29

30 able 4: Colon data: p-values for testing equal means of those genes with the absolute difference of sample means less than the threshold c 3. c 3 number of genes CQ JEL e-0 2.3e e e e e e e e e e-09.36e-0 30

Empirical Likelihood Tests for High-dimensional Data

Empirical Likelihood Tests for High-dimensional Data Empirical Likelihood Tests for High-dimensional Data Department of Statistics and Actuarial Science University of Waterloo, Canada ICSA - Canada Chapter 2013 Symposium Toronto, August 2-3, 2013 Based on

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

Effects of data dimension on empirical likelihood

Effects of data dimension on empirical likelihood Biometrika Advance Access published August 8, 9 Biometrika (9), pp. C 9 Biometrika Trust Printed in Great Britain doi:.9/biomet/asp7 Effects of data dimension on empirical likelihood BY SONG XI CHEN Department

More information

High-dimensional two-sample tests under strongly spiked eigenvalue models

High-dimensional two-sample tests under strongly spiked eigenvalue models 1 High-dimensional two-sample tests under strongly spiked eigenvalue models Makoto Aoshima and Kazuyoshi Yata University of Tsukuba Abstract: We consider a new two-sample test for high-dimensional data

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

On testing the equality of mean vectors in high dimension

On testing the equality of mean vectors in high dimension ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.

More information

On corrections of classical multivariate tests for high-dimensional data

On corrections of classical multivariate tests for high-dimensional data On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics

More information

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0}

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0} VARIATION OF ITERATED BROWNIAN MOTION Krzysztof Burdzy University of Washington 1. Introduction and main results. Suppose that X 1, X 2 and Y are independent standard Brownian motions starting from 0 and

More information

Strong Convergence of the Empirical Distribution of Eigenvalues of Large Dimensional Random Matrices

Strong Convergence of the Empirical Distribution of Eigenvalues of Large Dimensional Random Matrices Strong Convergence of the Empirical Distribution of Eigenvalues of Large Dimensional Random Matrices by Jack W. Silverstein* Department of Mathematics Box 8205 North Carolina State University Raleigh,

More information

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions

Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions Central Limit Theorems for Classical Likelihood Ratio Tests for High-Dimensional Normal Distributions Tiefeng Jiang 1 and Fan Yang 1, University of Minnesota Abstract For random samples of size n obtained

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

Self-normalized Cramér-Type Large Deviations for Independent Random Variables

Self-normalized Cramér-Type Large Deviations for Independent Random Variables Self-normalized Cramér-Type Large Deviations for Independent Random Variables Qi-Man Shao National University of Singapore and University of Oregon qmshao@darkwing.uoregon.edu 1. Introduction Let X, X

More information

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department

More information

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Statistica Sinica 19 (2009), 71-81 SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Song Xi Chen 1,2 and Chiu Min Wong 3 1 Iowa State University, 2 Peking University and

More information

Package ShrinkCovMat

Package ShrinkCovMat Type Package Package ShrinkCovMat Title Shrinkage Covariance Matrix Estimators Version 1.2.0 Author July 11, 2017 Maintainer Provides nonparametric Steinian shrinkage estimators

More information

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS Communications in Statistics - Simulation and Computation 33 (2004) 431-446 COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS K. Krishnamoorthy and Yong Lu Department

More information

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR Introduction a two sample problem Marčenko-Pastur distributions and one-sample problems Random Fisher matrices and two-sample problems Testing cova On corrections of classical multivariate tests for high-dimensional

More information

Testing Block-Diagonal Covariance Structure for High-Dimensional Data

Testing Block-Diagonal Covariance Structure for High-Dimensional Data Testing Block-Diagonal Covariance Structure for High-Dimensional Data MASASHI HYODO 1, NOBUMICHI SHUTOH 2, TAKAHIRO NISHIYAMA 3, AND TATJANA PAVLENKO 4 1 Department of Mathematical Sciences, Graduate School

More information

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS

KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Bull. Korean Math. Soc. 5 (24), No. 3, pp. 7 76 http://dx.doi.org/34/bkms.24.5.3.7 KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Yicheng Hong and Sungchul Lee Abstract. The limiting

More information

Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek

Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek J. Korean Math. Soc. 41 (2004), No. 5, pp. 883 894 CONVERGENCE OF WEIGHTED SUMS FOR DEPENDENT RANDOM VARIABLES Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek Abstract. We discuss in this paper the strong

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Advanced Statistics II: Non Parametric Tests

Advanced Statistics II: Non Parametric Tests Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney

More information

Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou

Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics. Wen-Xin Zhou Cramér-Type Moderate Deviation Theorems for Two-Sample Studentized (Self-normalized) U-Statistics Wen-Xin Zhou Department of Mathematics and Statistics University of Melbourne Joint work with Prof. Qi-Man

More information

Introduction to Self-normalized Limit Theory

Introduction to Self-normalized Limit Theory Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized

More information

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA

More information

On the Testing and Estimation of High- Dimensional Covariance Matrices

On the Testing and Estimation of High- Dimensional Covariance Matrices Clemson University TigerPrints All Dissertations Dissertations 1-009 On the Testing and Estimation of High- Dimensional Covariance Matrices Thomas Fisher Clemson University, thomas.j.fisher@gmail.com Follow

More information

Bootstrapping high dimensional vector: interplay between dependence and dimensionality

Bootstrapping high dimensional vector: interplay between dependence and dimensionality Bootstrapping high dimensional vector: interplay between dependence and dimensionality Xianyang Zhang Joint work with Guang Cheng University of Missouri-Columbia LDHD: Transition Workshop, 2014 Xianyang

More information

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Takahiro Nishiyama a,, Masashi Hyodo b, Takashi Seo a, Tatjana Pavlenko c a Department of Mathematical

More information

University of California, Berkeley

University of California, Berkeley University of California, Berkeley U.C. Berkeley Division of Biostatistics Working Paper Series Year 24 Paper 153 A Note on Empirical Likelihood Inference of Residual Life Regression Ying Qing Chen Yichuan

More information

Analysis of variance, multivariate (MANOVA)

Analysis of variance, multivariate (MANOVA) Analysis of variance, multivariate (MANOVA) Abstract: A designed experiment is set up in which the system studied is under the control of an investigator. The individuals, the treatments, the variables

More information

Grouped Network Vector Autoregression

Grouped Network Vector Autoregression Statistica Sinica: Supplement Grouped Networ Vector Autoregression Xuening Zhu 1 and Rui Pan 2 1 Fudan University, 2 Central University of Finance and Economics Supplementary Material We present here the

More information

Statistica Sinica Preprint No: SS R2

Statistica Sinica Preprint No: SS R2 Statistica Sinica Preprint No: SS-2016-0063R2 Title Two-sample tests for high-dimension, strongly spiked eigenvalue models Manuscript ID SS-2016-0063R2 URL http://www.stat.sinica.edu.tw/statistica/ DOI

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

CHAPTER 10 Shape Preserving Properties of B-splines

CHAPTER 10 Shape Preserving Properties of B-splines CHAPTER 10 Shape Preserving Properties of B-splines In earlier chapters we have seen a number of examples of the close relationship between a spline function and its B-spline coefficients This is especially

More information

Large-Scale Multiple Testing of Correlations

Large-Scale Multiple Testing of Correlations Large-Scale Multiple Testing of Correlations T. Tony Cai and Weidong Liu Abstract Multiple testing of correlations arises in many applications including gene coexpression network analysis and brain connectivity

More information

Estimation of multivariate 3rd moment for high-dimensional data and its application for testing multivariate normality

Estimation of multivariate 3rd moment for high-dimensional data and its application for testing multivariate normality Estimation of multivariate rd moment for high-dimensional data and its application for testing multivariate normality Takayuki Yamada, Tetsuto Himeno Institute for Comprehensive Education, Center of General

More information

Bounds for pairs in partitions of graphs

Bounds for pairs in partitions of graphs Bounds for pairs in partitions of graphs Jie Ma Xingxing Yu School of Mathematics Georgia Institute of Technology Atlanta, GA 30332-0160, USA Abstract In this paper we study the following problem of Bollobás

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

High-dimensional covariance estimation based on Gaussian graphical models

High-dimensional covariance estimation based on Gaussian graphical models High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

High-dimensional asymptotic expansions for the distributions of canonical correlations

High-dimensional asymptotic expansions for the distributions of canonical correlations Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic

More information

Cross-Validation with Confidence

Cross-Validation with Confidence Cross-Validation with Confidence Jing Lei Department of Statistics, Carnegie Mellon University UMN Statistics Seminar, Mar 30, 2017 Overview Parameter est. Model selection Point est. MLE, M-est.,... Cross-validation

More information

Vast Volatility Matrix Estimation for High Frequency Data

Vast Volatility Matrix Estimation for High Frequency Data Vast Volatility Matrix Estimation for High Frequency Data Yazhen Wang National Science Foundation Yale Workshop, May 14-17, 2009 Disclaimer: My opinion, not the views of NSF Y. Wang (at NSF) 1 / 36 Outline

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

Chapter 7, continued: MANOVA

Chapter 7, continued: MANOVA Chapter 7, continued: MANOVA The Multivariate Analysis of Variance (MANOVA) technique extends Hotelling T 2 test that compares two mean vectors to the setting in which there are m 2 groups. We wish to

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS

A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Statistica Sinica 26 (2016), 1117-1128 doi:http://dx.doi.org/10.5705/ss.202015.0240 A CENTRAL LIMIT THEOREM FOR NESTED OR SLICED LATIN HYPERCUBE DESIGNS Xu He and Peter Z. G. Qian Chinese Academy of Sciences

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Modified Simes Critical Values Under Positive Dependence

Modified Simes Critical Values Under Positive Dependence Modified Simes Critical Values Under Positive Dependence Gengqian Cai, Sanat K. Sarkar Clinical Pharmacology Statistics & Programming, BDS, GlaxoSmithKline Statistics Department, Temple University, Philadelphia

More information

On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables

On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables Deli Li 1, Yongcheng Qi, and Andrew Rosalsky 3 1 Department of Mathematical Sciences, Lakehead University,

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests

More information

Effective Linear Discriminant Analysis for High Dimensional, Low Sample Size Data

Effective Linear Discriminant Analysis for High Dimensional, Low Sample Size Data Effective Linear Discriant Analysis for High Dimensional, Low Sample Size Data Zhihua Qiao, Lan Zhou and Jianhua Z. Huang Abstract In the so-called high dimensional, low sample size (HDLSS) settings, LDA

More information

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data

Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations of High-Dimension, Low-Sample-Size Data Sri Lankan Journal of Applied Statistics (Special Issue) Modern Statistical Methodologies in the Cutting Edge of Science Asymptotic Distribution of the Largest Eigenvalue via Geometric Representations

More information

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University

More information

Elliptically Contoured Distributions

Elliptically Contoured Distributions Elliptically Contoured Distributions Recall: if X N p µ, Σ), then { 1 f X x) = exp 1 } det πσ x µ) Σ 1 x µ) So f X x) depends on x only through x µ) Σ 1 x µ), and is therefore constant on the ellipsoidal

More information

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

Copyright 2010 Pearson Education, Inc. Publishing as Prentice Hall.

Copyright 2010 Pearson Education, Inc. Publishing as Prentice Hall. .1 Limits of Sequences. CHAPTER.1.0. a) True. If converges, then there is an M > 0 such that M. Choose by Archimedes an N N such that N > M/ε. Then n N implies /n M/n M/N < ε. b) False. = n does not converge,

More information

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i := 2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]

More information

Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles

Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles Generalized Multivariate Rank Type Test Statistics via Spatial U-Quantiles Weihua Zhou 1 University of North Carolina at Charlotte and Robert Serfling 2 University of Texas at Dallas Final revision for

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA

TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA Statistica Sinica (014): Preprint 1 TWO-SAMPLE BEHRENS-FISHER PROBLEM FOR HIGH-DIMENSIONAL DATA Long Feng 1, Changliang Zou 1, Zhaojun Wang 1, Lixing Zhu 1 Nankai University and Hong Kong Baptist University

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

More Empirical Process Theory

More Empirical Process Theory More Empirical Process heory 4.384 ime Series Analysis, Fall 2008 Recitation by Paul Schrimpf Supplementary to lectures given by Anna Mikusheva October 24, 2008 Recitation 8 More Empirical Process heory

More information

j=1 π j = 1. Let X j be the number

j=1 π j = 1. Let X j be the number THE χ 2 TEST OF SIMPLE AND COMPOSITE HYPOTHESES 1. Multinomial distributions Suppose we have a multinomial (n,π 1,...,π k ) distribution, where π j is the probability of the jth of k possible outcomes

More information

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA J. Japan Statist. Soc. Vol. 37 No. 1 2007 53 86 MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA M. S. Srivastava* In this article, we develop a multivariate theory for analyzing multivariate datasets

More information

A note on profile likelihood for exponential tilt mixture models

A note on profile likelihood for exponential tilt mixture models Biometrika (2009), 96, 1,pp. 229 236 C 2009 Biometrika Trust Printed in Great Britain doi: 10.1093/biomet/asn059 Advance Access publication 22 January 2009 A note on profile likelihood for exponential

More information

Stein s Method and Characteristic Functions

Stein s Method and Characteristic Functions Stein s Method and Characteristic Functions Alexander Tikhomirov Komi Science Center of Ural Division of RAS, Syktyvkar, Russia; Singapore, NUS, 18-29 May 2015 Workshop New Directions in Stein s method

More information

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX Submitted to the Annals of Statistics HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX By Shurong Zheng, Zhao Chen, Hengjian Cui and Runze Li Northeast Normal University, Fudan

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data Yujun Wu, Marc G. Genton, 1 and Leonard A. Stefanski 2 Department of Biostatistics, School of Public Health, University of Medicine

More information

Minimum distance tests and estimates based on ranks

Minimum distance tests and estimates based on ranks Minimum distance tests and estimates based on ranks Authors: Radim Navrátil Department of Mathematics and Statistics, Masaryk University Brno, Czech Republic (navratil@math.muni.cz) Abstract: It is well

More information

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay Solutions to Final Exam 1. (13 pts) Consider the monthly log returns, in percentages, of five

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient

More information

arxiv:math/ v5 [math.ac] 17 Sep 2009

arxiv:math/ v5 [math.ac] 17 Sep 2009 On the elementary symmetric functions of a sum of matrices R. S. Costas-Santos arxiv:math/0612464v5 [math.ac] 17 Sep 2009 September 17, 2009 Abstract Often in mathematics it is useful to summarize a multivariate

More information

A Note on Jackknife Based Estimates of Sampling Distributions. Abstract

A Note on Jackknife Based Estimates of Sampling Distributions. Abstract A Note on Jackknife Based Estimates of Sampling Distributions C. Houdré Abstract Tail estimates of statistics are given; they depend on jackknife estimates of variance of the statistics of interest. 1

More information

arxiv: v5 [math.na] 16 Nov 2017

arxiv: v5 [math.na] 16 Nov 2017 RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem

More information

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis TAMS39 Lecture 10 Principal Component Analysis Factor Analysis Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content - Lecture Principal component analysis

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Improved Methods for Monte Carlo Estimation of the Fisher Information Matrix

Improved Methods for Monte Carlo Estimation of the Fisher Information Matrix 008 American Control Conference Westin Seattle otel, Seattle, Washington, USA June -3, 008 ha8. Improved Methods for Monte Carlo stimation of the isher Information Matrix James C. Spall (james.spall@jhuapl.edu)

More information

Lecture 5: Hypothesis tests for more than one sample

Lecture 5: Hypothesis tests for more than one sample 1/23 Lecture 5: Hypothesis tests for more than one sample Måns Thulin Department of Mathematics, Uppsala University thulin@math.uu.se Multivariate Methods 8/4 2011 2/23 Outline Paired comparisons Repeated

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho

Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho M. de Carvalho and F. J. Marques Abstract We discuss jackknife Euclidean likelihood-based inference methods, with a special focus on the

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Addressing ourliers 1 Addressing ourliers 2 Outliers in Multivariate samples (1) For

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case The largest eigenvalues of the sample covariance matrix 1 in the heavy-tail case Thomas Mikosch University of Copenhagen Joint work with Richard A. Davis (Columbia NY), Johannes Heiny (Aarhus University)

More information

ON THE STRONG LIMIT THEOREMS FOR DOUBLE ARRAYS OF BLOCKWISE M-DEPENDENT RANDOM VARIABLES

ON THE STRONG LIMIT THEOREMS FOR DOUBLE ARRAYS OF BLOCKWISE M-DEPENDENT RANDOM VARIABLES Available at: http://publications.ictp.it IC/28/89 United Nations Educational, Scientific Cultural Organization International Atomic Energy Agency THE ABDUS SALAM INTERNATIONAL CENTRE FOR THEORETICAL PHYSICS

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

STT 843 Key to Homework 1 Spring 2018

STT 843 Key to Homework 1 Spring 2018 STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ

More information

Extending the two-sample empirical likelihood method

Extending the two-sample empirical likelihood method Extending the two-sample empirical likelihood method Jānis Valeinis University of Latvia Edmunds Cers University of Latvia August 28, 2011 Abstract In this paper we establish the empirical likelihood method

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

Journal Club: Brief Introduction to Tensor Network

Journal Club: Brief Introduction to Tensor Network Journal Club: Brief Introduction to Tensor Network Wei-Han Hsiao a a The University of Chicago E-mail: weihanhsiao@uchicago.edu Abstract: This note summarizes the talk given on March 8th 2016 which was

More information

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING Statistica Sinica 3 (013), 135-155 doi:http://dx.doi.org/10.5705/ss.01.075 SEMIPARAMERIC ESIMAION OF CONDIIONAL HEEROSCEDASICIY VIA SINGLE-INDEX MODELING Liping Zhu, Yuexiao Dong and Runze Li Shanghai

More information

A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1.

A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1. A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE THOMAS CHEN AND NATAŠA PAVLOVIĆ Abstract. We prove a Beale-Kato-Majda criterion

More information