Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples

Size: px
Start display at page:

Download "Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples"

Transcription

1 Expected probabilities of misclassification in linear discriminant analysis based on 2-Step monotone missing samples Nobumichi Shutoh, Masashi Hyodo and Takashi Seo 2 Department of Mathematics, Graduate School of Science, Tokyo University of Science, -3, Kagurazaka, Shinjuku-ku, Tokyo , Japan 2 Department of Mathematical Information Science, Faculty of Science, Tokyo University of Science, -3, Kagurazaka, Shinjuku-ku, Tokyo , Japan Absact As concerns the asymptotic approximation for expected probabilities of misclassification in linear or quadratic discriminant analysis, theoretical results are given under some situations with complete data However it has not been enough to discuss that under missing samples In this paper, we derive that with explicit form in linear discriminant function for two groups consucted by 2-Step monotone missing samples For our purpose, some results of expectations concerning inverted Wishart maices are also obtained Finally the numerical evaluations of our result by Monte Carlo simulations are presented AMS 2000 Mathematics Subject Classification 62H2, 62H30 Key Words and Phrases: linear discriminant function; expected probability of misclassification; asymptotic approximation; monotone missing samples; missing observations Inoduction Linear discriminant analysis is well known as one of statistical procedures to assign p dimensional observation vector x into one of two populations Π and Π 2 If Π g g =, 2 has p dimensional normal disibution with mean vector µ g and common covariance maix Σ which are known where µ µ 2 and Σ is positive definite, linear discriminant function LDF is consucted as follows: Corresponding author address: nshutoh28@yahoocojp W = µ µ 2 Σ x 2 µ µ 2

2 x may be assigned to Π if W > 0, or Π 2 if W < 0 The probability of misclassification in equals Φ 2 exactly, where Φ denotes the cumulative density function of standard normal disibution and 2 µ µ 2 Σ µ µ 2 is Mahalanobis squared distance between Π and Π 2 However it is usual that both of parameters µ g and Σ are unknown Therefore their own estimators are substituted for µ g and Σ respectively in If N g observation vectors x g j g =, 2, j =,, N g which are can be observed completely from Πg, LDF suggested by Wald 944 is as follows: W c = x x 2 S x 2 x x 2, 2 where x g and S are sample mean vector from Π g and pooled sample covariance maix, and they are the maximum likelihood estimators MLEs of µ g and Σ respectively Although it is hard to obtain exact probabilities of misclassification under the situation where parameters are unknown, two types of asymptotic approximations for expected probabilities of misclassification EPMC have been considered Type I approximations are ones of framework such that N and N 2 are large and p is fixed Okamoto 963, 968 gave the approximation of this type by deriving asymptotic expansion formulas of LDF up to terms of the second order with respect to N, N 2, n, where n = N N 2 2 Memon and Okamoto 97 considered it in the case of quadratic discriminant function QDF Siotani and Wang 977 extended Okamoto 963, 968 and Memon and Okamoto 97 up to terms of the third order It has been known that these approximations are good for small p, but are poor as p is large In the face of this problem, Type II approximations are suggested under the framework such that N, N 2, p and n p are large As concerns the asymptotic approximations for EPMC in this type, Raudis 972 considered one of them in LDF for N = N 2 Fujikoshi and Seo 998 extended Raudis 972 to the case of N N 2 and also gave it in QDF Moreover Fujikoshi 2000 gave explicit error bounds for approximation for EPMC proposed by Lachenbruch 968 and discussed with a method of obtaining asymptotic expansion formulas for EPMC and their error bounds Besides the asymptotic approximation for EPMC proposed by Lachenbruch 968 is the one which can fit the situations in both Type I and Type II, so that the asymptotic approximation proposed Lachenbruch 968 can be used in various situations of data set without missing data concerns discussion under elliptical populations, see Wakaki 994 On the other hand, as concerns MLEs consucted by observation vectors including missing data, Srivastava 985 derived MLEs for one and two sample problems without condition of missing patterns by using incidence maix which can remove missing data from each observation vector Although they can not be obtained explicitly, Srivastava and Carter 986 proposed the 2 As

3 procedure to obtain them numerically by using Newton-Raphson method Also MLEs for one sample in the case of 2-Step monotone missing samples from multivariate normal disibution were derived explicitly by Anderson and Olkin 985 The k-step monotone missing samples from Π g are defined as data set having following form: x g x g p x g p x g p k x g p k x g N g x g N g x g N g x g x g N g 2 x g N g x g p N g p x g N g p N g p x g N g 2 p x g N g 2 p x g N g p k x g x g N g p k N g p k x g N g 2 p k x g N g p, 3 where p = p p k, N g = N g N g k and each part of denotes missing data Further Kanda and Fujikoshi 998 inoduced some properties of MLEs in the case of k = 2, 3 and general k They also gave asymptotic expansions for the disibutions of MLEs in the situation when the sample size is large In these days, Batsidis, Zografos and Loukas 2006 considered LDF based on 2-Step monotone missing samples and gave the asymptotic expression for probability density function of the probability of misclassification In this paper, as one of extension for Lachenbruch 968, we propose asymptotic approximation for EPMC where aining data has 2-Step monotone missing samples That is, we extend the results of Lachenbruch 968 in the case of LDF consucted by Batsidis, Zografos and Loukas 2006 Although Kanda and Fujikoshi 2004 derived that under known Σ, we propose that under unknown case as main result in this paper The organization of this paper is as follows In Section 2, we show method for obtaining MLEs in two groups in the case of 2-Step monotone missing samples and give some Lemmas which can support us to give the asymptotic approximation for EPMC considered in Section 3 In Section 3, we consider asymptotic approximation for EPMC in this case by using the results of Section 2 and derive the unbiased estimators of Mahalanobis squared distances which come in useful for obtaining the approximation of EPMC Further we evaluate our results in Section 3 numerically by Monte Carlo simulations in Section 4 Finally we present conclusion in this paper and future problems in Section 5 3

4 2 MLEs Based on 2-Step Monotone Missing Samples In this section, we derive the MLEs under assuming that 2-Step monotone missing samples are obtained from each group, ie, we extend MLEs given by Anderson and Olkin 985 to the case of two groups where k = 2 in 3 Therefore we assume each data set from Π g as follows: x g x g 2 x g p x g p x g p x g 2 x g 22 x g 2 p x g 2 p x g 2p x g x g N g N g 2 x g x g N g N g x g N g x g N g 2 x g x g N g 2 x g p N g p N g p x g N g p, x g N g p where p = p p 2 and N g = N g N g 2 Then we prepare notations and assumption concerning observation vector Let x g j x g j, xg 2j g =, 2, j =,, N g be the p dimensional observation vector from Π g which has complete data, where x g j = x g j,, xg jp j =,, N g and x g 2j = x g j p,, xg g jp j =,, N Now we assume disibution of observation vector: x g j N p µ g, Σ j =,, N g and xg j N p µ g, Σ j = N g,, N g respectively, where µ g = µ g µ g 2 and Σ = Σ Σ 2 Σ 2 Σ 22 µ g and Σ are partitioned according to blocks of data set Therefore µ g is p dimensional vector, µ g 2 is p 2 dimensional vector, Σ is p p maix, Σ 2 = Σ 2 is p p 2 maix and Σ 22 is p 2 p 2 maix respectively Then we consider ansformation of observation vector x g j by multiplying I p 0 Σ 2 Σ I p2 on the left-hand The mean vector η g and covariance maix of ansformed observation vector are η g η g η g 2 = µ g µ g 2 Σ 2 Σ µg 4 and Σ 0 0 Σ 22,

5 where Σ 22 = Σ 22 Σ 2 Σ Σ 2 and Σ has nonsingularity Therefore we consider the new parameters {η, η 2, Ψ} which are one to one corresponding to {µ, µ 2, Σ}, where Ψ Ψ Ψ = 2 Σ Σ = Σ 2 Ψ 2 Ψ 22 Σ 2 Σ Σ 22 Next the likelihood function to obtain MLEs of η, η 2 and Ψ is as follows: Lη, η 2, Ψ = Const Ψ 2 N N 2 Ψ 22 2 e 2 N g y g j 2 e 2 g= j= 2 N g g= j= y g 2j 2 N N ηg Ψ yg j ηg ηg 2 Ψ 22 yg 2j ηg 2, 2 where y g j = xg j and y g 2j = xg 2j Ψ 2x g j If we define the sample mean vectors x g T = x g 2F = N g N g x g j, N g j= N g j= x g 2j xg F = N g and x g L = N g 2 N g j= x g j, N g g xj, j=n g then by 2 and the well known correction of their own coefficients, we can obtain following MLEs and where Ψ = n Ψ 2 = 2 N g g= j= 2 N g g= j= x g j x g j η g = Ψ 22 = { 2 N g x g 2j n g= j= η g η g 2 = Ψ = Ψ Ψ 2 Ψ 2 Ψ 22 xg T xg j xg T, xg F xg j xg F xg 2F xg 2j xg 2F 5 x g T x g 2F Ψ 2 x g F 2, N g x g j g= j= xg F xg 2j xg 2F,

6 2 N g x g 2j g= j= 2 N g x g j g= j= xg 2F xg j xg F xg F xg j xg F 2 N g x g j g= j= xg F xg 2j xg 2F n = N N 2 2 and n = N N 2 2 Also these maices are expressed as follows: Ψ = n Γ Γ2, Ψ 2 = Γ Γ 2 and Ψ 22 = Γ n respectively Γ and Γ 2 are maices such that Γ Γ = Γ 2 Γ 2 Γ 22 and where = n S { 2 g N Γ 2 = n 2 S 2 N g } 2 N g= g x g F xg L xg F xg L, S = N 2 g x g j n g= j= S 2 = 2 N g g xj n 2 g= j=n g x g x g j x g, xg L xg j xg L and n 2 = N 2 N Clearly we can find that S and S 2 are pooled sample covariance maices of x g j g =, 2, j =,, N g and xg j g =, 2, j = N g,, N g respectively Also n is needed to be larger than p by nonsingularity of Γ MLEs derived in this section have natural forms of extension of Anderson and Olkin 985 Here we have Lemmas concerning the disibutions and expectations of 22 Lemma 2 Γ, Γ Γ2, Γ 2 given Γ and Γ 22 have following disibutions respectively: Γ W p n, Σ, Γ Γ2 = n Ψ W p n, Σ, VecΓ 2 Γ N p p 2 VecΓ Σ Σ 2, Σ 22 Γ, Γ 22 = n Ψ 22 W p2 n p, Σ 22, where W p n, Σ denotes Wishart disibution with the parameters n and Σ }, 6

7 Proof: Since Γ has W p n, Σ, it has seen that Γ has W p n, Σ and Γ 22 has W p 2 n p, Σ 22 The conditional disibution of Γ 2 function of Γ, Γ 2 given Γ is derived by joint probability density For a proof, eg, see Siotani, Hayakawa and Fujikoshi 985 Moreover the disibution of Γ 2 can be derived since n 2 S 2 has W p n 2, Σ, the other maix has W p 2, Σ and they are independent Also Γ and Γ2 are independently disibuted, which proves the result Lemma 22 Let C i and D i be p i p i i =, 2 constant maices respectively, then the expectations as follows can be obtained: E Γ = E Γ Γ2 = E Γ 22 = E Γ C Γ = n p Σ, n p Σ, n p Σ 22, n p 2 n p n p n p 3 Σ C Σ n p n p n p 3 { } Σ C Σ C Σ, Σ E Γ Γ2 C Γ Γ2 = E Γ 22 C 2Γ 22 = n p 2 n p n p n p 3 Σ C Σ n p n p n p 3 { Σ C Σ C Σ Σ n p 2 n pn p n p 3 Σ 22 C 2Σ 22 n pn p n p 3 { Σ 22 C 2Σ 22 C 2 Σ 22 }, Σ 22 }, E C Γ D Γ = n p 2 n p n p n p 3 C Σ D Σ 7

8 n p n p n p 3 { } D Σ C Σ D Σ C Σ, respectively Proof: Lemma 22 is derived by the original properties concerning the expectation of inverted Wishart maix The results can be derived by Haff 979 and Marx 98 At the end of this section, we show the other expression of some terms included in V m defined by Section 3 Here we give the definition of Type I beta disibution Definition 23 A p p symmeic random maix Y disibuted as Type I beta disibution with the parameters a and b is one which has the following probability density function: {β p a, b} dety a 2 p deti p Y b 2 p, 0 < Y < I p, where a > p /2, b > p /2 and β p a, b is p dimensional beta function given by β p a, b = Γ paγ p b Γ p a b for the real parts of a and b are larger than p /2 respectively Further the function Γ p a consucting β p a, b is the following function such that p Γ p a = π 4 pp Γ a 2 i, where Γ is Gamma function i= As concerns the details and properties concerning the above definition and the processes of calculating β p a, b and Γ p a, eg, see Gupta and Nagar 2000 Besides Type I beta disibution with the parameters a and b is denoted by B I pa, b in this paper Lemma 24 Let A be a p p constant and symmeic maix Then Γ Γ2 Σ Γ A can be also rewritten with following expression including random maices: β I,p W W Σ 2 A Σ 2 W W 2 2, where β I,p = W W 2 2 W W W 2 2 has Bp I n /2, n 2 2/2, W and W 2 have W p n, I p and W p n 2 2, I p respectively Proof: Note that β I,p has B I p n /2, n 2 2/2 For a proof, see, eg, Gupta and Nagar 2000 Moreover by making use of the properties of ace and Wishart maices, the proof can be completed 8

9 3 Asymptotic Approximation for EPMC Now we consider the asymptotic approximation for EPMC when the aining data has 2-Step monotone missing samples The approximation is obtained by making use of Lachenbruch 968 and results obtained in Section 2 If µ g and Σ are unknown, the aining data has 2-Step monotone samples and x does not have missing data, then LDF is consucted by the same way as 2: W m = µ µ 2 Σ x 2 µ µ 2, 3 where µ, µ 2 and Σ are MLEs obtained by aining data variance of W m given µ, µ 2 and Σ under x Π : Now we find expectation and EW m µ, µ 2, Σ = U m, VarW m µ, µ 2, Σ = V m, where U m = µ µ 2 Σ µ µ 2 D2 m, 32 D 2 m = µ µ 2 Σ µ µ 2, 33 V m = µ µ 2 Σ Σ Σ µ µ 2 34 Further if we define then Z m Z m = V 2 m µ µ 2 Σ x µ, 35 is independent of U m, V m and disibuted as N0, under the situation given µ, µ 2, Σ By using 32 35, 3 can be also rewritten as W m = V 2 m Z m U m 36 Thus W m can be expressed by U m, V m and Z m in the same way as Lachenbruch 968 in spite of 2-Step monotone missing samples As concerns the probability of misclassification where x is assigned to Π 2 when x comes from Π, it is expressed by using 36 as Pr W m 0 x Π = Pr Z m V 2 m U m x Π 37 Since Z m is disibuted as N0, if x comes from Π, 37 can be rewritten easily as ΦV 2 m U m Therefore EPMC of 3 where x is assigned to Π 2 when x comes from Π is obtained as e m 2 = E Um,V m ΦV 2 m U m 38 9

10 Here we consider the asymptotic approximation for EPMC in 3 disibuted as normal disibution exactly, but is closely normal asymptotically asymptotic approximation of the right-hand of 38, we propose In actual, W m is not Thus as an e m 2 Φ{EV m } 2 EUm 39 obtained by substituting their expectations of U m and V m For the case that Σ is known, Kanda and Fujikoshi 2004 proposed 39 where EU m = 2 p N N 2 2 N N 2 p N N 2 N N 2 EV m = 2 p N N 2 N N 2 p 2N N 2 N N 2 under the framework such that N N N positive const and positive const 2 N 2 as N, N 2, N 2 and N 2 2 tend to infinity if p and p 2 are fixed or N N 2 positive const, N N 2 positive const and δ δ = O as N, N 2, N 2, N 2 2, p and p 2 tend to infinity, where δ = µ µ 2 If p and p 2 are fixed, then this approximation corresponds to Type I approximation If p and p 2 are also large, that corresponds to Type II approximation Under unknown Σ, for the purpose of obtaining EU m and EV m, ie, deriving approximation given in 39, we have the following Lemmas Lemma 3 Let VecX N p p 2 VecM, Σ Γ, M = m ij, Σ = σ ij and Γ = γ ij Then the expectations with respect to X are Ex i j x i2 j 2 = σ i i 2 γ j j 2 m i j m i2 j 2, Ex i j x i2 j 2 x i3 j 3 = m i j σ i2 i 3 γ j2 j 3 m i2 j 2 σ i i 3 γ j j 3 m i3 j 3 σ i i 2 γ j j 2 m i j m i2 j 2 m i3 j 3, Ex i j x i2 j 2 x i3 j 3 x i4 j 4 = σ i i 4 γ j j 4 σ i2 i 3 γ j2 j 3 σ i i 2 γ j j 2 σ i4 i 3 γ j4 j 3 σ i i 3 γ j j 3 σ i4 i 2 γ j4 j 2 m i j m i2 j 2 σ i4 i 3 γ j4 j 3 m i j m i3 j 3 σ i4 i 2 γ j4 j 2 m i2 j 2 m i3 j 3 σ i i 4 γ j j 4 m i j m i4 j 4 σ i2 i 3 γ j2 j 3 m i4 j 4 m i2 j 2 σ i i 3 γ j j 3 m i4 j 4 m i3 j 3 σ i i 2 γ j j 2 m i j m i2 j 2 m i3 j 3 m i4 j 4, 0

11 Proof: This Lemma was derived by considering the characteristic function of X eg, see Gupta and Nagar 2000 Moreover Lemma 2 and Lemma 3 with putting X = Γ 2 concerning conditional expectations of Ψ 2 given Γ lead the following Lemma Lemma 32 Let C ij p i p j and D ij p i p j be constant maices respectively, then the conditional expectations given Γ as follows can be obtained: E Ψ 2 Γ = Σ Σ 2, E Ψ 2 C 2 Ψ 2 Γ = Γ C 2Σ 22 Σ Σ 2C 2 Σ Σ 2, E Ψ 2 C 22 Ψ 2 Γ = C 22Σ 22 Γ Σ Σ 2C 22 Σ 2 Σ, E Ψ 2 C Ψ 2 Γ = Γ C Σ 22 Σ 2 Σ C Σ Σ 2, E Ψ 2 C Ψ 22 2 C Ψ 2 Γ = C Γ Σ Σ 2C 22 Σ 22 Γ C Σ Σ 2C 22Σ 22 C 22 Σ 22 Γ C Σ Σ 2 Σ Σ 2C 22 Σ 2 Σ C Σ Σ 2, E Ψ 2 C Ψ 2 2 C Ψ 22 2 Γ = C 22 Σ 22 Σ Σ 2C 2 Γ C 2 Σ Σ 2C 22 Σ 22 Γ Γ C 2Σ 22 C 22 Σ 2 Σ Σ Σ 2C 2 Σ Σ 2C 22 Σ 2 Σ, E Ψ 2 C 22 Ψ 2 C Ψ 2 D 22 Ψ 2 Γ = C 22 Σ 22 D 22 Σ 22 Γ C C 22 Σ 22 D 22 Σ 22 Γ C Γ Σ Σ 2C 22 Σ 22 D 22Σ 2 Σ C Γ Γ C Γ C 22 Σ 22 D 22Σ 22 Γ C Γ D 22 Σ 22 Σ Σ 2C 22 Σ 2 Σ C Γ C 22 Σ 2 Σ C Σ Σ 2D 22 Σ 22 Γ Σ Σ 2C 22 Σ 22 D 22 Σ 2 Σ Γ C Σ Σ 2C 22Σ 22 D 22 Σ 2 Σ C 22 Σ 22 Γ C Σ Σ 2D 22 Σ 2 Σ Σ Σ 2C 22 Σ 2 Σ C Σ Σ 2D 22 Σ 2 Σ, E C Ψ 2 2 D Ψ 2 2 Γ = C 2 Γ D 2Σ 22 C 2 Σ Σ 2 D 2 Σ Σ 2, E = C Ψ 22 2 C Ψ 2 C 22 Σ 2 Σ C 22 Σ 22 C Γ Γ Ψ 2 C 2 C 22 Σ 2 Σ C Σ Σ 2 Γ C 2Σ 22 C Σ Σ 2C 2 Σ Σ 2C 2, C 22 Σ 22 C 2 Γ C Σ Σ 2

12 E C Ψ 22 2 C Ψ 2 D Ψ 22 2 D Ψ 2 Γ = C 22 Σ 22 D 22 Σ 22 Γ C Γ D C 22 Σ 22 Γ C D 22 Σ 22 Γ D C 22 Σ 22 D 22Σ 22 Γ C Γ D C 22 Σ 2 Σ C Σ Σ 2 D 22 Σ 22 Γ D C 22 Σ 2 Σ C Γ D Σ Σ 2D 22Σ 22 C 22 Σ 22 D 22 Σ 2 Σ D Γ C Σ Σ 2 C 22 Σ 2 Σ C Γ D Σ Σ 2D 22 Σ 22 C 22 Σ 22 D 22Σ 2 Σ D Γ C Σ Σ 2 C 22 Σ 22 Γ C D 22 Σ 2 Σ D Σ Σ 2 C 22 Σ 2 Σ C Σ Σ 2 D 22 Σ 2 Σ D Σ Σ 2 Proof: By making use of results of Lemma 2 and Lemma 3, we can obtain the above results For some expectations of the above, the concrete methods for calculating them will be given in Appendix A The other expectations can be also obtained by the methods similarly Furthermore we also give the some results of calculating expectations by Lemma 24, Konno 988 and Marx 98 Lemma 33 The following expectations can be calculated as E Γ Γ2 Σ Γ Σ E µ µ 2 Γ Γ2 Σ Γ µ µ 2 = p c, = c δ 2, where c = n /{n p n p 3n p }, δ 2 = µ µ 2 Σ µ µ 2 Proof: The proof will be completed in Appendix B By Lemma 22, Lemma 32 and Lemma 33, we can obtain the following theorem Theorem 34 If µ g and Σ are unknown, then the expectations of U m and V m defined by 32 and 34 are EU m = 2 u δ 2 p N N 2 N N 2 2 u 3 2 p N N 2 N N 2 p 2N N 2 N N 2 2 p 2u 2 δ 2 p N N 2, N N 2 EV m = v δ 2 p N N 2 N N 2 p 2 v 2 δ 2 p N N 2 N N 2 2

13 where u = u 2 = u 3 = v = v 2 = v 3 = v 3 2 p N N 2 N N 2 p 2N N 2 N N 2 n n p n n p, n n p n p, n n p, n 2 n n p n p n p 3 2p 2 nn n n p n p 3n p n p n 2 n n pn p n p 3, n 2 n pn p n p 3 p 2 n 2 n n pn p 3n p n p n p 3 2n 2 n n pn p n p 3n p n p n p 3 p n 2 n p 2 n pn p n p 3n p n p 3 2n 2 n pn p n p 3n p n p 3, n 2 n n pn p n p 3, Proof: The proof will be completed in Appendix C Thus we can derive the asymptotic approximation for EPMC in 39 with explicit form By Theorem 34, we derive that under the framework such that N N N positive const and 2 N 2 positive const as N, N 2, N 2 and N 2 2 tend to infinity if p and p 2 are fixed or N N 2 positive const, N N 2 positive const, δ δ = O, n p and n p as N, N 2, N 2, N 2 2, p and p 2 tend to infinity In addition, we can find the following corollary as checking on the results in Theorem 34 3

14 Corollary 35 If we put N g = N g g =, 2 and note that p = p p 2, EU m and EV m given by Theorem 34 can be reduced to EU and EV derived in Lachenbruch 968 as follows: n EU = 2n p EV = { 2 N N 2 p N N 2 n 2 n n pn p n p 3 }, { 2 N N 2 p } N N 2 Besides Lachenbruch 968 proposed the asymptotic approximation for EPMC of LDF in 2 by using EU and EV : e c 2 Φ{EV } 2 EU 30 As it were, this corollary implies asymptotic approximation proposed in this paper is one of extension for Lachenbruch 968 Although we can obtain the asymptotic approximation for EPMC, we can not use µ g and Σ actually since they are usually unknown Then it is natural to consider substituting d 2 m µ µ 2 Σ µ µ 2 and D2 m for δ 2 and 2 respectively However under the situation where sample sizes are not very large, the approximation may be affected by the biases of d 2 m and D2 m Therefore we also propose the unbiased estimators of δ 2 and 2 respectively at the end of this section Theorem 36 The unbiased estimators δ 2 and 2 of δ 2 and 2 which are included in the asymptotic approximation for EPMC given in Theorem 34 can be obtained as following forms: δ 2 = b d 2 m p N N 2 N N 2, 3 2 = b 2 D 2 m b 2 b b 3 δ 2 N N 2 p 2N N 2 b b 2 p N N 2 N N 2 p p 2 N N 2 b 3 n N N 2 where b = n/n p, b 2 = n /n p and b 3 = n /n p Proof: The proof will be completed in Appendix D, 32 Besides in the case of complete data, the unbiased estimator of Mahalanobis squared distance has the following form by using D 2 = x x 2 S x x 2 : n p D 2 N n N 2 N N 2 p 33 As concerns the asymptotic approximation proposed in this section, it will be evaluated by Monte Carlo simulations in Section 4 Then the above unbiased estimators will be used in stead of Mahalanobis squared distances 4

15 4 Simulation Studies We are interested in the accuracy of the asymptotic approximation for EPMC proposed in this paper Although we are also interested in comparing it with the other ones, the asymptotic approximations for EPMC in the case of missing samples had not been proposed before Therefore we compare the accuracy of it with the other ones for the case of complete data, eg, Lachenbruch 968 and Fujikoshi and Seo 998 As concerns the asymptotic approximation for EPMC proposed by Lachenbruch 968, see Corollary 35 On the other hand, Fujikoshi and Seo 998 proposed the following the asymptotic approximation for EPMC of LDF presented in 2: where ζ = N N 2 /2 p { 2 2 N N 2 e c 2 Φζ, 4 pn N 2 N N 2 } { 2 pn N 2 N N 2 } /2 Then we start to compare the accuracy of 39, 30 and 4 by Monte Carlo simulations As you know, in actual, the exact Mahalanobis squared distances δ 2 and 2 are unknown respectively Thus we use the unbiased estimator 33 in stead of 2 included in the asymptotic approximations 30 and 4 In similar, the unbiased estimators 3 and 32 are used in stead of δ 2 and 2 in 39 respectively As concerns the settings of sample sizes concerning complete data, dimension and Mahalanobis distance, we chose the parameters which are similar to Fujikoshi and Seo 998 The settings of sample sizes and dimensions concerning missing samples are as follows: N 2, N 2 2 = 0, 0, 20, 20, 20, 40, 40, 40, and p, p 2 =,, 3, 2, 5, 5 The tables in these simulations are separated by the setting of dimensions Besides the values of asymptotic approximations in tables are average values of the ones calculated,000,000 times by Monte Carlo simulations and evaluated the accuracy of ones by them The results are shown in Tables 3 respectively 5 Conclusion and Future Problems In this paper, we extended the MLEs of 2-Step monotone missing samples to that of two groups in Section 2 By making use of the MLEs, we considered LDF based on 2-Step monotone missing samples Moreover by giving the many results concerning the expectations which can 5

16 not be calculated easilyeg, Lemma 33, we proposed the asymptotic approximation for EPMC with an explicit form which is Lachenbruch type of the LDF considered by us As it were, it is one of theoretical extensions of Lachenbruch 968 We proposed not only the asymptotic approximation but also the unbiased estimators which are used in calculating the approximation actually in Section 3 Finally we evaluated our result numerically by comparing with the other approximations in Section 4 By simulations, we find that our result is more accurate than the ones which had been proposed before Although the asymptotic approximation proposed in this paper is limited only in the case of 2-Step monotone missing samples, consucting LDF by missing samples holds good method for making EPMC lower and giving more accurate asymptotic approximation for EPMC These results imply that method proposed in this paper can be considered as one of useful methods for experimenters who have put linear discriminant analysis to practical use In the case of complete data, in general, the results of Fujikoshi and Seo 998 may be more accurate than Lachenbruch 968 As left problems, although it may be easy to be suck with an idea of the extension of the result in this paper to k-step monotone missing samples, the calculations of expectations will be much complicated If we limit the discussion in the case of 2-Step monotone missing samples, also we will be able to derive theoretical error bounds for the asymptotic approximation proposed in this paper by the method which is similar to Fujikoshi 2000 By the results of simulations in the case of complete data, the extensions of the other types of approximations eg, Okamoto 963, 968, Fujikoshi and Seo 998 to the case of 2-Step monotone missing samples will be needed Finally we point many discussions eg, testing equality procedure, MANOVA based on monotone missing samples have been left Acknowledgements We wish to express much gratitude to Professor Yasunori Fujikoshi for suggesting the ideas and giving useful and valuable advices in this paper 6

17 References Anderson, T W and Olkin, I 985 Maximum-likelihood Estimation of the Parameters of a Multivariate Normal Disibution Linear Algebra and Its Applications, 70, 47 7 Batsidis, A, Zografos, K and Loukas, S 2006 Errors in Discrimination with Monotone Missing Data from Multivariate Normal Populations Computational Statistics & Data Analysis, 50, Fujikoshi, Y 2000 Error Bounds for Asymptotic Approximations of the Linear Discriminant Function When the Sample Sizes and Dimensionality are Large Journal of Multivariate Analysis, 73, 7 Fujikoshi, Y and Seo, T 998 Asymptotic Approximations for EPMC s of the Linear and Quadratic Discriminant Functions When the Sample Sizes and the Dimension are Large Random Operators and Stochastic Equations, 6, Gupta, A K and Nagar, D K 2000 Maix Variate Disibutions Chapman and Hall/CRC, Monographs and Surveys in Pure and Applied Mathematics Haff, L R 979 An Identity for the Wishart Disibution with Applivations Journal of Multivariate Analysis, 9, Kanda, T and Fujikoshi, Y 998 Some Basic Properties of the MLE s for a Multivariate Normal Disibution with Monotone Missing Data American Journal of Mathematical and Management Sciences, 8, 6 90 Kanda, T and Fujikoshi, Y 2004 Linear Discriminant Function and Probabilities of Misclassification with Monotone Missing Data Proceedings of The Eighth China-Japan Symposium on Statistics, Konno, Y 988 Exact Moments of the Multivariate F and Beta Disibutions Journal of The Japan Statistical Society, 8, Lachenbruch, P A 968 On Expected Probabilities of Misclassification in Discriminant Analysis, Necessary Sample Size, and a Relation with the Multiple Correlation Coefficient Biomeics, 24, Marx, D G 98 Aspects of the Maic t-disibution Ph D thesis, Free State, RSA University of The Orange Memon, A Z and Okamoto, M 97 Asymptotic Expansion of the Disibution of the Z Statistic in Discriminant Analysis Journal of Multivariate Analysis,, Okamoto, M 963 An Asymptotic Expansion for the Disibution of the Linear Discriminant Function Annals of Mathematical Statistics, 34,

18 Okamoto, M 968 Correction to An Asymptotic Expansion for the Disibution of the Linear Discriminant Function Annals of Mathematical Statistics, 39, Raudis, S 972 On the Amount of Priori Information in Designing the Classification Algorithm Engineering Cybernetics, 0, 7 78 in Russian Siotani, M, Hayakawa, T and Fujikoshi, Y 985 Modern Multivariate Statistical Analysis : A Graduate Course Handbook American Sciences Press, Columbus, Ohio Siotani, M and Wang, R H 977 Asymptotic Expansions for Error Rates and Comparison of the W - procedure and Z-procedure in Discriminant Analysis Multivariate Analysis-IV P R Krishnaiah, ed, North Holland, Srivastava, M S 985 Multivariate Data with Missing Observations Theory and Methods, 4, Communication in Statistics Srivastava, M S and Carter, E M 986 The Maximum Likelihood Method for Non-Response in Sample Survey Survey Methodology, 2, 6 72 Wakaki, H 994 Discriminant Analysis under Elliptical Populations Hiroshima Mathematical Journal, 24, Wald, A 944 On a Statistical Problem Arising in the Classification of an Individual into One of Two Groups Annals of Mathematical Statistics, 5,

19 Appendix A Proof of Lemma 32 We show the proof for calculations E Ψ 2 C Ψ 22 2 C Ψ 2 Γ A and E C Ψ 2 2 D Ψ 2 2 Γ A2 At First, we complete the proof of A A is also expressed as follows: Γ E Γ 2 C 22Γ 2 C Γ 2 Γ, A3 where C = Γ C Γ Therefore i, jth element of A3 is expressed by Γ 2 = Γij 2, Γ 2 = Γij 2, C 22 = C ij 22 and C = C ij as follows: E Γ 2 C 22Γ 2 C Γ 2 Γ ij By Lemma 3 with putting = = p 2 p 2 p p i = i 2 = i 3 = i 4 = p 2 p 2 p p i = i 2 = i 3 = i 4 = E Γ ii 2 Ci i 2 22 Γi 2i 3 2 Ci 3i 4 Γi 4j 2 Γ C i i 2 22 Ci 3i 4 E Γ i i 2 Γi 2i 3 2 Γji 4 2 Γ A4 A4 is also expressed by X = Γ 2, M = Σ 2Σ Γ, Σ = Σ 22 and Γ = Γ, M = m ij, Σ = σ ij, Γ = γ ij, M = m ij, Σ = σ ij and Γ = γ ij as follows: E Γ 2 C 22Γ 2 C Γ 2 Γ ij = p 2 p 2 p p i = i 2 = i 3 = i 4 = γ ii4 C i 4i 3 m i 3 i 2 C i 2i { m ii C i i 2 22 σ i 2 jc i 3i 4 γ i 4 i 3 22 σ i j γ ii3 C i 3i 4 } m ii C i i 2 22 m i 2 i 3 C i 3i 4 m i 4 j m i 4 jc i i 2 22 σ i 2 i Therefore by noting Σ and Γ are symmeic, E Γ 2 C 22Γ 2 C Γ 2 Γ = C Γ Γ Σ Σ 2C 22 Σ 22 Γ C Γ Σ Σ 2C 22Σ 22 Γ C Γ Σ Σ 2 C 22 Σ 22 Γ Σ Σ 2C 22 Σ 2 Σ Γ C Γ Σ Σ 2 A5 9

20 By A3, multiplying Γ on the left-hand of A5 and putting C = Γ C Γ into A5 lead the result to prove Next we complete the proof of A2 A2 is also expressed as E C 2Γ 2 D 2Γ 2 Γ, A6 where C 2 = C 2 Γ and D 2 = D 2 Γ Therefore A6 is expressed by Γ 2 = Γij 2, Γ 2 = Γij 2, C 2 = C ij 2 and D 2 = D ij 2 as follows: E C 2Γ 2 D 2Γ 2 By using Lemma 3 with putting Γ = = p 2 p p 2 p i = i 2 = i 3 = i 4 = p 2 p p 2 p i = i 2 = i 3 = i 4 = E Γ i i 2 2 Γi 3i 4 2 Γ E C i i 2 2 Γi 2i 2 Di 3i 4 2 Γi 4i 3 2 Γ C i i 2 2 Di 3i 4 2 A7 again, A7 is also expressed by X = Γ 2, M = Σ 2Σ Γ, Σ = Σ 22 and Γ = Γ M = m ij, Σ = σ ij, Γ = γ ij, M = m ij and Σ = σ ij as follows: E C 2Γ 2 D 2Γ 2 Γ = p 2 p p 2 p C i i 2 i = i 2 = i 3 = i 4 = C i i 2 2 m i 2 i D i 3i 4 2 m i 4 i 3 2 γ i 2 i 4 D i 4i 3 2 σ i 3 i Therefore by noting that Σ is symmeic, E C 2Γ 2 D 2Γ 2 Γ = C 2Γ D 2Σ 22 D 2Γ Σ Σ 2 C 2Γ Σ Σ 2 A8 By A6, putting C 2 = C 2 Γ and D 2 = D 2 Γ into A8 leads the result to prove The other expectations can be obtained in the same way as the above Thus Lemma 32 has been completed 20

21 Appendix B Proof of Lemma 33 By Lemma 24, Γ Γ2 Σ Γ A = β I,p W W Σ 2 A Σ 2 W W 2 2, B where β I,p W W 2 2 W W W 2 2 has B I p n /2, n 2 2/2 Since β I,p is independent of W W 2, we can calculate conditional expectation of B given W W 2 Besides Konno 988 gave the expectation of β I,p as follows: E β I,p = n p n p I p Thus the conditional expectation to obtain can be calculated as n p n p W W 2 Σ 2 A Σ 2 W W 2 B2 Next we also calculate the expectation of B2 concerning W W 2 Since W W 2 is disibuted as W p n, I p, the expectation can be obtained by Marx 98: E Γ Γ2 Σ Γ A = c Σ 2 A Σ 2, B3 where c = n /{n p n p 3n p } By putting A = Σ and A = µ µ 2 µ µ 2 into B3, Lemma 33 has been completed Appendix C Proof of Theorem 34 Since µ g = Ip 0 Ψ 2 I p2 η g, Ip 0 Σ = Ψ 2 I p2 Ψ 0 0 Ψ 22 I p Ψ 2 0 I p2 and Ip 0 Ψ 0 Ip Ψ 2, Σ = Ψ 2 I p2 0 Ψ 22 0 I p2 U m is also expressed as U m = η η 2 Ψ η µ η 2 η Ψ 22 η 2 Ψ 2 µ µ 2

22 2 η η 2 Ψ η η 2 2 η = x T x2 T Ψ x T µ x 2F x2 2F x F x2 F Ψ 2 2 x 2F x2 2F x 2 η 2 2 2F x2 2F Ψ 22 η 2 η 2 2 Ψ 22 x 2F µ 2 Ψ 22 Ψ 2 x F µ x F x2 F Ψ 2 Ψ 22 Ψ 2 x F µ 2 x Ψ 22 x 2F x2 2F x 2 x F x2 F Ψ 2 Ψ Ψ 22 2 x F x2 F 2F x2 2F T x2 T Ψ 22 x 2F µ 2 Ψ x T x2 T Ψ 22 Ψ 2 x F x2 F Here all sample mean vectors x g T, xg F and xg 2F are independent of Γ, Γ 2 respectively Then the expectation of U m given Γ and Γ 2 is EU m Γ, Γ 2 = n N N 2 2N N 2 n N N 2 2N N 2 nn N 2 2N N 2 n N N 2 N N 2 n N N 2 2N N 2 Γ 22 Σ 22 Γ 22 Σ 2Σ Σ 2 Γ Γ2 Σ Γ Γ 2 Γ 22 Σ 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ n 2 µ µ 2 Γ Γ2 µ µ 2 n µ µ 2 Γ Γ 2 Γ 22 µ 2 µ 2 2 n 2 µ 2 µ 2 2 Γ 22 µ 2 µ 2 2 n 2 µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 Now we express EU m Γ, Γ 2 by using EU m Γ, Γ 2 and EU m2 Γ, Γ 2 which are separated this expectation given Γ and Γ 2 into the parts containing Γ Γ 2 and not respectively Therefore where EU m Γ, Γ 2 = EU m Γ, Γ 2 EU m2 Γ, Γ 2, EU m Γ, Γ 2 = n N N 2 N N 2 Γ Γ 2 Γ 22 Σ 2 22

23 n N N 2 2N N 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ n µ µ 2 Γ Γ 2 Γ 22 µ 2 µ 2 2 n 2 µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 and EU m2 Γ, Γ 2 = n N N 2 2N N 2 n N N 2 2N N 2 nn N 2 2N N 2 Γ 22 Σ 22 Γ 22 Σ 2Σ Σ 2 Γ Γ2 Σ n 2 µ µ 2 Γ Γ2 µ µ 2 n 2 µ 2 µ 2 2 Γ 22 µ 2 µ 2 2 Then we consider EU m at first Since U m does not include Γ 2 and Γ 22 is independent of Γ, Γ 2, we can obtain EU m Γ, Γ 2 By Lemma 22, EU m Γ, Γ 2 is as follows: EU m Γ, Γ 2 = n n p µ µ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 n 2n p µ µ 2 n N N 2 N N 2 n p n N N 2 2N N 2 n p Γ Γ 2 Σ 22 Γ 2 Γ µ Σ 22 Σ 2Γ Γ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ Moreover EU m Γ can be obtained by Lemma 22 and Lemma 32 EU m Γ = p 2 n 2n p µ µ 2 Γ µ µ 2 n n p µ µ 2 n 2n p µ µ 2 p 2n N N 2 2N N 2 n p n N N 2 2N N 2 n p Σ Σ 2Σ 22 µ 23 2 µ 2 2 Σ Σ 2Σ 22 Σ 2Σ µ Γ Σ Σ 22 Σ 2Σ Σ 2 µ 2 µ 2

24 Finally by Lemma 22, p 2 n EU m = 2n p n p µ µ 2 n n p µ µ 2 n 2n p µ µ 2 Σ Σ 2Σ 22 µ p p 2 n N N 2 2N N 2 n p n p n N N 2 2N N 2 n p Σ µ 2 µ 2 µ 2 2 Σ Σ 2Σ 22 Σ 2Σ µ On the other hand, we can calculate EU m2 with leaving Γ Γ 2 making use of Lemma 22, n EU m2 = 2n p µ µ 2 n 2n p µ 2 µ 2 n N N 2 2N N 2 n p p nn N 2 2N N 2 n p µ 2 Σ 22 Σ 2Σ Σ 2 C Σ µ 2 Σ 22 µ out of consideration By µ 2 2 µ 2 2 Σ 22 Σ 2Σ Σ 2 p 2n N N 2 2N N 2 n p C2 Thus EU m = EU m EU m2 is derived by C and C2 On the other hand, V m is expressed as V m = η η 2 Ψ Ψ Ψ η η 2 2 η η 2 Ψ Ψ Ψ 2 Ψ 22 η 2 η η η 2 Ψ Ψ Ψ 2 Ψ 22 η 2 η 2 2 η 2 η 2 2 Ψ Ψ 22 2 Ψ Ψ 2 Ψ 22 η 2 η η 2 η 2 2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 η 2 η 2 2 η 2 η 2 2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 η 2 η 2 2 η 2 η 2 2 Ψ 22 Ψ 22 Ψ 22 η 2 η 2 2 = x T x2 T Ψ Ψ Ψ x T x2 T 2x T x2 T Ψ Ψ Ψ 2 24 Ψ 22 x 2F x2 2F

25 2x 2x 2x T x2 T T x2 T T x2 T x 2F x2 2F 2x F x2 F Ψ 2 x F x2 F Ψ 2 2x 2F x2 2F 2x F x2 F Ψ 2 2x F x2 F Ψ 2 2x F x2 F Ψ 2 x 2F x2 2F 2x F x2 F Ψ 2 x F x2 F Ψ 2 x 2F x2 2F Ψ Ψ Ψ 2 Ψ Ψ 22 2 x F x2 Ψ Ψ Ψ 2 Ψ 22 x 2F x2 2F F Ψ Ψ Ψ 2 Ψ Ψ 22 2 x F x2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 Ψ 2 Ψ Ψ 2 F Ψ 22 x 2F x2 2F Ψ 22 x 2F x2 2F Ψ Ψ 22 2 Ψ Ψ 2 Ψ Ψ 22 2 x F x2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 x 2F x2 2F Ψ Ψ 22 2 Ψ Ψ 2 Ψ 22 x 2F x2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 x 2F x2 2F 2F F Ψ 22 Ψ 2 Ψ Ψ 2 Ψ Ψ 22 2 x F x2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 Ψ 22 2x F x2 F Ψ 2 x F x2 F Ψ 2 Ψ 22 Ψ 2 Ψ Ψ 2 Ψ 22 x 2F x2 2F Ψ 22 x 2F x2 2F Ψ 22 Ψ 2 Ψ Ψ 2 Ψ Ψ 22 2 x F x2 Ψ 22 x 2F x2 Ψ 22 Ψ 22 2F Ψ 22 x 2F x2 2F Ψ 22 Ψ 22 Ψ Ψ 22 2 x F x2 F Also in this calculation, we separate EV m Γ, Γ 2 into the terms containing Γ Γ 2 in the same way as calculating EU m Thus we express EV m Γ, Γ 2 as follows: EV m Γ, Γ 2 = EV m Γ, Γ 2 EV m2 Γ, Γ 2, where EV m Γ, Γ 2 and EV m2 Γ, Γ 2 are the terms which contain Γ Γ 2 and do not respectively First EV m Γ, Γ 2 is calculated as follows: EV m Γ, Γ 2 = 4nn N N 2 N N 2 Γ Γ2 Σ Γ Γ 2 Γ 22 Σ 2 2nn µ µ 2 Γ Γ2 Σ Γ Γ 2 Γ 22 µ 2 µ 2 2n2 N N 2 N N 2 F F Γ 22 Σ 22 Γ 22 Γ 2 Γ Σ 2 2n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 22 Γ 22 µ 2 µ 2 2 2nn µ µ 2 Γ Γ2 Σ 2 Γ 22 Γ 2 Γ µ µ 2 2n2 N N 2 N N 2 Γ 22 Σ 2Γ Γ 2 Γ 22 Σ

26 2n2 N N 2 N N 2 Γ 22 Σ 2Γ Γ 2 Γ 22 Σ 2Σ Σ 2 2n 2 µ 2 µ 2 2 Γ 22 Σ 2Γ Γ 2 Γ 22 µ 2 µ 2 2n2 N N 2 N N 2 2 Γ 22 Σ 2Σ Σ 2Γ 22 Γ 2 Γ Σ 2 2n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 2Σ Σ 2Γ 22 µ 2 µ 2 2 2nn N N 2 N N 2 Γ Γ2 Σ Γ Γ 2 Γ 22 Γ 2 Γ Σ 2nn µ µ 2 Γ Γ2 Σ Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 n N N 2 N N 2 n N N 2 N N 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 Σ 22 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 Σ 2Σ Σ 2 n 2 µ 2 µ 2 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 µ 2 µ 2 n2 N N 2 N N 2 Γ Γ 2 Γ 22 Σ 22 Γ 22 Γ 2 Γ Σ n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 22 Γ 22 Γ 2 Γ µ µ 2 2n2 N N 2 N N 2 Γ 22 Σ 2Γ Γ 2 Γ 22 Γ 2 Γ Σ 2 2n 2 µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ 2 Γ 22 µ 2 µ 2 2n2 N N 2 N N 2 Γ Γ 2 Γ 22 Σ 2Γ Γ 2 Γ 22 Σ n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 2Γ Γ 2 Γ 22 µ 2 µ 2 n2 N N 2 N N 2 2 Γ Γ 2 Γ 22 Σ 2Σ Σ 2Γ 22 Γ 2 Γ Σ n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 2Σ Σ 2 Γ 22 Γ 2 Γ µ µ 2 4n2 N N 2 N N 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 Σ 2 2n 2 µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ 26

27 Γ Γ 2 Γ 22 µ µ 2 2n 2 µ µ 2 Γ Γ 2 Γ 22 Σ 2 Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 n2 N N N N 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 Γ 2 Γ Σ n 2 µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 Now we consider separating EV m Γ, Γ 2 again by the number of Ψ 2, ie, EV m Γ, Γ 2 EV m Γ, Γ 2 EV m2 Γ, Γ 2 EV m3 Γ, Γ 2 EV m4 Γ, Γ 2, where V mr r =,, 4 denotes the terms including r Ψ 2 s in V m Since we find that Γ 22 is independent of Γ, Γ 2, Γ2 and we have the conditional expectation of Ψ 2 given Γ in Lemma 32, we can calculate conditional expectation of V mr given Γ, Γ2 for each r By making use of Lemma 22, we can obtain EV m Γ, Γ 2, Γ2 as follows: EV m Γ, Γ 2, Γ2 = 4nn N N 2 N N 2 n p Γ Γ2 Σ Γ Γ 2 Σ 22 Σ 2 2nn n p µ µ 2 Γ Γ2 Σ Γ Γ 2 Σ 22 µ 2 µ 2 2 2nn n p µ µ 2 Γ Γ2 Σ 2 Σ 22 Γ 2 Γ µ µ 2 4n 2 N N 2 N N 2 n pn p 3 4n 2 N N 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Γ 2 Σ 22 Σ 2Σ Σ 2 N N 2 4p 2 n 2 N N 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ 2 27 Σ 22 Σ 2Γ Γ 2

28 4n 2 N N 2 N N 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ 2 Σ 22 Σ 2Σ Σ 2 2n 2 n pn p 3 µ 2 µ 2 2 Σ 22 Σ 2Γ Γ 2 Σ 22 µ 2 µ 2 2n 2 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 2n 2 n pn p 3 µ µ 2 Γ Γ 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 2n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 2n 2 n p n pn p n p 3 µ µ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 By making use of Lemma 32, we can obtain EV m Γ, Γ2 as follows: EV m Γ, Γ2 = 4nn N N 2 N N 2 n p Γ Γ2 Σ 2 Σ 22 Σ 2 2 2nn n p µ µ 2 Γ Γ2 Σ 2 Σ 22 µ 2 µ 2 2 2nn n p µ µ 2 Γ Γ2 Σ 2 Σ 22 Σ 2Σ µ µ 2 4n 2 N N 2 n p N N 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 4n 2 N N 2 N N 2 n pn p 3 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 N N 2 { 4n 2 N N 2 n pn p n p 3 } Σ 22 Σ 2Σ 2 Σ 2 28

29 2n 2 n pn p 3 µ 2 µ 2 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 2n 2 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 2n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 2n 2 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 2n 2 n p n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 By making use of Lemma 22, EV m2 Γ, Γ 2, Γ2 can be calculated as follows: EV m2 Γ, Γ 2, Γ2 = 2nn N N 2 N N 2 n p Γ Γ2 Σ Γ Γ 2 Σ 22 Γ 2 Γ Σ 2nn n p µ µ 2 Γ Γ2 Σ Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 N N 2 2n 2 N N 2 n p n pn p n p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 n 2 N N 2 N N 2 n pn p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 Σ 2Σ Σ 2 N N 2 Σ 22 Σ 2Σ Σ 2 2n 2 N N 2 n pn p n p 3 n 2 n pn p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 µ 2 µ 2 2 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 µ 2 µ

30 n 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 N N 2 2n 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Γ 2 Σ 22 Γ 2 Γ Σ 2 N N 2 4n 2 N N 2 n pn p n p 3 Σ 22 Σ 2Γ Γ 2 2n 2 n p 2 n pn p n p 3 Σ 22 Σ 2Γ Γ 2 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ 2 Σ 22 µ 2 µ 2 2n 2 n pn p n p 3 µ µ 2 Γ Γ 2 Σ 22 Σ 2Γ Γ 2 Σ 22 µ 2 µ 2 4n 2 n pn p n p 3 Σ 22 Σ 2Γ Γ 2 µ µ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 2n 2 N N 2 N N 2 n pn p 3 Γ Γ 2 Σ 22 Σ 2Γ Γ 2 Σ 22 Σ 2 2n 2 n p 2 n pn p n p µ µ 2 Γ Γ 2 Σ 22 Σ 2Γ Γ 2 Σ 22 µ 2 µ 2 2n 2 n pn p n p 3 2 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ 2 Σ 22 µ 2 µ 2 2 n 2 N N 2 N N 2 n pn p 3 Γ Γ 2 Σ 22 Σ 2Σ Σ 2Σ 22 Γ 2 Γ Σ n 2 n pn p 3 µ µ 2 Γ Γ 2 Σ 22 Σ 2 30

31 Σ Σ 2Σ 22 Γ 2 Γ µ µ 2 n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 n 2 n p n pn p n p 3 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 By making use of Lemma 32, EV m2 Γ, Γ2 can be calculated as follows: EV m2 Γ, Γ2 = 2p 2nn N N 2 N N 2 n p Γ Γ2 Σ Γ Σ 2nn N N 2 N N 2 n p Γ Γ2 Σ 2 Σ 22 Σ 2 2p 2nn n p µ µ 2 Γ Γ2 Σ Γ µ µ 2 2nn n p µ µ 2 Γ Γ2 Σ 2 Σ 22 Σ 2Σ µ µ 2 2p 2 n 2 N N 2 n p N N 2 n pn p n p 3 Γ Σ N N 2 2n 2 N N 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 6n 2 N N 2 N N 2 n pn p 3 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 N N 2 2n 2 N N 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 Γ Σ 3

32 N N 2 { 6n 2 N N 2 n pn p n p 3 } Σ 22 Σ 2Σ 2 Σ 2 n 2 n pn p 3 µ 2 µ 2 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 n 2 n p n pn p n p 3 Γ Σ µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 n 2 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 N N 2 2p 2 n 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Σ 2 N N 2 4n 2 N N 2 n pn p n p 3 Σ 22 Σ 2Γ Σ 2 2p 2 n 2 n p 2 n pn p n p 3 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ 2 4n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2 Σ Σ 2Σ 22 µ 2 µ 2 6n 2 2 n pn p n p 3 2 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ 2 4n 2 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 N N 2 2n 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Σ 2 32

33 2n 2 n p 2 n pn p n p 3 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ 2 n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ Σ 2 Σ 22 Σ 2Σ µ µ 2 n 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 2 µ µ 2 Γ µ µ 2 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 p 2 n 2 n p n pn p n p 3 µ µ 2 Γ µ µ 2 n 2 n p n pn p n p 3 Since V m3 does not include Γ 2, EV m3 Γ Lemma 22 as follows: µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2, Γ 2 can be calculated by making use of EV m3 Γ, Γ 2 = 2n 2 N N 2 N N 2 n pn p 3 Γ Γ 2 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 Σ 2 N N 2 Γ 2 Γ 2n 2 N N 2 n pn p n p 3 Σ Γ Γ 2 Σ 22 Σ 22 Σ 2Γ Γ 2 2n 2 n pn p 3 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 µ 2 µ 2 2n 2 2 n pn p n p 3 33

34 Γ 2 Γ Σ Γ Γ 2 Σ 22 µ µ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 2n 2 N N 2 N N 2 n pn p 3 Γ Γ 2 Σ 22 Σ 2Γ Γ 2 Σ 22 Γ 2 Γ Σ 2n 2 N N 2 N N 2 Σ 22 Σ 2Γ Γ 2 n pn p n p 3 Γ Γ 2 Σ 22 Γ 2 Γ Σ 2n 2 n pn p 3 µ µ 2 Γ Γ 2 Σ 22 Σ 2Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 2n 2 n pn p n p 3 Σ 22 Σ 2Γ Γ 2 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 By making use of Lemma 32, EV m3 Γ can be calculated as follows: N N N 2 EV m3 Γ = 4n 2 N 2 n pn p 3 Σ 22 Σ 2Γ Σ 2 N N 2 4p 2 n 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Σ 2 4n 2 N N 2 N N 2 n pn p 3 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 2n 2 n pn p 3 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ 2 2p 2 n 2 n pn p 3 2 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ

35 2n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 4n 2 n pn p n p 3 µ µ 2 Γ Σ 2 Σ 22 µ 2 µ 2 2 2n 2 n p n pn p n p 3 Γ Σ µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 2n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 N N 2 8n 2 N N 2 n pn p n p 3 Σ 22 Σ 2Γ Σ 2 N N 2 4n 2 N N 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 Γ Σ N N 2 { 4n 2 N N 2 n pn p n p 3 } Σ 22 Σ 2Σ 2 Σ 2 2p 2 n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Γ µ µ 2 2n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Γ µ µ 2 2n 2 n pn p 3 2 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ µ µ 2 4n 2 n pn p n p 3 µ µ 2 Γ Σ 2 Σ 22 Σ 2Σ µ µ 2 35

36 2n 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Γ µ µ 2 2n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 Since V m4 also does not include Γ 2, EV m4 Γ Lemma 22 as follows: N N N 2 EV m4 Γ, Γ 2 = n 2 N 2, Γ 2 can be calculated by making use of n pn p 3 Γ Γ 2 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 Γ 2 Γ Σ N N 2 n 2 N N 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 n 2 n pn p 3 µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 n 2 n pn p n p 3 Σ 22 Γ 2 Γ Σ Γ Γ 2 Γ Γ 2 Σ 22 Γ 2 Γ Σ µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 By making use of Lemma 32, EV m4 Γ can be calculated as follows: EV m4 Γ = p 2 2 n2 N N 2 N N 2 n pn p 3 Γ Σ Γ Σ N N 2 p 2 n 2 N N 2 n pn p 3 Γ Σ Γ Σ 2p 2 n 2 N N 2 N N 2 n pn p 3 36

37 Σ 22 Σ 2Γ Σ 2 N N 2 2n 2 N N 2 n pn p 3 Σ 22 Σ 2Γ Σ 2 N N 2 2n 2 N N 2 n pn p 3 Γ Σ Σ 22 Σ 2Σ Σ 2 n 2 N N 2 N N 2 n pn p 3 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 N N 2 2p 2 n 2 N N 2 n pn p n p 3 Γ Σ Γ Σ N N 2 p 2 n 2 N N 2 n p n pn p n p 3 Γ Σ Γ Σ N N 2 Γ Σ N N 2 2p 2 n 2 N N 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 4n 2 N N 2 n pn p n p 3 Σ 22 Σ 2Γ Σ 2 N N 2 { n 2 N N 2 n pn p n p 3 } Σ 22 Σ 2Σ 2 Σ 2 p 2 2 n2 n pn p 3 µ µ 2 Γ Σ Γ µ µ 2 p 2 n 2 n pn p 3 µ µ 2 Γ Σ Γ µ µ 2 2p 2 n 2 n pn p 3 37

38 µ µ 2 Σ Σ 2Σ 22 Σ 2Γ µ µ 2 2n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Γ µ µ 2 n 2 n pn p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ µ µ 2 2p 2 n 2 n pn p n p 3 µ µ 2 Γ Σ Γ µ µ 2 p 2 n 2 n p n pn p n p 3 Γ Σ µ µ 2 Γ µ µ 2 n 2 n p n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Γ µ µ 2 4n 2 n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Γ µ µ 2 n 2 n p n pn p n p 3 Γ Σ µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 Now that EV m Γ, Γ2 = EV m Γ, Γ2 EV m2 Γ, Γ2 EV m3 Γ EV m4 Γ, EV m Γ, Γ2 = p 2 n 2 n p n pn p n p 3 µ µ 2 Γ µ µ 2 2p 2nn n p µ µ 2 Γ Γ2 Σ Γ µ µ 2 p 2 2 n2 n pn p 3 µ µ 2 Γ Σ Γ µ µ 2 p 2 n 2 n pn p 3 µ µ 2 38 Γ Σ Γ µ µ 2

39 2p 2 n 2 n pn p n p 3 µ µ 2 Γ Σ Γ µ µ 2 p 2 n 2 n p n pn p n p 3 Γ Σ µ µ 2 Γ Σ Γ µ µ 2 n 2 n p n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 n 2 n p n pn p n p 3 Γ Σ µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 2n 2 n p n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 2nn n p µ µ 2 Γ Γ2 Σ 2 Γ 22 µ 2 µ 2 2 2n 2 n p n pn p n p 3 Γ Σ µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 2 n 2 n p n pn p n p 3 Γ Σ µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 n 2 n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 n 2 n pn p 3 µ 2 µ 2 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 2p 2 n 2 N N 2 n p N N 2 n pn p n p 3 p 2 2 n2 N N 2 N N 2 n pn p 3 p 2 n 2 N N 2 N N 2 n pn p 3 2 Γ Σ Γ Σ Γ Σ Γ Σ Γ Σ 39

40 2p 2 n 2 N N 2 N N 2 n pn p n p 3 2p 2nn N N 2 N N 2 n p Γ Γ2 Σ Γ Σ N N 2 p 2 n 2 N N 2 n p n pn p n p 3 Γ Σ Γ Σ 2n 2 N N 2 n p N N 2 n pn p n p 3 Γ Σ Γ Σ Σ 22 Σ 2Σ Σ 2 2nn N N 2 N N 2 n p Σ 22 Σ 2Γ Γ2 Σ 2 n 2 N N 2 N N 2 n pn p n p 3 n 2 N N 2 N N 2 n pn p 3 { } Σ 22 Σ 2Σ 2 Σ 2 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 Finally EV m can be derived by making use of Lemma 22 and Lemma 33 as follows: EV m = p 2 n 2 n pn p n p 3 µ µ 2 Σ µ µ 2 2p 2 nn n n p n p 3n p n p µ µ 2 p 2 p 2 n 2 n n pn p 3n p n p n p 3 Σ µ µ µ 2 Σ µ µ 2 2p 2 n 2 n n pn p n p 3n p n p n p 3 µ µ 2 Σ µ µ 2 p p 2 n 2 n p 2 n pn p n p 3n p n p 3 µ µ 2 Σ µ µ 2 2p 2 n 2 n pn p n p 3n p n p 3 µ µ 2 Σ µ µ 2 n 2 n n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 Σ 2Σ µ µ 2 2n 2 n n pn p n p 3 µ µ 2 Σ Σ 2Σ 22 µ µ 2 2 µ

41 2nn n p n p µ µ 2 p n 2 n pn p n p 3 µ 2 µ 2 n 2 Σ Σ 2Σ 22 µ 2 Σ 22 µ n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 n 2 n pn p 3 µ 2 µ 2 2p p 2 n 2 N N 2 N N 2 n pn p n p 3 2 µ 2 2 µ 2 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ µ 2 p p 2 p 2 n 2 n N N 2 N N 2 n pn p 3n p n p n p 3 2p p 2 n 2 n N N 2 N N 2 n pn p n p 3n p n p n p 3 2p p 2 nn n N N 2 N N 2 n p n p 3n p n p p 2 p 2n 2 N N 2 n p 2 N N 2 n pn p n p 3n p n p 3 2p p 2 n 2 N N 2 N N 2 n pn p n p 3n p n p 3 2n 2 N N 2 n p N N 2 n pn p n p 3 2nn N N 2 N N 2 n p n p n 2 N N 2 N N 2 n pn p n p 3 n 2 N N 2 N N 2 n pn p 3 On the other hand, EV m2 Γ, Γ 2 is calculated as follows: Σ 22 Σ 2Σ Σ 2 Σ 22 Σ 2Σ Σ 2 2 { } Σ 22 Σ 2Σ 2 Σ 2 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2 EV m2 Γ, Γ 2 = n2 N N 2 N N 2 Γ Γ2 Σ Γ Γ2 Σ n 2 µ µ 2 Γ Γ2 Σ Γ Γ2 µ µ 2 2nn N N 2 N N 2 4 Γ 22 Σ 2Γ Γ2 Σ 2 C3

42 2nn µ µ 2 Γ Γ2 Σ 2 Γ 22 µ 2 µ 2 n2 N N 2 N N 2 n2 N N 2 N N 2 2 Γ 22 Σ 22 Γ 22 Σ 22 Γ 22 Σ 22 Γ 22 Σ 2Σ Σ 2 n 2 µ 2 µ 2 2 Γ 22 Σ 22 Γ 22 µ 2 µ 2 n2 N N 2 N N 2 n2 N N 2 N N 2 2 Γ 22 Σ 2Σ Σ 2Γ 22 Σ 22 Γ 22 Σ 2Σ Σ 2Γ 22 Σ 2Σ Σ 2 n 2 µ 2 µ 2 2 Γ 22 Σ 2Σ Σ 2Γ 22 µ 2 µ 2 2 By making use of Lemma 22 and noting that Γ Γ2 and Γ 22 are independent, EV m2 can be calculated easily as follows: EV m2 = n 2 n n p n p n p 3 µ µ 2 Σ µ µ 2 2nn n p n p µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 n 2 n p n pn p n p 3 µ 2 µ 2 n 2 2 Σ 22 µ n pn p n p 3 Σ 22 Σ 2Σ Σ 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 n 2 n pn p 3 µ 2 µ 2 2 Σ 22 Σ 2Σ Σ 2Σ 22 µ 2 µ 2 2n 2 N N 2 n p N N 2 n pn p n p 3 2nn N N 2 N N 2 n p n p n 2 N N 2 N N 2 n pn p n p 3 n 2 N N 2 N N 2 n pn p 3 p n 2 n N N 2 N N 2 n p n p n p µ 2 2 Σ 22 Σ 2Σ Σ 2 Σ 22 Σ 2Σ Σ 2 { } Σ 22 Σ 2Σ 2 Σ 2 Σ 22 Σ 2Σ Σ 2Σ 22 Σ 2Σ Σ 2

43 p 2 n 2 n p N N 2 N N 2 n pn p n p 3 C4 Now we have obtained EV m = EV m EV m2 by C3 and C4 Thus Theorem 34 has been completed Appendix D Proof of Theorem 36 First we derive the unbiased estimator of δ 2 Since d 2 m = µ µ 2 Σ µ µ 2 we find the expectation of it given Γ, Γ 2 : = nx T x2 T Γ Γ2 x T x2 T, E d 2 m Γ, Γ 2 = nn N 2 N N 2 Γ Γ2 Σ nµ µ 2 Γ Γ2 µ µ 2 Thus we can obtain the following expectation as follows by Lemma 22: E d 2 m = n δ 2 p N N 2 n p N N 2 Therefore we derive the following statistic as the unbiased estimator of δ 2 : δ 2 = n p d 2 m p N N 2 n N N 2 D Next we derive the unbiased estimator of 2 Since Dm 2 = µ µ 2 Σ µ µ 2 µ µ 2 Σ Σ 2 Σ Σ 22 2 Σ µ µ 2 2 µ µ 2 Σ Σ 2 Σ 22 µ 2 µ 2 2 µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 = nx T x2 T Γ Γ2 x T x2 T n x 2F x2 2F Γ 22 x 2F x2 2F 2n x F x2 F Γ Γ 2 Γ 22 x 2F x2 2F n x F x2 F Γ Γ 2 Γ 22 Γ 2 Γ x F x2 F, 43

44 the conditional expectation of D 2 m given Γ, Γ 2 can be obtained: E D m Γ 2, Γ 2 = nn N 2 N N 2 Γ Γ2 Σ nµ µ 2 Γ Γ2 µ µ 2 n N N 2 N N 2 n N N 2 N N 2 Γ 22 Σ 22 Γ 22 Σ 2Σ Σ 2 n µ 2 µ 2 2 Γ 22 µ 2 µ 2 2n N N 2 N N 2 2 Γ 22 Σ 2Γ Γ 2 2n µ µ 2 Γ Γ 2 Γ 22 µ 2 µ 2 n N N 2 N N 2 2 Γ 22 Γ 2 Γ Σ Γ Γ 2 n µ µ 2 Γ Γ 2 Γ 22 Γ 2 Γ µ µ 2 Therefore we find the conditional expectation of D 2 m given Γ, Γ 2, Γ2 by Lemma 22 as follows: E Dm Γ 2, Γ 2, Γ2 = nn N 2 N N 2 Γ Γ2 Σ nµ µ 2 Γ Γ2 µ µ 2 p 2n N N 2 N N 2 n p n N N 2 N N 2 n p n n p µ 2 µ 2 2n N N 2 N N 2 2n n p µ µ 2 n p Σ 22 Σ 2Σ Σ 2 2 Σ 22 µ 2 µ 2 2 Σ 22 Σ 2Γ Γ 2 Γ Γ 2 Σ 22 µ 2 µ 2 2 n N N 2 N N 2 n p Σ 22 Γ 2 Γ Σ Γ Γ 2 n n p µ µ 2 Γ Γ 2 Σ 22 Γ 2 Γ µ µ 2 44

45 Moreover we can calculate the conditional expectation of D 2 m given Γ, Γ2 by Lemma 32 as follows: E Dm Γ 2, Γ2 = nn N 2 N N 2 Γ Γ2 Σ nµ µ 2 Γ Γ2 µ µ 2 p 2n N N 2 N N 2 n p n n p µ 2 µ 2 2 Σ 22 µ 2 µ 2 2 2n n p µ µ 2 Σ Σ 2Σ 22 µ 2 µ 2 p 2n N N 2 N N 2 n p p 2n n p µ µ 2 n n p µ µ 2 Γ Σ Γ µ µ 2 2 Σ Σ 2Σ 22 Σ 2Σ µ By using the result of Lemma 22 again, we can obtain ED 2 m as follows: ED 2 m = n n p 2 n n p n n p δ 2 µ 2 p nn N 2 N N 2 n p p 2n N N 2 N N 2 n p p p 2 n N N 2 N N 2 n p n p D2 By the results of D and D2, Theorem 36 has been completed 45

46 Table The Comparisons of the Accuracy of Asymptotic Approximations for EPMC where p = 2 p = 2 p =, p 2 = N, N 2, N 2, N 2 2 Lachenbruch/Shutoh Fujikoshi and Seo Simulation 0, 0, 0, , 0, 20, , 0, 20, , 0, 40, , 20, 0, , 20, 20, , 20, 20, , 20, 40,

47 Table Continued p = 2 p =, p 2 = N, N 2, N 2, N 2 2 Lachenbruch/Shutoh Fujikoshi and Seo Simulation 0, 30, 0, , 30, 20, , 30, 20, , 30, 40, , 0, 0, , 0, 20, , 0, 20, , 0, 40,

48 Table Continued p = 2 p =, p 2 = N, N 2, N 2, N 2 2 Lachenbruch/Shutoh Fujikoshi and Seo Simulation 20, 20, 0, , 20, 20, , 20, 20, , 20, 40, , 40, 0, , 40, 20, , 40, 20, , 40, 40,

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work

Approximate interval estimation for EPMC for improved linear discriminant rule under high dimensional frame work Hiroshima Statistical Research Group: Technical Report Approximate interval estimation for PMC for improved linear discriminant rule under high dimensional frame work Masashi Hyodo, Tomohiro Mitani, Tetsuto

More information

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large

EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large EPMC Estimation in Discriminant Analysis when the Dimension and Sample Sizes are Large Tetsuji Tonda 1 Tomoyuki Nakagawa and Hirofumi Wakaki Last modified: March 30 016 1 Faculty of Management and Information

More information

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA

SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA SIMULTANEOUS CONFIDENCE INTERVALS AMONG k MEAN VECTORS IN REPEATED MEASURES WITH MISSING DATA Kazuyuki Koizumi Department of Mathematics, Graduate School of Science Tokyo University of Science 1-3, Kagurazaka,

More information

T 2 Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem

T 2 Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem T Type Test Statistic and Simultaneous Confidence Intervals for Sub-mean Vectors in k-sample Problem Toshiki aito a, Tamae Kawasaki b and Takashi Seo b a Department of Applied Mathematics, Graduate School

More information

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control

On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control On the conservative multivariate multiple comparison procedure of correlated mean vectors with a control Takahiro Nishiyama a a Department of Mathematical Information Science, Tokyo University of Science,

More information

High-dimensional asymptotic expansions for the distributions of canonical correlations

High-dimensional asymptotic expansions for the distributions of canonical correlations Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic

More information

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors

On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors On the conservative multivariate Tukey-Kramer type procedures for multiple comparisons among mean vectors Takashi Seo a, Takahiro Nishiyama b a Department of Mathematical Information Science, Tokyo University

More information

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi

High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis. Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi High-Dimensional AICs for Selection of Redundancy Models in Discriminant Analysis Tetsuro Sakurai, Takeshi Nakada and Yasunori Fujikoshi Faculty of Science and Engineering, Chuo University, Kasuga, Bunkyo-ku,

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

Estimation of the large covariance matrix with two-step monotone missing data

Estimation of the large covariance matrix with two-step monotone missing data Estimation of the large covariance matrix with two-ste monotone missing data Masashi Hyodo, Nobumichi Shutoh 2, Takashi Seo, and Tatjana Pavlenko 3 Deartment of Mathematical Information Science, Tokyo

More information

Testing equality of two mean vectors with unequal sample sizes for populations with correlation

Testing equality of two mean vectors with unequal sample sizes for populations with correlation Testing equality of two mean vectors with unequal sample sizes for populations with correlation Aya Shinozaki Naoya Okamoto 2 and Takashi Seo Department of Mathematical Information Science Tokyo University

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis

Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Consistency of Distance-based Criterion for Selection of Variables in High-dimensional Two-Group Discriminant Analysis Tetsuro Sakurai and Yasunori Fujikoshi Center of General Education, Tokyo University

More information

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices

Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Testing linear hypotheses of mean vectors for high-dimension data with unequal covariance matrices Takahiro Nishiyama a,, Masashi Hyodo b, Takashi Seo a, Tatjana Pavlenko c a Department of Mathematical

More information

Estimation in an Exponentiated Half Logistic Distribution under Progressively Type-II Censoring

Estimation in an Exponentiated Half Logistic Distribution under Progressively Type-II Censoring Communications of the Korean Statistical Society 2011, Vol. 18, No. 5, 657 666 DOI: http://dx.doi.org/10.5351/ckss.2011.18.5.657 Estimation in an Exponentiated Half Logistic Distribution under Progressively

More information

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems

More information

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data Yujun Wu, Marc G. Genton, 1 and Leonard A. Stefanski 2 Department of Biostatistics, School of Public Health, University of Medicine

More information

An Unbiased C p Criterion for Multivariate Ridge Regression

An Unbiased C p Criterion for Multivariate Ridge Regression An Unbiased C p Criterion for Multivariate Ridge Regression (Last Modified: March 7, 2008) Hirokazu Yanagihara 1 and Kenichi Satoh 2 1 Department of Mathematics, Graduate School of Science, Hiroshima University

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

An Introduction to Multivariate Statistical Analysis

An Introduction to Multivariate Statistical Analysis An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents

More information

Inferences about Parameters of Trivariate Normal Distribution with Missing Data

Inferences about Parameters of Trivariate Normal Distribution with Missing Data Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 7-5-3 Inferences about Parameters of Trivariate Normal Distribution with Missing

More information

Estimation of a multivariate normal covariance matrix with staircase pattern data

Estimation of a multivariate normal covariance matrix with staircase pattern data AISM (2007) 59: 211 233 DOI 101007/s10463-006-0044-x Xiaoqian Sun Dongchu Sun Estimation of a multivariate normal covariance matrix with staircase pattern data Received: 20 January 2005 / Revised: 1 November

More information

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring A. Ganguly, S. Mitra, D. Samanta, D. Kundu,2 Abstract Epstein [9] introduced the Type-I hybrid censoring scheme

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Two sample T 2 test 1 Two sample T 2 test 2 Analogous to the univariate context, we

More information

Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data

Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data Applied Mathematical Sciences, Vol 3, 009, no 54, 695-70 Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data Evelina Veleva Rousse University A Kanchev Department of Numerical

More information

CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL

CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL J. Japan Statist. Soc. Vol. 37 No. 1 2007 135 155 CONDITIONS FOR ROBUSTNESS TO NONNORMALITY ON TEST STATISTICS IN A GMANOVA MODEL Hirokazu Yanagihara* This paper presents the conditions for robustness

More information

Estimation for Mean and Standard Deviation of Normal Distribution under Type II Censoring

Estimation for Mean and Standard Deviation of Normal Distribution under Type II Censoring Communications for Statistical Applications and Methods 2014, Vol. 21, No. 6, 529 538 DOI: http://dx.doi.org/10.5351/csam.2014.21.6.529 Print ISSN 2287-7843 / Online ISSN 2383-4757 Estimation for Mean

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

Supplement to the paper Accurate distributions of Mallows C p and its unbiased modifications with applications to shrinkage estimation

Supplement to the paper Accurate distributions of Mallows C p and its unbiased modifications with applications to shrinkage estimation To aear in Economic Review (Otaru University of Commerce Vol. No..- 017. Sulement to the aer Accurate distributions of Mallows C its unbiased modifications with alications to shrinkage estimation Haruhiko

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order

More information

Testing Block-Diagonal Covariance Structure for High-Dimensional Data

Testing Block-Diagonal Covariance Structure for High-Dimensional Data Testing Block-Diagonal Covariance Structure for High-Dimensional Data MASASHI HYODO 1, NOBUMICHI SHUTOH 2, TAKAHIRO NISHIYAMA 3, AND TATJANA PAVLENKO 4 1 Department of Mathematical Sciences, Graduate School

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Estimation for generalized half logistic distribution based on records

Estimation for generalized half logistic distribution based on records Journal of the Korean Data & Information Science Society 202, 236, 249 257 http://dx.doi.org/0.7465/jkdi.202.23.6.249 한국데이터정보과학회지 Estimation for generalized half logistic distribution based on records

More information

Department of Statistics

Department of Statistics Research Report Department of Statistics Research Report Department of Statistics No. 05: Testing in multivariate normal models with block circular covariance structures Yuli Liang Dietrich von Rosen Tatjana

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Hypothesis testing in multilevel models with block circular covariance structures

Hypothesis testing in multilevel models with block circular covariance structures 1/ 25 Hypothesis testing in multilevel models with block circular covariance structures Yuli Liang 1, Dietrich von Rosen 2,3 and Tatjana von Rosen 1 1 Department of Statistics, Stockholm University 2 Department

More information

Approximate Distributions of the Likelihood Ratio Statistic in a Structural Equation with Many Instruments

Approximate Distributions of the Likelihood Ratio Statistic in a Structural Equation with Many Instruments CIRJE-F-466 Approximate Distributions of the Likelihood Ratio Statistic in a Structural Equation with Many Instruments Yukitoshi Matsushita CIRJE, Faculty of Economics, University of Tokyo February 2007

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45 Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved

More information

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION

A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION J. Japan Statist. Soc. Vol. 32 No. 2 2002 165 181 A NOTE ON ESTIMATION UNDER THE QUADRATIC LOSS IN MULTIVARIATE CALIBRATION Hisayuki Tsukuma* The problem of estimation in multivariate linear calibration

More information

CURRICULUM VITAE. Yoshihiko KONNO. August 31, Personal. Japan Women s University, Mejirodai, Bunkyo-ku, Tokyo , Japan

CURRICULUM VITAE. Yoshihiko KONNO. August 31, Personal. Japan Women s University, Mejirodai, Bunkyo-ku, Tokyo , Japan CURRICULUM VITAE Yoshihiko KONNO August 31, 2017 Personal Date/Place of Birth Citizenship Mailing address Email address Weg-page September 9, 1961, Japan Japan Japan Women s University, 2-8-1 Mejirodai,

More information

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction Acta Math. Univ. Comenianae Vol. LXV, 1(1996), pp. 129 139 129 ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES V. WITKOVSKÝ Abstract. Estimation of the autoregressive

More information

Properties and approximations of some matrix variate probability density functions

Properties and approximations of some matrix variate probability density functions Technical report from Automatic Control at Linköpings universitet Properties and approximations of some matrix variate probability density functions Karl Granström, Umut Orguner Division of Automatic Control

More information

Optimum Sampling Vectors for Wiener Filter Noise Reduction

Optimum Sampling Vectors for Wiener Filter Noise Reduction 58 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 1, JANUARY 2002 Optimum Sampling Vectors for Wiener Filter Noise Reduction Yukihiko Yamashita, Member, IEEE Absact Sampling is a very important and

More information

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX

HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX Submitted to the Annals of Statistics HYPOTHESIS TESTING ON LINEAR STRUCTURES OF HIGH DIMENSIONAL COVARIANCE MATRIX By Shurong Zheng, Zhao Chen, Hengjian Cui and Runze Li Northeast Normal University, Fudan

More information

Likelihood Ratio Tests in Multivariate Linear Model

Likelihood Ratio Tests in Multivariate Linear Model Likelihood Ratio Tests in Multivariate Linear Model Yasunori Fujikoshi Department of Mathematics, Graduate School of Science, Hiroshima University, 1-3-1 Kagamiyama, Higashi Hiroshima, Hiroshima 739-8626,

More information

Estimation of parametric functions in Downton s bivariate exponential distribution

Estimation of parametric functions in Downton s bivariate exponential distribution Estimation of parametric functions in Downton s bivariate exponential distribution George Iliopoulos Department of Mathematics University of the Aegean 83200 Karlovasi, Samos, Greece e-mail: geh@aegean.gr

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS

SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS SIMULTANEOUS ESTIMATION OF SCALE MATRICES IN TWO-SAMPLE PROBLEM UNDER ELLIPTICALLY CONTOURED DISTRIBUTIONS Hisayuki Tsukuma and Yoshihiko Konno Abstract Two-sample problems of estimating p p scale matrices

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information

Linear Models for Multivariate Repeated Measures Data

Linear Models for Multivariate Repeated Measures Data THE UNIVERSITY OF TEXAS AT SAN ANTONIO, COLLEGE OF BUSINESS Working Paper SERIES Date December 3, 200 WP # 007MSS-253-200 Linear Models for Multivariate Repeated Measures Data Anuradha Roy Management Science

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample

More information

Advanced Multivariate Statistics with Matrices

Advanced Multivariate Statistics with Matrices Advanced Multivariate Statistics with Matrices Mathematics and Its Applications Managing Editor: M. HAZEWINKEL Centre for Mathematics and Computer Science, Amsterdam, The Netherlands Volume 579 Advanced

More information

Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification

Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification Jackknife Bias Correction of the AIC for Selecting Variables in Canonical Correlation Analysis under Model Misspecification (Last Modified: May 22, 2012) Yusuke Hashiyama, Hirokazu Yanagihara 1 and Yasunori

More information

Gaussian Models (9/9/13)

Gaussian Models (9/9/13) STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate

More information

A note on profile likelihood for exponential tilt mixture models

A note on profile likelihood for exponential tilt mixture models Biometrika (2009), 96, 1,pp. 229 236 C 2009 Biometrika Trust Printed in Great Britain doi: 10.1093/biomet/asn059 Advance Access publication 22 January 2009 A note on profile likelihood for exponential

More information

Approximate and Fiducial Confidence Intervals for the Difference Between Two Binomial Proportions

Approximate and Fiducial Confidence Intervals for the Difference Between Two Binomial Proportions Approximate and Fiducial Confidence Intervals for the Difference Between Two Binomial Proportions K. Krishnamoorthy 1 and Dan Zhang University of Louisiana at Lafayette, Lafayette, LA 70504, USA SUMMARY

More information

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12 Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following

More information

An Analysis of Record Statistics based on an Exponentiated Gumbel Model

An Analysis of Record Statistics based on an Exponentiated Gumbel Model Communications for Statistical Applications and Methods 2013, Vol. 20, No. 5, 405 416 DOI: http://dx.doi.org/10.5351/csam.2013.20.5.405 An Analysis of Record Statistics based on an Exponentiated Gumbel

More information

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:

More information

Bias-corrected AIC for selecting variables in Poisson regression models

Bias-corrected AIC for selecting variables in Poisson regression models Bias-corrected AIC for selecting variables in Poisson regression models Ken-ichi Kamo (a), Hirokazu Yanagihara (b) and Kenichi Satoh (c) (a) Corresponding author: Department of Liberal Arts and Sciences,

More information

Selection of Model Selection Criteria for Multivariate Ridge Regression

Selection of Model Selection Criteria for Multivariate Ridge Regression Selection of Model Selection Criteria for Multivariate Ridge Regression (Last Modified: January, 24) Isamu Nagai, Department of Mathematics, Graduate School of Science, Hiroshima University -3- Kagamiyama,

More information

Asymptotic Analysis of the Generalized Coherence Estimate

Asymptotic Analysis of the Generalized Coherence Estimate IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 1, JANUARY 2001 45 Asymptotic Analysis of the Generalized Coherence Estimate Axel Clausen, Member, IEEE, and Douglas Cochran, Senior Member, IEEE Abstract

More information

Input-Dependent Estimation of Generalization Error under Covariate Shift

Input-Dependent Estimation of Generalization Error under Covariate Shift Statistics & Decisions, vol.23, no.4, pp.249 279, 25. 1 Input-Dependent Estimation of Generalization Error under Covariate Shift Masashi Sugiyama (sugi@cs.titech.ac.jp) Department of Computer Science,

More information

THE STATISTICAL ANALYSIS OF MONOTONE INCOMPLETE MULTIVARIATE NORMAL DATA

THE STATISTICAL ANALYSIS OF MONOTONE INCOMPLETE MULTIVARIATE NORMAL DATA The Pennsylvania State University The Graduate School Department of Statistics THE STATISTICAL ANALYSIS OF MONOTONE INCOMPLETE MULTIVARIATE NORMAL DATA A Dissertation in Statistics by Megan M. Romer c

More information

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE J Jaan Statist Soc Vol 34 No 2004 9 26 ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE Yasunori Fujikoshi*, Tetsuto Himeno

More information

The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1

The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1 Applied Mathematical Sciences, Vol. 2, 28, no. 48, 2377-2391 The Distributions of Sums, Products and Ratios of Inverted Bivariate Beta Distribution 1 A. S. Al-Ruzaiza and Awad El-Gohary 2 Department of

More information

Canonical Correlation Analysis of Longitudinal Data

Canonical Correlation Analysis of Longitudinal Data Biometrics Section JSM 2008 Canonical Correlation Analysis of Longitudinal Data Jayesh Srivastava Dayanand N Naik Abstract Studying the relationship between two sets of variables is an important multivariate

More information

Robustness of the Quadratic Discriminant Function to correlated and uncorrelated normal training samples

Robustness of the Quadratic Discriminant Function to correlated and uncorrelated normal training samples DOI 10.1186/s40064-016-1718-3 RESEARCH Open Access Robustness of the Quadratic Discriminant Function to correlated and uncorrelated normal training samples Atinuke Adebanji 1,2, Michael Asamoah Boaheng

More information

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)

More information

Web Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models

Web Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models Web Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models Diaa Noureldin Department of Economics, University of Oxford, & Oxford-Man Institute, Eagle House, Walton Well Road, Oxford OX

More information

Analysis of variance, multivariate (MANOVA)

Analysis of variance, multivariate (MANOVA) Analysis of variance, multivariate (MANOVA) Abstract: A designed experiment is set up in which the system studied is under the control of an investigator. The individuals, the treatments, the variables

More information

A note on multiple imputation for general purpose estimation

A note on multiple imputation for general purpose estimation A note on multiple imputation for general purpose estimation Shu Yang Jae Kwang Kim SSC meeting June 16, 2015 Shu Yang, Jae Kwang Kim Multiple Imputation June 16, 2015 1 / 32 Introduction Basic Setup Assume

More information

Chapter 4: Asymptotic Properties of the MLE (Part 2)

Chapter 4: Asymptotic Properties of the MLE (Part 2) Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response

More information

THE EFFECT OF SAMPLING DESIGN ON ANDERSON S EXPANSION OF THE DISTRIBUTION OF FISHER S SAMPLE DISCRIMINANT FUNCTION

THE EFFECT OF SAMPLING DESIGN ON ANDERSON S EXPANSION OF THE DISTRIBUTION OF FISHER S SAMPLE DISCRIMINANT FUNCTION Statistica Sinica 8(1998), 1115-1130 THE EFFECT OF SAMPLING ESIGN ON ANERSON S EXPANSION OF THE ISTRIBUTION OF FISHER S SAMPLE ISCRIMINANT FUNCTION Kam-Wah Tsui and Ching-Ho Leu University of Wisconsin-Madison

More information

A Note on Bootstraps and Robustness. Tony Lancaster, Brown University, December 2003.

A Note on Bootstraps and Robustness. Tony Lancaster, Brown University, December 2003. A Note on Bootstraps and Robustness Tony Lancaster, Brown University, December 2003. In this note we consider several versions of the bootstrap and argue that it is helpful in explaining and thinking about

More information

Closed-form sampling formulas for the coalescent with recombination

Closed-form sampling formulas for the coalescent with recombination 0 / 21 Closed-form sampling formulas for the coalescent with recombination Yun S. Song CS Division and Department of Statistics University of California, Berkeley September 7, 2009 Joint work with Paul

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA

MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA J. Japan Statist. Soc. Vol. 37 No. 1 2007 53 86 MULTIVARIATE THEORY FOR ANALYZING HIGH DIMENSIONAL DATA M. S. Srivastava* In this article, we develop a multivariate theory for analyzing multivariate datasets

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

c 2005 Society for Industrial and Applied Mathematics

c 2005 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. XX, No. X, pp. XX XX c 005 Society for Industrial and Applied Mathematics DISTRIBUTIONS OF THE EXTREME EIGENVALUES OF THE COMPLEX JACOBI RANDOM MATRIX ENSEMBLE PLAMEN KOEV

More information

KWAME NKRUMAH UNIVERSITY OF SCIENCE AND TECHNOLOGY

KWAME NKRUMAH UNIVERSITY OF SCIENCE AND TECHNOLOGY KWAME NKRUMAH UNIVERSITY OF SCIENCE AND TECHNOLOGY Performance Evaluation of Classification Methods: the case of Equal Mean Discrimination. By Michael Asamoah-Boaheng (Bsc. Statistics) A THESIS SUBMITTED

More information

A Study on the Correlation of Bivariate And Trivariate Normal Models

A Study on the Correlation of Bivariate And Trivariate Normal Models Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 11-1-2013 A Study on the Correlation of Bivariate And Trivariate Normal Models Maria

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

The Bayes classifier

The Bayes classifier The Bayes classifier Consider where is a random vector in is a random variable (depending on ) Let be a classifier with probability of error/risk given by The Bayes classifier (denoted ) is the optimal

More information

Two Different Shrinkage Estimator Classes for the Shape Parameter of Classical Pareto Distribution

Two Different Shrinkage Estimator Classes for the Shape Parameter of Classical Pareto Distribution Two Different Shrinkage Estimator Classes for the Shape Parameter of Classical Pareto Distribution Meral EBEGIL and Senay OZDEMIR Abstract In this study, biased estimators for the shape parameter of a

More information

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and Athens Journal of Sciences December 2014 Discriminant Analysis with High Dimensional von Mises - Fisher Distributions By Mario Romanazzi This paper extends previous work in discriminant analysis with von

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017 Department of Statistical Science Duke University FIRST YEAR EXAM - SPRING 017 Monday May 8th 017, 9:00 AM 1:00 PM NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam;

More information

Likelihood Ratio Criterion for Testing Sphericity from a Multivariate Normal Sample with 2-step Monotone Missing Data Pattern

Likelihood Ratio Criterion for Testing Sphericity from a Multivariate Normal Sample with 2-step Monotone Missing Data Pattern The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 473-481 Likelihood Ratio Criterion for Testing Sphericity from a Multivariate Normal Sample with 2-step Monotone Missing Data Pattern Byungjin

More information

On Properties of QIC in Generalized. Estimating Equations. Shinpei Imori

On Properties of QIC in Generalized. Estimating Equations. Shinpei Imori On Properties of QIC in Generalized Estimating Equations Shinpei Imori Graduate School of Engineering Science, Osaka University 1-3 Machikaneyama-cho, Toyonaka, Osaka 560-8531, Japan E-mail: imori.stat@gmail.com

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a

More information

On testing the equality of mean vectors in high dimension

On testing the equality of mean vectors in high dimension ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.

More information

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Takeshi Emura and Hisayuki Tsukuma Abstract For testing the regression parameter in multivariate

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

Notes on Random Vectors and Multivariate Normal

Notes on Random Vectors and Multivariate Normal MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution

More information