KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS

Size: px
Start display at page:

Download "KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS"

Transcription

1 Bull. Korean Math. Soc. 5 (24), No. 3, pp KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Yicheng Hong and Sungchul Lee Abstract. The limiting distribution for the linear placement statistics under the null hypotheses has been provided by Orban and Wolfe [9] and Kim [5] when one of the sample sizes goes to infinity, and by Kim, Lee and Wang [6] when the sample sizes of each group go to infinity simultaneously. In this paper we establish the generalized Kruskal-Wallis one-way analysis of variance for the linear placement statistics.. Introduction and statement of main results To determine whether the mutually independent random samples come from a common underlying distribution, we use various kinds of two-sample tests. Two-sample tests that are valid with no assumptions (except the continuity of the underlying distribution), are called the distribution-free procedures. Some distribution-free procedures use the sample ranks; the rank-sum test of Wilcoxon [] and the median test of Mood [8]. Other distribution-free tests utilize the sample placements; the linear placement test of Orban and Wolfe [9] and the fixed and updated linear placement test of Kim [5]. Asymptotic theories for the sample rank tests are well developed; see Chernoff and Savage [], Govindarajulu, Le Cam and Raghavachari [3], Hajek [4], Pyke and Shorak [], and Dupac and Hajek [2]. However, asymptotic theories for the sample linear placement tests are quite limited. Orban and Wolfe [9] provided the limiting theory when one of the sample sizes goes to infinity. Kim and Wolfe [7] investigated the iterative asymptotic distribution when the control group sample size goes to infinity then the treatment group sample size goes to infinity. Similarly, Kim [5] established the limiting theory of the fixed and updated linear placement test when the control group sample size goes to Received March, Mathematics Subject Classification. Primary 6F5, 62E2, 62G2; Secondary 6E5, 6E. Key words and phrases. Kruskal-Wallis one-way analysis of variance, central limit theorem, linear placement statistic. This work was supported by the BK2 project of the Department of Mathematics, Yonsei University and Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education, Science and Technology(2-99). 7 c 24 Korean Mathematical Society

2 72 Y. HONG AND S. LEE infinity. Moreover, Kim, Lee, and Wang [6] further established the central limit theorem when the sample sizes of each group go to infinity simultaneously. In this paper we develop the Kruskal-Wallis one-way analysis of variance based on linear placements. The usual Kruskal-Wallis one-way analysis of variance is based on ranks but our Kruskal-Wallis one-way analysis of variance is based on linear placements, which can be used for comparing more than two samples that are independent, or not related. Our main results are follows. First we fix our notations. The data set consists of k groups and the group i has n i observations X ij, j n i. Overall we have N = k n i observations. We assume that X ij, i k, j n i, are mutually independent samples from the continuous cumulative distribution functions G i (x) = F(x θ i ). For each X ij let V ij be the rank of X ij among all observations excluding observations from the group that X ij belongs, and let W ij be its normalized rank: V ij = n s (.) (X st X ij ), s i t= V ij (.2) W ij =. N n i We further let V i. be the sample mean rank of group i observations and V.. be the sample mean rank of all observations: = n i (.3) V ij, n i (.4) V i. V.. = N j= n i V ij. j= When the sample sizes of all groups go to infinity simultaneously, we have the following Kruskal-Wallis one-way analysis of variance based on linear placements. Theorem. If θ = = θ k and if for each i, as N n i N λ i, < λ i <, then as N (.5) H := 2 N(N +) n i (V i. EV i. ) 2 χ 2 k, where by (2.7), EV i. = (N n i )/2. In Figure we create 2 samples of H using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure (a) we consider the equal group sizes and in Figure (b) we consider the unequal group sizes. In

3 KRUSKAL-WALLIS ONE-WAY ANALYSIS 73 empirical c.d.f. of H c.d.f. of χ 2 k H values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of H c.d.f. of χ 2 k H values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure. Comparison of the empirical distribution of H and the Chi-square distribution. We create 2 samples of H using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In both cases we see that the distribution of H is very close to the Chi-square distribution. both cases we see that the distribution of the Kruskal-Wallis test statistic H based on linear placements is very close to the Chi-square distribution as we present in Theorem. If we have the equal group sizes, we replace EV i. in H with V.. and we still get the Chi-square limit. Theorem 2. If θ = = θ k and if for each i, as N (.6) n i = N k +f i(n), f i (N) = O(N βi ), β i < 2, then as N (.7) Ĥ := 2 N(N +) n i (V i. V.. ) 2 χ 2 k. However, if we don t have the same group sizes, we cannot replace EV i. in H with V.. to get the Chi-square limit. Proposition. If θ = = θ k, if for each i, as N (.8) n i = N k +f i(n), f i (N) = O(N βi ), β i <, and if for some i, as N (.9) n i = N k +f i(n), f i (N) = Ω(N βi ), then as N 2 < β i <, E(H Ĥ) and Var(H Ĥ).

4 74 Y. HONG AND S. LEE Remark. If in Proposition, instead of (.8) and (.9) we assume for each i, as N (.) n i = N k +f i(n), f i (N) = O(N βi ), β i 2, and if for some i, as N (.) n i = N k +f i(n), f i (N) = Ω(N βi ), β i = 2, we cannot draw any meaningful conclusion for this borderline case. empirical c.d.f. of Ĥ c.d.f. of χ 2 k Ĥ values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of Ĥ c.d.f. of χ 2 k Ĥ values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure 2. Comparison of the empirical distribution of Ĥ and the Chi-square distribution. We create 2 samples of Ĥ using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In the equal group sizes case (a) we see that the distribution of Ĥ is very close to the Chi-square distribution. However, in the unequal group sizes case (b) the distribution of Ĥ is far away from the Chi-square distribution. Proposition 2. If θ = = θ k and if for each i, as N (.2) and n i N λ i, < λ i <, (.3) λ i k for some i, then as N E(H Ĥ) and Var(H Ĥ).

5 KRUSKAL-WALLIS ONE-WAY ANALYSIS 75 In Figure 2 we create 2 samples of Ĥ using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure 2(a) we consider the equal group sizes and in Figure 2(b) we consider the unequal group sizes. In the equal group sizes case we see that the distribution of the Kruskal-Wallis test statistic Ĥ based on linear placements is very close to the Chi-square distribution as we present in Theorem 2. However, in the unequal group sizes case the distribution of Ĥ is far away from the Chi-square distribution as we present in Propositions and 2. Let S N be the total sum of the normalized ranks W ij, i.e., (.4) S N = n i W ij. j= Then we also have the following variant of the central limit theorem of the linear placement statistics of Kim, Lee, and Wang [6]. Theorem 3. If θ = = θ k and if for each i, as N (.5) and n i N λ i, < λ i <, (.6) λ i k for some i, then as N (.7) Ŝ := S N E(S N ) Var(SN ) N(,) in distribution. Remark. If in Theorem 3, instead of (.6) we assume for each i (.8) λ i = k, we cannot prove the CLT. However, even in this case based on the simulation we think the CLT should hold under certain mild conditions. A little caution is needed here. As we see in (3.), since in the extremely balanced case n = = n k = N/k, the sum k ni j= R ij is a deterministic constant N(N+)/2, S N is not random and hence in this case the CLT cannot be true. In Figure 3 we create 2 samples of Ŝ using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure 3(a) we consider the equal group sizes and in Figure 3(b) we consider the unequal group sizes. In both caseswe see that the distribution ofŝ is very closeto the standard normal distribution as we present in Theorem 3.

6 76 Y. HONG AND S. LEE empirical c.d.f. of Ŝ c.d.f. of N(,) 5 5 Ŝ values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of Ŝ c.d.f. of N(,) 5 5 Ŝ values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure 3. Comparison of the empirical distribution of Ŝ and the standard normal distribution. We create 2 samples of Ŝ using iid standardnormal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In both cases we see that the distribution of Ŝ is very close to the standard normal distribution. We can still consider the generalization of Theorem 3 in line with Kim, Lee, and Wang [6]. For a non-constant C function ϕ : [,] R we let n i S ϕ N = ϕ(w ij ). j= If θ = = θ k and if for each i, as N n i N λ i, < λ i <, then as N we conjecture that Ŝ ϕ := Sϕ N E(Sϕ N ) Var(S ϕ N ) N(,) in distribution. In Figure 4 with ϕ (x) = x 2, ϕ 2 (x) = x, ϕ 3 (x) = log(+x), ϕ 4 (x) = e x, we create 2 samples of Ŝϕi using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, on the left we consider the equal group sizes and on the right we consider the unequal group sizes. In these equal and unequal cases we see that the distribution of Ŝϕi is close to the standard normal distribution as we conjecture. More interestingly, we notice that the distribution of Ŝϕi for the unequal group sizes is closer to the standard normal distribution than that for the equal group sizes. This agrees with the fact that in the case ϕ(x) = x of Theorem 3, S N in the extremely balanced case n = = n k = N/k, is not random and hence in this case the CLT does not hold. So, for general ϕ we conjecture even further that in the extremely balanced case n = = n k =

7 KRUSKAL-WALLIS ONE-WAY ANALYSIS 77 empirical c.d.f. of Ŝϕ c.d.f. of N(,) 5 5 Ŝ ϕ values empirical c.d.f. of Ŝϕ c.d.f. of N(,) 5 5 Ŝ ϕ values empirical c.d.f. of Ŝϕ2 c.d.f. of N(,) 5 5 Ŝ ϕ2 values empirical c.d.f. of Ŝϕ2 c.d.f. of N(,) 5 5 Ŝ ϕ2 values empirical c.d.f. of Ŝϕ3 c.d.f. of N(,) 5 5 Ŝ ϕ3 values empirical c.d.f. of Ŝϕ3 c.d.f. of N(,) 5 5 Ŝ ϕ3 values empirical c.d.f. of Ŝϕ4 c.d.f. of N(,) Ŝ ϕ4 values empirical c.d.f. of Ŝϕ4 c.d.f. of N(,) 5 5 Ŝ ϕ4 values Figure 4. Comparison of the empirical distribution of Ŝ ϕi and the standard normal distribution. With ϕ (x) = x 2, ϕ 2 (x) = x, ϕ 3 (x) = log( + x), ϕ 4 (x) = e x, we create 2 samples of Ŝϕ using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes n =, n 2 =, n 3 = 2, n 4 = 995, n 5 = 5 on the left and the unequal group sizes n =, n 2 = 5, n 3 = 2, n 4 = 995, n 5 = 5 on the right. In these equal and unequal cases we see that the distribution of Ŝϕi is close to the standard normal distribution. We also note that the distribution of Ŝϕi for the unequal group sizes on the right is closer to the standard normal distribution than that for the equal group sizes on the left. N/k, the CLT does not hold and that the convergence rate for the unequal group sizes is greater than that for the equal group sizes.

8 78 Y. HONG AND S. LEE In Section 2, we prove Kruskal-Wallis one-way analysis of variance based on linear placements, Theorems, 2, and Proposition using the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. In Section 3, we prove the variant of the central limit theorem of the linear placement statistics of Kim, Lee, and Wang [6] using Lyapunov CLT. 2. Kruskal-Wallis one-way analysis of variance based on linear placements In this section, we prove Kruskal-Wallis one-way analysis of variance based on linear placements, Theorems, 2, and Proposition using the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. To prove this, we start with the following lemma that characterizes the Chi-square distribution. Lemma. Let X = (X,...,X n ) T be an n dimensional multivariate normal random vector with mean vector and covariance matrix B. If the covariance matrix B satisfies the following two conditions then n k : X2 i B 2 = B and rank(b) = k, comes from the Chi-square distribution with the degree of freedom n Xi 2 χ 2 k. Proof. SinceBissymmetric, thereisanorthogonalmatrixpsuchthatp BP = Λ, where Λ is a diagonal matrix. Since rank(b) = k, Λ has k non-zero diagonal elements. By rearranging the coordinates if needed, we may assume that the first k diagonal elements of Λ are non-zero. Since B 2 = B, all the non-zero diagonal elements of Λ are in fact. Since X N n (,B) and since P BP = Λ, we see that Z := P X = (Z,...,Z n ) T is an n dimensional multivariate normal random vector with mean vector and covariance matrix Λ, i.e., Z N n (,Λ). Since the covariance matrix Λ is diagonal and since the first k diagonal elements of Λ are and all other elements of Λ are, Z,...,Z k are iid standard normal and Z k+ = = Z n =. Therefore, n n Xi 2 = X T X = Z T P PZ = Z T Z = Zi 2 = Zi 2 χ 2 k. Proof of Theorem. Let T := (T,...,T k ) T be given by 2n i (2.) T i := N(N +) (V i. EV i. ), i =,...,k. By Lemma, to prove Theorem it suffices to prove the following three; (2.2) T N k (,B),

9 (2.3) (2.4) KRUSKAL-WALLIS ONE-WAY ANALYSIS 79 B = B 2, rank(b) = k. To prove these we use the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. For each X ij let R ij be the rank of X ij among all the observations from all groups, and let L ij be the rank of X ij among all the observations from group i: n s R ij = (X st X ij ), L ij = s= t= n i (X it X ij ). t= We further let R i. be the sample mean rank of group i observations and R.. be the sample mean rank of all observations: (2.5) R i. = n i R ij := R i., n i n i (2.6) R.. = N j= n i R ij := N R... j= By the definitions of V ij, R ij, and L ij, we can rewrite V ij as (2.7) V ij = R ij L ij. Since (2.8) n i j= n i L ij = j= j = n i(n i +), 2 by (2.7) we see V i. EV i. = R i. ER i. and hence (2.9) V i. EV i. = R i. ER i.. So, we can rewrite T i as 2n i (2.) T i = N(N +) (R i. ER i. ), i =,...,k. Now the standard results on the Kruskal-Wallis one-way analysis of variance based on ranks say that T := (T,...,T k ) T with T i given by (2.) satisfies (2.2)-(2.4). Therefore, Theorem follows from Lemma. Lemma 2. For R ij, L ij, and R i., we have ER ij = N+ 2, Var(R ij) = (N )(N+) 2, Cov(R ij,r st ) = N+ 2, EL ij = ni+ 2, Var(L ij ) = (ni )(ni+) 2, Cov(L ij,l ik ) = ni+ 2, ER i. = ni(n+) 2, Var(R i. ) = ni(n ni)(n+) 2, Cov(R i.,r j. ) = ninj(n+) 2.

10 7 Y. HONG AND S. LEE Proof. By symmetry R ij is uniform on {,2,...,N}, i.e., for any observation X ij P(Rij = r) = for r N. N Also, by symmetry for any two different observations X ij and X st P(R ij = r,r st = l) = N(N ) for r l N. The lemma for R ij and R i. follows from the direct calculation using the above joint probability mass function. By the same reasoning one can get the lemma for L ij. By the definitions of Ĥ and H, we can decompose Ĥ into H and C; (2.) Ĥ = = where, by (2.9) 2 N(N +) 2 N(N +) n i (V i. V.. ) 2 n i ((V i. EV i. )+(EV i. V.. )) 2 = H +C, (2.2) [ ] 2 C := n i (EV i. V.. ) 2 + 2n i (EV i. V.. )(V i. EV i. ) N(N +) [ ] 2 = n i (EV i. V.. ) 2 + 2n i (EV i. V.. )(R i. ER i. ). N(N +) So by Theorem, to prove Theorem 2 it suffices to prove that under the conditions of Theorem 2, C in probability, or (2.3) EC, (2.4) Var(C), and to prove Propositions and 2 we just need to show that under the conditions of Propositions and 2, (2.5) EC, (2.6) Var(C). By (2.7), (2.8) and Lemma 2, (2.7) EV i. = n i E n i j= R ij n i E j= L ij = N n i. 2

11 KRUSKAL-WALLIS ONE-WAY ANALYSIS 7 By (2.7), (2.7) and (2.8), we see that EV i. V.. is actually a deterministic constant; (2.8) EV i. V.. = N n i 2 N n s V st = N s= t= s= n 2 s 2 n i 2. 2 k By (2.2) and (2.8), the random part of C is just N(N+) 2n i(ev i. V.. )R i.. Again, by (2.8) we can rewrite the major part k 2n i(ev i. V.. )R i. of the random part of C as (2.9) 2n i (EV i. V.. )R i. = = ( ) n 2 s 2n i N 2 n i 2 s= ( ) n 2 s N n i R i.. Takethe expectationand varianceonthe both sidesof(2.9). Then, bylemma 2 we have ( ) ( ) (2.2) E 2n i (EV i. V.. )R i. = n 2 s N n i E(R i. ) =, and (2.2) ( ) ( ( ) ) Var 2n i (EV i. V.. )R i. = Var n 2 s n i R i. N s= ( = N + ) 2 N n 3 i n 2 s. 2 So, by (2.2), (2.8), (2.2), and (2.2) we have (2.22) E(C) = 3f(N), (2.23) Var(C) = 2f(N), where (2.24) f(n) := N N 2 (N +) s= s= ( n 3 i s= n 2 s ) 2 Since n i = N k +f i(n) and since k f i(n) =, we have (2.25) f(n) = N N 2 (N +) ( ) 3 N k +f i(n) (. s= R i. ( ) ) 2 2 N k +f i(n)

12 72 Y. HONG AND S. LEE = N(N +) fi 3 (N)+ k(n +) ( 2 fi (N)) 2. N 2 (N +) fi 2 (N) Proof of Theorem 2. By dropping the last negative term in the right hand side of (2.25), we have (2.26) f(n) N(N +) fi 3 (N)+ k(n +) fi 2 (N). If by (.6) for all i, f i (N) = O(N β ), β < 2, by (2.26) as N f(n), and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.3) and (2.4) Theorem 2 follows. Proof of Proposition. By (.2) the first term in the right hand side of (2.25) is bounded by k N f2 i (N); ( ) (2.27) fi 3 (N) = o fi 2 (N). N(N +) N Again by (.2) (2.28) k N 2 fi 2 (N) k. So, by (2.28) among the second and third terms in the right hand side of (2.25) the second is the major term; ( ) 2 fi 2 k(n +) (N) (2.29) N 2 fi 2 (N +) (N) ( )( ) = fi 2 N + (N) k N 2 fi 2 (N). Combining(2.27),(2.28),and(2.29)weseethatf(N)isoforder N k f2 i (N); ( ) (2.3) f(n) = Ω fi 2 (N). N If by (.9) for some i, f i (N) = Ω(N βi ), 2 < β i <, by (2.3) as N f(n),

13 KRUSKAL-WALLIS ONE-WAY ANALYSIS 73 and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.5) and (2.6) Proposition follows. Proof of Proposition 2. Under the conditions(.2) and(.3) we have unequal < λ i < with k λ i =. For these λ i we consider a random variable X with P(X = λ i ) = λ i for i k. Let s calculate its variance. Since EX = k λ2 i and since EX2 = k we have 2 (2.3) Var(X) = EX 2 (EX) 2 = λi) 2. λ 3 i ( λ3 i, Since X is not a constant but finite random variable, it has a non-zero but finite variance. Therefore, by (2.3) we see that, for unequal < λ i < with k λ i = (2.32) < ( 2 λ 3 i λi) 2 <. By (2.24) and (2.32), as N ( 2 f(n) = N λ 3 i λi) 2 +o(), and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.5) and (2.6) Proposition 2 follows. 3. Proof of Theorem 3 In this section we use Lyapunov s CLT and prove Theorem 3. By (2.7) and (2.8) we can rewrite S N as (3.) S N = n i j= N + N n i R ij N + n i (n i +) 2(N n i ). Since {U ij := Rij N+ } have the following (joint) probability mass function ( P U ij = k ) = for k N, N + N ( P U ij = k ) N +,U i j = k = for k k N, N + N(N ) {U ij } have the following mean, variance, and covariance (3.2) EU ij = 2,

14 74 Y. HONG AND S. LEE (3.3) Var(U ij ) = 2N + 6N +6 4 = 2 6 N +O(N 2 ), (3.4) Cov(U ij,u i j ) = 3N2 N 2 2(N +)(N ) 4 = 2 N +O(N 2 ). Since N+ N n i λ i and since k n i(n i+) 2N(N n i) is non-random, we see that the asymptotics of S N is the same as that of T N, where T N is defined as n i (3.5) T N = A i U ij, where j= (3.6) A i = λ i. By (3.3), (3.4), and (3.6), n i (3.7) Var(T N ) = Var A i U ij = ij A 2 i ij [ = N 2 = N 2 j= ( 2 6 N +O(N 2 ) i j ij A i A i λ i ( λ i ) 2 λ i ( λ i ) 2 ) ( ) 2 N +O(N 2 ) = i = ( λ i λ i ( λ i )( λ i ) +O(N ) λ i λ i ) 2 +O(N ) Under the conditions (.5) and (.6) we have unequal < λ i < with k λ i =. For these λ i we consider a random variable Y with ( P Y = ) = λ i for i k. λ i Let s calculate its variance. Since EY = k λ i/( λ i ) and since EY 2 = k λ i/( λ i ) 2, we have ( ) 2 (3.8) Var(Y) = EY 2 (EY) 2 λ i = ( λ i ) 2 λ i. λ i Since by (.6), Y is not a constant but finite random variable, it has a nonzero but finite variance. Therefore, by (3.8) we see that, for unequal < λ i <. ]

15 with k λ i = (3.9) < KRUSKAL-WALLIS ONE-WAY ANALYSIS 75 λ i ( λ i ) 2 Therefore, by (3.7) and (3.9) we have ( λ i λ i (3.) Var(T N ) = Ω(N). ) 2 <. Since U ij, for any fixed δ >, E U ij EU ij 2+δ is bounded by. So, n i (3.) E A i U ij EA i U ij 2+δ = N j= n i A 2+δ i E U ij EU ij 2+δ j= [ λ i +o() ( λ i ) 2+δ ]. n i A 2+δ i By (3.) and (3.) one can easily check the Lyapunov condition; lim (Var(T n i N)) (+δ 2 ) E A i U ij EA i U ij 2+δ. n j= Hence, by Lyapunov s CLT Theorem 3 follows. References [] H. Chernoff and I. R. Savage, Asymptotic normality and efficiency of certain nonparametric test statistics, Ann. Math. Statist. 29 (958), [2] V. Dupac and J. Hajek, Asymptotic normality of simple linear rank statistics under alternative II., Ann. Math. Statist. 4 (969), [3] Z. Govindarajulu, L. Le Cam, and M. Raghavachari, Generalizations of theorems of Chernoff and Savage on the asymptotic normality of test statistics, Proc. Fifth Berkeley Sympos. Math. Statist. and Probability (Berkeley, Calif., 965/66) Vol. I: Statistics, pp Univ. California Press, Berkeley, Calif, 967. [4] J. Hajek, Asymptotic normality of simple linear rank statistics under alternatives, Ann. Math. Statist. 39 (968), [5] D. Kim, A class of distribution-free treatments versus control tests based on placements, Far East J. Theor. Stat. 3 (999), no., [6] D. Kim, S. Lee, and W. Wang, The asymptotic behavior of linear placement statistics, Statist. Probab. Lett. 8 (2), no. 2, [7] D. Kim and D. A. Wolfe, Properties of distribution-free two-sample procedures based on placements, Far East J. Math. Sci. (993), no. 2, [8] A. M. Mood, Introduction to the Theory of Statistics, McGraw-Hill, New York, 95. [9] J. Orban and D. A. Wolfe, A class of distribution-free two-sample tests based on placements, J. Amer. Statist. Assoc. 77 (982), no. 379, [] R. Pyke and G. R. Shorack, Weak convergence of a two-sample empirical process and a new approach to Chernoff-Savage theorems, Ann. Math. Statist. 39 (968),

16 76 Y. HONG AND S. LEE [] F. Wilcoxon, Individual comparisons by ranking methods, Biometrics (945), no. 6, Yicheng Hong Department of Mathematics Yanbian University Yanji 332, P. R. China and Department of Mathematics Yonsei University Seoul 2-749, Korea address: tophych@gmail.com Sungchul Lee Department of Mathematics Yonsei University Seoul 2-749, Korea address: sungchul@yonsei.ac.kr

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip

TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississippi University, MS38677 K-sample location test, Koziol-Green

More information

Nonparametric hypothesis tests and permutation tests

Nonparametric hypothesis tests and permutation tests Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

Dr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines)

Dr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines) Dr. Maddah ENMG 617 EM Statistics 10/12/12 Nonparametric Statistics (Chapter 16, Hines) Introduction Most of the hypothesis testing presented so far assumes normally distributed data. These approaches

More information

Nonparametric tests. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 704: Data Analysis I

Nonparametric tests. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 704: Data Analysis I 1 / 16 Nonparametric tests Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I Nonparametric one and two-sample tests 2 / 16 If data do not come from a normal

More information

A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES

A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES Bull. Korean Math. Soc. 52 (205), No. 3, pp. 825 836 http://dx.doi.org/0.434/bkms.205.52.3.825 A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES Yongfeng Wu and Mingzhu

More information

Unit 14: Nonparametric Statistical Methods

Unit 14: Nonparametric Statistical Methods Unit 14: Nonparametric Statistical Methods Statistics 571: Statistical Methods Ramón V. León 8/8/2003 Unit 14 - Stat 571 - Ramón V. León 1 Introductory Remarks Most methods studied so far have been based

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

ON TWO RESULTS IN MULTIPLE TESTING

ON TWO RESULTS IN MULTIPLE TESTING ON TWO RESULTS IN MULTIPLE TESTING By Sanat K. Sarkar 1, Pranab K. Sen and Helmut Finner Temple University, University of North Carolina at Chapel Hill and University of Duesseldorf Two known results in

More information

Advanced Statistics II: Non Parametric Tests

Advanced Statistics II: Non Parametric Tests Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

This condition was introduced by Chandra [1]. Ordonez Cabrera [5] extended the notion of Cesàro uniform integrability

This condition was introduced by Chandra [1]. Ordonez Cabrera [5] extended the notion of Cesàro uniform integrability Bull. Korean Math. Soc. 4 (2004), No. 2, pp. 275 282 WEAK LAWS FOR WEIGHTED SUMS OF RANDOM VARIABLES Soo Hak Sung Abstract. Let {a ni, u n i, n } be an array of constants. Let {X ni, u n i, n } be {a ni

More information

Complete moment convergence of weighted sums for processes under asymptotically almost negatively associated assumptions

Complete moment convergence of weighted sums for processes under asymptotically almost negatively associated assumptions Proc. Indian Acad. Sci. Math. Sci.) Vol. 124, No. 2, May 214, pp. 267 279. c Indian Academy of Sciences Complete moment convergence of weighted sums for processes under asymptotically almost negatively

More information

Lecture 28: Asymptotic confidence sets

Lecture 28: Asymptotic confidence sets Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level

More information

Empirical Power of Four Statistical Tests in One Way Layout

Empirical Power of Four Statistical Tests in One Way Layout International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo

More information

Final Examination Statistics 200C. T. Ferguson June 11, 2009

Final Examination Statistics 200C. T. Ferguson June 11, 2009 Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean

More information

Lecture 11. Multivariate Normal theory

Lecture 11. Multivariate Normal theory 10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances

More information

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution

More information

IMPROVING TWO RESULTS IN MULTIPLE TESTING

IMPROVING TWO RESULTS IN MULTIPLE TESTING IMPROVING TWO RESULTS IN MULTIPLE TESTING By Sanat K. Sarkar 1, Pranab K. Sen and Helmut Finner Temple University, University of North Carolina at Chapel Hill and University of Duesseldorf October 11,

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Empirical Likelihood Tests for High-dimensional Data

Empirical Likelihood Tests for High-dimensional Data Empirical Likelihood Tests for High-dimensional Data Department of Statistics and Actuarial Science University of Waterloo, Canada ICSA - Canada Chapter 2013 Symposium Toronto, August 2-3, 2013 Based on

More information

Inferential Statistics

Inferential Statistics Inferential Statistics Eva Riccomagno, Maria Piera Rogantin DIMA Università di Genova riccomagno@dima.unige.it rogantin@dima.unige.it Part G Distribution free hypothesis tests 1. Classical and distribution-free

More information

ON PITMAN EFFICIENCY OF

ON PITMAN EFFICIENCY OF 1. Summary ON PITMAN EFFICIENCY OF SOME TESTS OF SCALE FOR THE GAMMA DISTRIBUTION BARRY R. JAMES UNIVERSITY OF CALIFORNIA, BERKELEY A comparison is made of several two sample rank tests for scale change

More information

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),

More information

Nonparametric Location Tests: k-sample

Nonparametric Location Tests: k-sample Nonparametric Location Tests: k-sample Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 04-Jan-2017 Nathaniel E. Helwig (U of Minnesota)

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Lecture 3 - Expectation, inequalities and laws of large numbers

Lecture 3 - Expectation, inequalities and laws of large numbers Lecture 3 - Expectation, inequalities and laws of large numbers Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 3 - Expectation, inequalities and laws of large numbersapril 19, 2009 1 / 67 Part

More information

A Signed-Rank Test Based on the Score Function

A Signed-Rank Test Based on the Score Function Applied Mathematical Sciences, Vol. 10, 2016, no. 51, 2517-2527 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.66189 A Signed-Rank Test Based on the Score Function Hyo-Il Park Department

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Yunhi Cho and Young-One Kim

Yunhi Cho and Young-One Kim Bull. Korean Math. Soc. 41 (2004), No. 1, pp. 27 43 ANALYTIC PROPERTIES OF THE LIMITS OF THE EVEN AND ODD HYPERPOWER SEQUENCES Yunhi Cho Young-One Kim Dedicated to the memory of the late professor Eulyong

More information

f (1 0.5)/n Z =

f (1 0.5)/n Z = Math 466/566 - Homework 4. We want to test a hypothesis involving a population proportion. The unknown population proportion is p. The null hypothesis is p = / and the alternative hypothesis is p > /.

More information

Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek

Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek J. Korean Math. Soc. 41 (2004), No. 5, pp. 883 894 CONVERGENCE OF WEIGHTED SUMS FOR DEPENDENT RANDOM VARIABLES Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek Abstract. We discuss in this paper the strong

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

STAT 135 Lab 8 Hypothesis Testing Review, Mann-Whitney Test by Normal Approximation, and Wilcoxon Signed Rank Test.

STAT 135 Lab 8 Hypothesis Testing Review, Mann-Whitney Test by Normal Approximation, and Wilcoxon Signed Rank Test. STAT 135 Lab 8 Hypothesis Testing Review, Mann-Whitney Test by Normal Approximation, and Wilcoxon Signed Rank Test. Rebecca Barter March 30, 2015 Mann-Whitney Test Mann-Whitney Test Recall that the Mann-Whitney

More information

Lecture 32: Asymptotic confidence sets and likelihoods

Lecture 32: Asymptotic confidence sets and likelihoods Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence

More information

ECON Fundamentals of Probability

ECON Fundamentals of Probability ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,

More information

Institute of Actuaries of India

Institute of Actuaries of India Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the

More information

Bivariate Paired Numerical Data

Bivariate Paired Numerical Data Bivariate Paired Numerical Data Pearson s correlation, Spearman s ρ and Kendall s τ, tests of independence University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

WLLN for arrays of nonnegative random variables

WLLN for arrays of nonnegative random variables WLLN for arrays of nonnegative random variables Stefan Ankirchner Thomas Kruse Mikhail Urusov November 8, 26 We provide a weak law of large numbers for arrays of nonnegative and pairwise negatively associated

More information

4. Distributions of Functions of Random Variables

4. Distributions of Functions of Random Variables 4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n

More information

1 Probability theory. 2 Random variables and probability theory.

1 Probability theory. 2 Random variables and probability theory. Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major

More information

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Hypothesis testing. Anna Wegloop Niels Landwehr/Tobias Scheffer

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Hypothesis testing. Anna Wegloop Niels Landwehr/Tobias Scheffer Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Hypothesis testing Anna Wegloop iels Landwehr/Tobias Scheffer Why do a statistical test? input computer model output Outlook ull-hypothesis

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

6 The normal distribution, the central limit theorem and random samples

6 The normal distribution, the central limit theorem and random samples 6 The normal distribution, the central limit theorem and random samples 6.1 The normal distribution We mentioned the normal (or Gaussian) distribution in Chapter 4. It has density f X (x) = 1 σ 1 2π e

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

NEW CONSTRUCTION OF THE EAGON-NORTHCOTT COMPLEX. Oh-Jin Kang and Joohyung Kim

NEW CONSTRUCTION OF THE EAGON-NORTHCOTT COMPLEX. Oh-Jin Kang and Joohyung Kim Korean J Math 20 (2012) No 2 pp 161 176 NEW CONSTRUCTION OF THE EAGON-NORTHCOTT COMPLEX Oh-Jin Kang and Joohyung Kim Abstract The authors [6 introduced the concept of a complete matrix of grade g > 3 to

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

The symmetry in the martingale inequality

The symmetry in the martingale inequality Statistics & Probability Letters 56 2002 83 9 The symmetry in the martingale inequality Sungchul Lee a; ;, Zhonggen Su a;b;2 a Department of Mathematics, Yonsei University, Seoul 20-749, South Korea b

More information

Gaussian vectors and central limit theorem

Gaussian vectors and central limit theorem Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables

More information

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory

More information

Class 8 Review Problems solutions, 18.05, Spring 2014

Class 8 Review Problems solutions, 18.05, Spring 2014 Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots

More information

Analysis of Experimental Designs

Analysis of Experimental Designs Analysis of Experimental Designs p. 1/? Analysis of Experimental Designs Gilles Lamothe Mathematics and Statistics University of Ottawa Analysis of Experimental Designs p. 2/? Review of Probability A short

More information

1 Review of Probability and Distributions

1 Review of Probability and Distributions Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote

More information

PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS

PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS REZA HASHEMI and MOLUD ABDOLAHI Department of Statistics, Faculty of Science, Razi University, 67149, Kermanshah,

More information

Introduction to Nonparametric Statistics

Introduction to Nonparametric Statistics Introduction to Nonparametric Statistics by James Bernhard Spring 2012 Parameters Parametric method Nonparametric method µ[x 2 X 1 ] paired t-test Wilcoxon signed rank test µ[x 1 ], µ[x 2 ] 2-sample t-test

More information

Notes on Time Series Modeling

Notes on Time Series Modeling Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

. Find E(V ) and var(v ).

. Find E(V ) and var(v ). Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number

More information

Multivariate Time Series: Part 4

Multivariate Time Series: Part 4 Multivariate Time Series: Part 4 Cointegration Gerald P. Dwyer Clemson University March 2016 Outline 1 Multivariate Time Series: Part 4 Cointegration Engle-Granger Test for Cointegration Johansen Test

More information

Intermediate Econometrics

Intermediate Econometrics Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the

More information

Shih-sen Chang, Yeol Je Cho, and Haiyun Zhou

Shih-sen Chang, Yeol Je Cho, and Haiyun Zhou J. Korean Math. Soc. 38 (2001), No. 6, pp. 1245 1260 DEMI-CLOSED PRINCIPLE AND WEAK CONVERGENCE PROBLEMS FOR ASYMPTOTICALLY NONEXPANSIVE MAPPINGS Shih-sen Chang, Yeol Je Cho, and Haiyun Zhou Abstract.

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7

UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7 UCSD ECE50 Handout #4 Prof Young-Han Kim Wednesday, June 6, 08 Solutions to Exercise Set #7 Polya s urn An urn initially has one red ball and one white ball Let X denote the name of the first ball drawn

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Conditional moment representations for dependent random variables

Conditional moment representations for dependent random variables Conditional moment representations for dependent random variables W lodzimierz Bryc Department of Mathematics University of Cincinnati Cincinnati, OH 45 22-0025 bryc@ucbeh.san.uc.edu November 9, 995 Abstract

More information

A REVERSE TO THE JEFFREYS LINDLEY PARADOX

A REVERSE TO THE JEFFREYS LINDLEY PARADOX PROBABILITY AND MATHEMATICAL STATISTICS Vol. 38, Fasc. 1 (2018), pp. 243 247 doi:10.19195/0208-4147.38.1.13 A REVERSE TO THE JEFFREYS LINDLEY PARADOX BY WIEBE R. P E S T M A N (LEUVEN), FRANCIS T U E R

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2

Some Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2 STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square

More information

THE EXPONENTIAL DISTRIBUTION ANALOG OF THE GRUBBS WEAVER METHOD

THE EXPONENTIAL DISTRIBUTION ANALOG OF THE GRUBBS WEAVER METHOD THE EXPONENTIAL DISTRIBUTION ANALOG OF THE GRUBBS WEAVER METHOD ANDREW V. SILLS AND CHARLES W. CHAMP Abstract. In Grubbs and Weaver (1947), the authors suggest a minimumvariance unbiased estimator for

More information

Orthogonal decompositions in growth curve models

Orthogonal decompositions in growth curve models ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 4, Orthogonal decompositions in growth curve models Daniel Klein and Ivan Žežula Dedicated to Professor L. Kubáček on the occasion

More information

Chapter 11 Sampling Distribution. Stat 115

Chapter 11 Sampling Distribution. Stat 115 Chapter 11 Sampling Distribution Stat 115 1 Definition 11.1 : Random Sample (finite population) Suppose we select n distinct elements from a population consisting of N elements, using a particular probability

More information

USING A BIMODAL KERNEL FOR A NONPARAMETRIC REGRESSION SPECIFICATION TEST

USING A BIMODAL KERNEL FOR A NONPARAMETRIC REGRESSION SPECIFICATION TEST Statistica Sinica 25 (2015), 1145-1161 doi:http://dx.doi.org/10.5705/ss.2014.008 USING A BIMODAL KERNEL FOR A NONPARAMETRIC REGRESSION SPECIFICATION TEST Cheolyong Park 1, Tae Yoon Kim 1, Jeongcheol Ha

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

Statistics (1): Estimation

Statistics (1): Estimation Statistics (1): Estimation Marco Banterlé, Christian Robert and Judith Rousseau Practicals 2014-2015 L3, MIDO, Université Paris Dauphine 1 Table des matières 1 Random variables, probability, expectation

More information

Contents. Acknowledgments. xix

Contents. Acknowledgments. xix Table of Preface Acknowledgments page xv xix 1 Introduction 1 The Role of the Computer in Data Analysis 1 Statistics: Descriptive and Inferential 2 Variables and Constants 3 The Measurement of Variables

More information

Negative Multinomial Model and Cancer. Incidence

Negative Multinomial Model and Cancer. Incidence Generalized Linear Model under the Extended Negative Multinomial Model and Cancer Incidence S. Lahiri & Sunil K. Dhar Department of Mathematical Sciences, CAMS New Jersey Institute of Technology, Newar,

More information

A Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms. by R. L. Eubank and Suojin Wang Texas A&M University

A Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms. by R. L. Eubank and Suojin Wang Texas A&M University A Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms by R. L. Eubank and Suojin Wang Texas A&M University ABSTRACT. A central limit theorem is established for the sum of stochastically

More information

Functions of Several Random Variables (Ch. 5.5)

Functions of Several Random Variables (Ch. 5.5) (Ch. 5.5) Iowa State University Mar 7, 2013 Iowa State University Mar 7, 2013 1 / 37 Outline Iowa State University Mar 7, 2013 2 / 37 several random variables We often consider functions of random variables

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

CS145: Probability & Computing

CS145: Probability & Computing CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,

More information

Simple Linear Regression: The Model

Simple Linear Regression: The Model Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random

More information

STA 4322 Exam I Name: Introduction to Statistics Theory

STA 4322 Exam I Name: Introduction to Statistics Theory STA 4322 Exam I Name: Introduction to Statistics Theory Fall 2013 UF-ID: Instructions: There are 100 total points. You must show your work to receive credit. Read each part of each question carefully.

More information

Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY. SECOND YEAR B.Sc. SEMESTER - III

Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY. SECOND YEAR B.Sc. SEMESTER - III Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY SECOND YEAR B.Sc. SEMESTER - III SYLLABUS FOR S. Y. B. Sc. STATISTICS Academic Year 07-8 S.Y. B.Sc. (Statistics)

More information

Rank-sum Test Based on Order Restricted Randomized Design

Rank-sum Test Based on Order Restricted Randomized Design Rank-sum Test Based on Order Restricted Randomized Design Omer Ozturk and Yiping Sun Abstract One of the main principles in a design of experiment is to use blocking factors whenever it is possible. On

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible

More information

NONPARAMETRICS. Statistical Methods Based on Ranks E. L. LEHMANN HOLDEN-DAY, INC. McGRAW-HILL INTERNATIONAL BOOK COMPANY

NONPARAMETRICS. Statistical Methods Based on Ranks E. L. LEHMANN HOLDEN-DAY, INC. McGRAW-HILL INTERNATIONAL BOOK COMPANY NONPARAMETRICS Statistical Methods Based on Ranks E. L. LEHMANN University of California, Berkeley With the special assistance of H. J. M. D'ABRERA University of California, Berkeley HOLDEN-DAY, INC. San

More information

Mark Scheme (Results) Summer 2007

Mark Scheme (Results) Summer 2007 Mark Scheme (Results) Summer 007 GCE GCE Mathematics Statistics S3 (6691) Edexcel Limited. Registered in England and Wales No. 4496750 Registered Office: One90 High Holborn, London WC1V 7BH June 007 6691

More information

Chapter 18 Resampling and Nonparametric Approaches To Data

Chapter 18 Resampling and Nonparametric Approaches To Data Chapter 18 Resampling and Nonparametric Approaches To Data 18.1 Inferences in children s story summaries (McConaughy, 1980): a. Analysis using Wilcoxon s rank-sum test: Younger Children Older Children

More information

Non-Parametric Statistics: When Normal Isn t Good Enough"

Non-Parametric Statistics: When Normal Isn t Good Enough Non-Parametric Statistics: When Normal Isn t Good Enough" Professor Ron Fricker" Naval Postgraduate School" Monterey, California" 1/28/13 1 A Bit About Me" Academic credentials" Ph.D. and M.A. in Statistics,

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information