KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS
|
|
- Gabriella Britney Price
- 5 years ago
- Views:
Transcription
1 Bull. Korean Math. Soc. 5 (24), No. 3, pp KRUSKAL-WALLIS ONE-WAY ANALYSIS OF VARIANCE BASED ON LINEAR PLACEMENTS Yicheng Hong and Sungchul Lee Abstract. The limiting distribution for the linear placement statistics under the null hypotheses has been provided by Orban and Wolfe [9] and Kim [5] when one of the sample sizes goes to infinity, and by Kim, Lee and Wang [6] when the sample sizes of each group go to infinity simultaneously. In this paper we establish the generalized Kruskal-Wallis one-way analysis of variance for the linear placement statistics.. Introduction and statement of main results To determine whether the mutually independent random samples come from a common underlying distribution, we use various kinds of two-sample tests. Two-sample tests that are valid with no assumptions (except the continuity of the underlying distribution), are called the distribution-free procedures. Some distribution-free procedures use the sample ranks; the rank-sum test of Wilcoxon [] and the median test of Mood [8]. Other distribution-free tests utilize the sample placements; the linear placement test of Orban and Wolfe [9] and the fixed and updated linear placement test of Kim [5]. Asymptotic theories for the sample rank tests are well developed; see Chernoff and Savage [], Govindarajulu, Le Cam and Raghavachari [3], Hajek [4], Pyke and Shorak [], and Dupac and Hajek [2]. However, asymptotic theories for the sample linear placement tests are quite limited. Orban and Wolfe [9] provided the limiting theory when one of the sample sizes goes to infinity. Kim and Wolfe [7] investigated the iterative asymptotic distribution when the control group sample size goes to infinity then the treatment group sample size goes to infinity. Similarly, Kim [5] established the limiting theory of the fixed and updated linear placement test when the control group sample size goes to Received March, Mathematics Subject Classification. Primary 6F5, 62E2, 62G2; Secondary 6E5, 6E. Key words and phrases. Kruskal-Wallis one-way analysis of variance, central limit theorem, linear placement statistic. This work was supported by the BK2 project of the Department of Mathematics, Yonsei University and Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education, Science and Technology(2-99). 7 c 24 Korean Mathematical Society
2 72 Y. HONG AND S. LEE infinity. Moreover, Kim, Lee, and Wang [6] further established the central limit theorem when the sample sizes of each group go to infinity simultaneously. In this paper we develop the Kruskal-Wallis one-way analysis of variance based on linear placements. The usual Kruskal-Wallis one-way analysis of variance is based on ranks but our Kruskal-Wallis one-way analysis of variance is based on linear placements, which can be used for comparing more than two samples that are independent, or not related. Our main results are follows. First we fix our notations. The data set consists of k groups and the group i has n i observations X ij, j n i. Overall we have N = k n i observations. We assume that X ij, i k, j n i, are mutually independent samples from the continuous cumulative distribution functions G i (x) = F(x θ i ). For each X ij let V ij be the rank of X ij among all observations excluding observations from the group that X ij belongs, and let W ij be its normalized rank: V ij = n s (.) (X st X ij ), s i t= V ij (.2) W ij =. N n i We further let V i. be the sample mean rank of group i observations and V.. be the sample mean rank of all observations: = n i (.3) V ij, n i (.4) V i. V.. = N j= n i V ij. j= When the sample sizes of all groups go to infinity simultaneously, we have the following Kruskal-Wallis one-way analysis of variance based on linear placements. Theorem. If θ = = θ k and if for each i, as N n i N λ i, < λ i <, then as N (.5) H := 2 N(N +) n i (V i. EV i. ) 2 χ 2 k, where by (2.7), EV i. = (N n i )/2. In Figure we create 2 samples of H using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure (a) we consider the equal group sizes and in Figure (b) we consider the unequal group sizes. In
3 KRUSKAL-WALLIS ONE-WAY ANALYSIS 73 empirical c.d.f. of H c.d.f. of χ 2 k H values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of H c.d.f. of χ 2 k H values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure. Comparison of the empirical distribution of H and the Chi-square distribution. We create 2 samples of H using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In both cases we see that the distribution of H is very close to the Chi-square distribution. both cases we see that the distribution of the Kruskal-Wallis test statistic H based on linear placements is very close to the Chi-square distribution as we present in Theorem. If we have the equal group sizes, we replace EV i. in H with V.. and we still get the Chi-square limit. Theorem 2. If θ = = θ k and if for each i, as N (.6) n i = N k +f i(n), f i (N) = O(N βi ), β i < 2, then as N (.7) Ĥ := 2 N(N +) n i (V i. V.. ) 2 χ 2 k. However, if we don t have the same group sizes, we cannot replace EV i. in H with V.. to get the Chi-square limit. Proposition. If θ = = θ k, if for each i, as N (.8) n i = N k +f i(n), f i (N) = O(N βi ), β i <, and if for some i, as N (.9) n i = N k +f i(n), f i (N) = Ω(N βi ), then as N 2 < β i <, E(H Ĥ) and Var(H Ĥ).
4 74 Y. HONG AND S. LEE Remark. If in Proposition, instead of (.8) and (.9) we assume for each i, as N (.) n i = N k +f i(n), f i (N) = O(N βi ), β i 2, and if for some i, as N (.) n i = N k +f i(n), f i (N) = Ω(N βi ), β i = 2, we cannot draw any meaningful conclusion for this borderline case. empirical c.d.f. of Ĥ c.d.f. of χ 2 k Ĥ values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of Ĥ c.d.f. of χ 2 k Ĥ values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure 2. Comparison of the empirical distribution of Ĥ and the Chi-square distribution. We create 2 samples of Ĥ using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In the equal group sizes case (a) we see that the distribution of Ĥ is very close to the Chi-square distribution. However, in the unequal group sizes case (b) the distribution of Ĥ is far away from the Chi-square distribution. Proposition 2. If θ = = θ k and if for each i, as N (.2) and n i N λ i, < λ i <, (.3) λ i k for some i, then as N E(H Ĥ) and Var(H Ĥ).
5 KRUSKAL-WALLIS ONE-WAY ANALYSIS 75 In Figure 2 we create 2 samples of Ĥ using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure 2(a) we consider the equal group sizes and in Figure 2(b) we consider the unequal group sizes. In the equal group sizes case we see that the distribution of the Kruskal-Wallis test statistic Ĥ based on linear placements is very close to the Chi-square distribution as we present in Theorem 2. However, in the unequal group sizes case the distribution of Ĥ is far away from the Chi-square distribution as we present in Propositions and 2. Let S N be the total sum of the normalized ranks W ij, i.e., (.4) S N = n i W ij. j= Then we also have the following variant of the central limit theorem of the linear placement statistics of Kim, Lee, and Wang [6]. Theorem 3. If θ = = θ k and if for each i, as N (.5) and n i N λ i, < λ i <, (.6) λ i k for some i, then as N (.7) Ŝ := S N E(S N ) Var(SN ) N(,) in distribution. Remark. If in Theorem 3, instead of (.6) we assume for each i (.8) λ i = k, we cannot prove the CLT. However, even in this case based on the simulation we think the CLT should hold under certain mild conditions. A little caution is needed here. As we see in (3.), since in the extremely balanced case n = = n k = N/k, the sum k ni j= R ij is a deterministic constant N(N+)/2, S N is not random and hence in this case the CLT cannot be true. In Figure 3 we create 2 samples of Ŝ using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, in Figure 3(a) we consider the equal group sizes and in Figure 3(b) we consider the unequal group sizes. In both caseswe see that the distribution ofŝ is very closeto the standard normal distribution as we present in Theorem 3.
6 76 Y. HONG AND S. LEE empirical c.d.f. of Ŝ c.d.f. of N(,) 5 5 Ŝ values (a) The equal group sizes; n =, n 2 =, n 3 = 2, n 4 = 995 and n 5 = 5. empirical c.d.f. of Ŝ c.d.f. of N(,) 5 5 Ŝ values (b) The unequal group sizes; n =, n 2 = 5, n 3 = 2, n 4 = 995 and n 5 = 5. Figure 3. Comparison of the empirical distribution of Ŝ and the standard normal distribution. We create 2 samples of Ŝ using iid standardnormal X ij, i =,...,5, j =,...,n i, from both the equal group sizes (a) and the unequal group sizes (b). In both cases we see that the distribution of Ŝ is very close to the standard normal distribution. We can still consider the generalization of Theorem 3 in line with Kim, Lee, and Wang [6]. For a non-constant C function ϕ : [,] R we let n i S ϕ N = ϕ(w ij ). j= If θ = = θ k and if for each i, as N n i N λ i, < λ i <, then as N we conjecture that Ŝ ϕ := Sϕ N E(Sϕ N ) Var(S ϕ N ) N(,) in distribution. In Figure 4 with ϕ (x) = x 2, ϕ 2 (x) = x, ϕ 3 (x) = log(+x), ϕ 4 (x) = e x, we create 2 samples of Ŝϕi using iid standard normal X ij, i =,...,5, j =,...,n i. More specifically, on the left we consider the equal group sizes and on the right we consider the unequal group sizes. In these equal and unequal cases we see that the distribution of Ŝϕi is close to the standard normal distribution as we conjecture. More interestingly, we notice that the distribution of Ŝϕi for the unequal group sizes is closer to the standard normal distribution than that for the equal group sizes. This agrees with the fact that in the case ϕ(x) = x of Theorem 3, S N in the extremely balanced case n = = n k = N/k, is not random and hence in this case the CLT does not hold. So, for general ϕ we conjecture even further that in the extremely balanced case n = = n k =
7 KRUSKAL-WALLIS ONE-WAY ANALYSIS 77 empirical c.d.f. of Ŝϕ c.d.f. of N(,) 5 5 Ŝ ϕ values empirical c.d.f. of Ŝϕ c.d.f. of N(,) 5 5 Ŝ ϕ values empirical c.d.f. of Ŝϕ2 c.d.f. of N(,) 5 5 Ŝ ϕ2 values empirical c.d.f. of Ŝϕ2 c.d.f. of N(,) 5 5 Ŝ ϕ2 values empirical c.d.f. of Ŝϕ3 c.d.f. of N(,) 5 5 Ŝ ϕ3 values empirical c.d.f. of Ŝϕ3 c.d.f. of N(,) 5 5 Ŝ ϕ3 values empirical c.d.f. of Ŝϕ4 c.d.f. of N(,) Ŝ ϕ4 values empirical c.d.f. of Ŝϕ4 c.d.f. of N(,) 5 5 Ŝ ϕ4 values Figure 4. Comparison of the empirical distribution of Ŝ ϕi and the standard normal distribution. With ϕ (x) = x 2, ϕ 2 (x) = x, ϕ 3 (x) = log( + x), ϕ 4 (x) = e x, we create 2 samples of Ŝϕ using iid standard normal X ij, i =,...,5, j =,...,n i, from both the equal group sizes n =, n 2 =, n 3 = 2, n 4 = 995, n 5 = 5 on the left and the unequal group sizes n =, n 2 = 5, n 3 = 2, n 4 = 995, n 5 = 5 on the right. In these equal and unequal cases we see that the distribution of Ŝϕi is close to the standard normal distribution. We also note that the distribution of Ŝϕi for the unequal group sizes on the right is closer to the standard normal distribution than that for the equal group sizes on the left. N/k, the CLT does not hold and that the convergence rate for the unequal group sizes is greater than that for the equal group sizes.
8 78 Y. HONG AND S. LEE In Section 2, we prove Kruskal-Wallis one-way analysis of variance based on linear placements, Theorems, 2, and Proposition using the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. In Section 3, we prove the variant of the central limit theorem of the linear placement statistics of Kim, Lee, and Wang [6] using Lyapunov CLT. 2. Kruskal-Wallis one-way analysis of variance based on linear placements In this section, we prove Kruskal-Wallis one-way analysis of variance based on linear placements, Theorems, 2, and Proposition using the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. To prove this, we start with the following lemma that characterizes the Chi-square distribution. Lemma. Let X = (X,...,X n ) T be an n dimensional multivariate normal random vector with mean vector and covariance matrix B. If the covariance matrix B satisfies the following two conditions then n k : X2 i B 2 = B and rank(b) = k, comes from the Chi-square distribution with the degree of freedom n Xi 2 χ 2 k. Proof. SinceBissymmetric, thereisanorthogonalmatrixpsuchthatp BP = Λ, where Λ is a diagonal matrix. Since rank(b) = k, Λ has k non-zero diagonal elements. By rearranging the coordinates if needed, we may assume that the first k diagonal elements of Λ are non-zero. Since B 2 = B, all the non-zero diagonal elements of Λ are in fact. Since X N n (,B) and since P BP = Λ, we see that Z := P X = (Z,...,Z n ) T is an n dimensional multivariate normal random vector with mean vector and covariance matrix Λ, i.e., Z N n (,Λ). Since the covariance matrix Λ is diagonal and since the first k diagonal elements of Λ are and all other elements of Λ are, Z,...,Z k are iid standard normal and Z k+ = = Z n =. Therefore, n n Xi 2 = X T X = Z T P PZ = Z T Z = Zi 2 = Zi 2 χ 2 k. Proof of Theorem. Let T := (T,...,T k ) T be given by 2n i (2.) T i := N(N +) (V i. EV i. ), i =,...,k. By Lemma, to prove Theorem it suffices to prove the following three; (2.2) T N k (,B),
9 (2.3) (2.4) KRUSKAL-WALLIS ONE-WAY ANALYSIS 79 B = B 2, rank(b) = k. To prove these we use the tools developed for Kruskal-Wallis one-way analysis of variance based on ranks. For each X ij let R ij be the rank of X ij among all the observations from all groups, and let L ij be the rank of X ij among all the observations from group i: n s R ij = (X st X ij ), L ij = s= t= n i (X it X ij ). t= We further let R i. be the sample mean rank of group i observations and R.. be the sample mean rank of all observations: (2.5) R i. = n i R ij := R i., n i n i (2.6) R.. = N j= n i R ij := N R... j= By the definitions of V ij, R ij, and L ij, we can rewrite V ij as (2.7) V ij = R ij L ij. Since (2.8) n i j= n i L ij = j= j = n i(n i +), 2 by (2.7) we see V i. EV i. = R i. ER i. and hence (2.9) V i. EV i. = R i. ER i.. So, we can rewrite T i as 2n i (2.) T i = N(N +) (R i. ER i. ), i =,...,k. Now the standard results on the Kruskal-Wallis one-way analysis of variance based on ranks say that T := (T,...,T k ) T with T i given by (2.) satisfies (2.2)-(2.4). Therefore, Theorem follows from Lemma. Lemma 2. For R ij, L ij, and R i., we have ER ij = N+ 2, Var(R ij) = (N )(N+) 2, Cov(R ij,r st ) = N+ 2, EL ij = ni+ 2, Var(L ij ) = (ni )(ni+) 2, Cov(L ij,l ik ) = ni+ 2, ER i. = ni(n+) 2, Var(R i. ) = ni(n ni)(n+) 2, Cov(R i.,r j. ) = ninj(n+) 2.
10 7 Y. HONG AND S. LEE Proof. By symmetry R ij is uniform on {,2,...,N}, i.e., for any observation X ij P(Rij = r) = for r N. N Also, by symmetry for any two different observations X ij and X st P(R ij = r,r st = l) = N(N ) for r l N. The lemma for R ij and R i. follows from the direct calculation using the above joint probability mass function. By the same reasoning one can get the lemma for L ij. By the definitions of Ĥ and H, we can decompose Ĥ into H and C; (2.) Ĥ = = where, by (2.9) 2 N(N +) 2 N(N +) n i (V i. V.. ) 2 n i ((V i. EV i. )+(EV i. V.. )) 2 = H +C, (2.2) [ ] 2 C := n i (EV i. V.. ) 2 + 2n i (EV i. V.. )(V i. EV i. ) N(N +) [ ] 2 = n i (EV i. V.. ) 2 + 2n i (EV i. V.. )(R i. ER i. ). N(N +) So by Theorem, to prove Theorem 2 it suffices to prove that under the conditions of Theorem 2, C in probability, or (2.3) EC, (2.4) Var(C), and to prove Propositions and 2 we just need to show that under the conditions of Propositions and 2, (2.5) EC, (2.6) Var(C). By (2.7), (2.8) and Lemma 2, (2.7) EV i. = n i E n i j= R ij n i E j= L ij = N n i. 2
11 KRUSKAL-WALLIS ONE-WAY ANALYSIS 7 By (2.7), (2.7) and (2.8), we see that EV i. V.. is actually a deterministic constant; (2.8) EV i. V.. = N n i 2 N n s V st = N s= t= s= n 2 s 2 n i 2. 2 k By (2.2) and (2.8), the random part of C is just N(N+) 2n i(ev i. V.. )R i.. Again, by (2.8) we can rewrite the major part k 2n i(ev i. V.. )R i. of the random part of C as (2.9) 2n i (EV i. V.. )R i. = = ( ) n 2 s 2n i N 2 n i 2 s= ( ) n 2 s N n i R i.. Takethe expectationand varianceonthe both sidesof(2.9). Then, bylemma 2 we have ( ) ( ) (2.2) E 2n i (EV i. V.. )R i. = n 2 s N n i E(R i. ) =, and (2.2) ( ) ( ( ) ) Var 2n i (EV i. V.. )R i. = Var n 2 s n i R i. N s= ( = N + ) 2 N n 3 i n 2 s. 2 So, by (2.2), (2.8), (2.2), and (2.2) we have (2.22) E(C) = 3f(N), (2.23) Var(C) = 2f(N), where (2.24) f(n) := N N 2 (N +) s= s= ( n 3 i s= n 2 s ) 2 Since n i = N k +f i(n) and since k f i(n) =, we have (2.25) f(n) = N N 2 (N +) ( ) 3 N k +f i(n) (. s= R i. ( ) ) 2 2 N k +f i(n)
12 72 Y. HONG AND S. LEE = N(N +) fi 3 (N)+ k(n +) ( 2 fi (N)) 2. N 2 (N +) fi 2 (N) Proof of Theorem 2. By dropping the last negative term in the right hand side of (2.25), we have (2.26) f(n) N(N +) fi 3 (N)+ k(n +) fi 2 (N). If by (.6) for all i, f i (N) = O(N β ), β < 2, by (2.26) as N f(n), and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.3) and (2.4) Theorem 2 follows. Proof of Proposition. By (.2) the first term in the right hand side of (2.25) is bounded by k N f2 i (N); ( ) (2.27) fi 3 (N) = o fi 2 (N). N(N +) N Again by (.2) (2.28) k N 2 fi 2 (N) k. So, by (2.28) among the second and third terms in the right hand side of (2.25) the second is the major term; ( ) 2 fi 2 k(n +) (N) (2.29) N 2 fi 2 (N +) (N) ( )( ) = fi 2 N + (N) k N 2 fi 2 (N). Combining(2.27),(2.28),and(2.29)weseethatf(N)isoforder N k f2 i (N); ( ) (2.3) f(n) = Ω fi 2 (N). N If by (.9) for some i, f i (N) = Ω(N βi ), 2 < β i <, by (2.3) as N f(n),
13 KRUSKAL-WALLIS ONE-WAY ANALYSIS 73 and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.5) and (2.6) Proposition follows. Proof of Proposition 2. Under the conditions(.2) and(.3) we have unequal < λ i < with k λ i =. For these λ i we consider a random variable X with P(X = λ i ) = λ i for i k. Let s calculate its variance. Since EX = k λ2 i and since EX2 = k we have 2 (2.3) Var(X) = EX 2 (EX) 2 = λi) 2. λ 3 i ( λ3 i, Since X is not a constant but finite random variable, it has a non-zero but finite variance. Therefore, by (2.3) we see that, for unequal < λ i < with k λ i = (2.32) < ( 2 λ 3 i λi) 2 <. By (2.24) and (2.32), as N ( 2 f(n) = N λ 3 i λi) 2 +o(), and hence by (2.22)-(2.24) both the mean and the variance of C go to. Therefore, by (2.5) and (2.6) Proposition 2 follows. 3. Proof of Theorem 3 In this section we use Lyapunov s CLT and prove Theorem 3. By (2.7) and (2.8) we can rewrite S N as (3.) S N = n i j= N + N n i R ij N + n i (n i +) 2(N n i ). Since {U ij := Rij N+ } have the following (joint) probability mass function ( P U ij = k ) = for k N, N + N ( P U ij = k ) N +,U i j = k = for k k N, N + N(N ) {U ij } have the following mean, variance, and covariance (3.2) EU ij = 2,
14 74 Y. HONG AND S. LEE (3.3) Var(U ij ) = 2N + 6N +6 4 = 2 6 N +O(N 2 ), (3.4) Cov(U ij,u i j ) = 3N2 N 2 2(N +)(N ) 4 = 2 N +O(N 2 ). Since N+ N n i λ i and since k n i(n i+) 2N(N n i) is non-random, we see that the asymptotics of S N is the same as that of T N, where T N is defined as n i (3.5) T N = A i U ij, where j= (3.6) A i = λ i. By (3.3), (3.4), and (3.6), n i (3.7) Var(T N ) = Var A i U ij = ij A 2 i ij [ = N 2 = N 2 j= ( 2 6 N +O(N 2 ) i j ij A i A i λ i ( λ i ) 2 λ i ( λ i ) 2 ) ( ) 2 N +O(N 2 ) = i = ( λ i λ i ( λ i )( λ i ) +O(N ) λ i λ i ) 2 +O(N ) Under the conditions (.5) and (.6) we have unequal < λ i < with k λ i =. For these λ i we consider a random variable Y with ( P Y = ) = λ i for i k. λ i Let s calculate its variance. Since EY = k λ i/( λ i ) and since EY 2 = k λ i/( λ i ) 2, we have ( ) 2 (3.8) Var(Y) = EY 2 (EY) 2 λ i = ( λ i ) 2 λ i. λ i Since by (.6), Y is not a constant but finite random variable, it has a nonzero but finite variance. Therefore, by (3.8) we see that, for unequal < λ i <. ]
15 with k λ i = (3.9) < KRUSKAL-WALLIS ONE-WAY ANALYSIS 75 λ i ( λ i ) 2 Therefore, by (3.7) and (3.9) we have ( λ i λ i (3.) Var(T N ) = Ω(N). ) 2 <. Since U ij, for any fixed δ >, E U ij EU ij 2+δ is bounded by. So, n i (3.) E A i U ij EA i U ij 2+δ = N j= n i A 2+δ i E U ij EU ij 2+δ j= [ λ i +o() ( λ i ) 2+δ ]. n i A 2+δ i By (3.) and (3.) one can easily check the Lyapunov condition; lim (Var(T n i N)) (+δ 2 ) E A i U ij EA i U ij 2+δ. n j= Hence, by Lyapunov s CLT Theorem 3 follows. References [] H. Chernoff and I. R. Savage, Asymptotic normality and efficiency of certain nonparametric test statistics, Ann. Math. Statist. 29 (958), [2] V. Dupac and J. Hajek, Asymptotic normality of simple linear rank statistics under alternative II., Ann. Math. Statist. 4 (969), [3] Z. Govindarajulu, L. Le Cam, and M. Raghavachari, Generalizations of theorems of Chernoff and Savage on the asymptotic normality of test statistics, Proc. Fifth Berkeley Sympos. Math. Statist. and Probability (Berkeley, Calif., 965/66) Vol. I: Statistics, pp Univ. California Press, Berkeley, Calif, 967. [4] J. Hajek, Asymptotic normality of simple linear rank statistics under alternatives, Ann. Math. Statist. 39 (968), [5] D. Kim, A class of distribution-free treatments versus control tests based on placements, Far East J. Theor. Stat. 3 (999), no., [6] D. Kim, S. Lee, and W. Wang, The asymptotic behavior of linear placement statistics, Statist. Probab. Lett. 8 (2), no. 2, [7] D. Kim and D. A. Wolfe, Properties of distribution-free two-sample procedures based on placements, Far East J. Math. Sci. (993), no. 2, [8] A. M. Mood, Introduction to the Theory of Statistics, McGraw-Hill, New York, 95. [9] J. Orban and D. A. Wolfe, A class of distribution-free two-sample tests based on placements, J. Amer. Statist. Assoc. 77 (982), no. 379, [] R. Pyke and G. R. Shorack, Weak convergence of a two-sample empirical process and a new approach to Chernoff-Savage theorems, Ann. Math. Statist. 39 (968),
16 76 Y. HONG AND S. LEE [] F. Wilcoxon, Individual comparisons by ranking methods, Biometrics (945), no. 6, Yicheng Hong Department of Mathematics Yanbian University Yanji 332, P. R. China and Department of Mathematics Yonsei University Seoul 2-749, Korea address: tophych@gmail.com Sungchul Lee Department of Mathematics Yonsei University Seoul 2-749, Korea address: sungchul@yonsei.ac.kr
Stat 710: Mathematical Statistics Lecture 31
Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:
More informationTESTS FOR LOCATION WITH K SAMPLES UNDER THE KOZIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississip
TESTS FOR LOCATION WITH K SAMPLES UNDER THE KOIOL-GREEN MODEL OF RANDOM CENSORSHIP Key Words: Ke Wu Department of Mathematics University of Mississippi University, MS38677 K-sample location test, Koziol-Green
More informationNonparametric hypothesis tests and permutation tests
Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationDr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines)
Dr. Maddah ENMG 617 EM Statistics 10/12/12 Nonparametric Statistics (Chapter 16, Hines) Introduction Most of the hypothesis testing presented so far assumes normally distributed data. These approaches
More informationNonparametric tests. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 704: Data Analysis I
1 / 16 Nonparametric tests Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I Nonparametric one and two-sample tests 2 / 16 If data do not come from a normal
More informationA NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES
Bull. Korean Math. Soc. 52 (205), No. 3, pp. 825 836 http://dx.doi.org/0.434/bkms.205.52.3.825 A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES Yongfeng Wu and Mingzhu
More informationUnit 14: Nonparametric Statistical Methods
Unit 14: Nonparametric Statistical Methods Statistics 571: Statistical Methods Ramón V. León 8/8/2003 Unit 14 - Stat 571 - Ramón V. León 1 Introductory Remarks Most methods studied so far have been based
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationON TWO RESULTS IN MULTIPLE TESTING
ON TWO RESULTS IN MULTIPLE TESTING By Sanat K. Sarkar 1, Pranab K. Sen and Helmut Finner Temple University, University of North Carolina at Chapel Hill and University of Duesseldorf Two known results in
More informationAdvanced Statistics II: Non Parametric Tests
Advanced Statistics II: Non Parametric Tests Aurélien Garivier ParisTech February 27, 2011 Outline Fitting a distribution Rank Tests for the comparison of two samples Two unrelated samples: Mann-Whitney
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend
More informationThis condition was introduced by Chandra [1]. Ordonez Cabrera [5] extended the notion of Cesàro uniform integrability
Bull. Korean Math. Soc. 4 (2004), No. 2, pp. 275 282 WEAK LAWS FOR WEIGHTED SUMS OF RANDOM VARIABLES Soo Hak Sung Abstract. Let {a ni, u n i, n } be an array of constants. Let {X ni, u n i, n } be {a ni
More informationComplete moment convergence of weighted sums for processes under asymptotically almost negatively associated assumptions
Proc. Indian Acad. Sci. Math. Sci.) Vol. 124, No. 2, May 214, pp. 267 279. c Indian Academy of Sciences Complete moment convergence of weighted sums for processes under asymptotically almost negatively
More informationLecture 28: Asymptotic confidence sets
Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level
More informationEmpirical Power of Four Statistical Tests in One Way Layout
International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo
More informationFinal Examination Statistics 200C. T. Ferguson June 11, 2009
Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationPSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests
PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution
More informationIMPROVING TWO RESULTS IN MULTIPLE TESTING
IMPROVING TWO RESULTS IN MULTIPLE TESTING By Sanat K. Sarkar 1, Pranab K. Sen and Helmut Finner Temple University, University of North Carolina at Chapel Hill and University of Duesseldorf October 11,
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationEmpirical Likelihood Tests for High-dimensional Data
Empirical Likelihood Tests for High-dimensional Data Department of Statistics and Actuarial Science University of Waterloo, Canada ICSA - Canada Chapter 2013 Symposium Toronto, August 2-3, 2013 Based on
More informationInferential Statistics
Inferential Statistics Eva Riccomagno, Maria Piera Rogantin DIMA Università di Genova riccomagno@dima.unige.it rogantin@dima.unige.it Part G Distribution free hypothesis tests 1. Classical and distribution-free
More informationON PITMAN EFFICIENCY OF
1. Summary ON PITMAN EFFICIENCY OF SOME TESTS OF SCALE FOR THE GAMMA DISTRIBUTION BARRY R. JAMES UNIVERSITY OF CALIFORNIA, BERKELEY A comparison is made of several two sample rank tests for scale change
More informationA COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky
A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),
More informationNonparametric Location Tests: k-sample
Nonparametric Location Tests: k-sample Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 04-Jan-2017 Nathaniel E. Helwig (U of Minnesota)
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationLecture 3 - Expectation, inequalities and laws of large numbers
Lecture 3 - Expectation, inequalities and laws of large numbers Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 3 - Expectation, inequalities and laws of large numbersapril 19, 2009 1 / 67 Part
More informationA Signed-Rank Test Based on the Score Function
Applied Mathematical Sciences, Vol. 10, 2016, no. 51, 2517-2527 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.66189 A Signed-Rank Test Based on the Score Function Hyo-Il Park Department
More informationSummary of Chapters 7-9
Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationYunhi Cho and Young-One Kim
Bull. Korean Math. Soc. 41 (2004), No. 1, pp. 27 43 ANALYTIC PROPERTIES OF THE LIMITS OF THE EVEN AND ODD HYPERPOWER SEQUENCES Yunhi Cho Young-One Kim Dedicated to the memory of the late professor Eulyong
More informationf (1 0.5)/n Z =
Math 466/566 - Homework 4. We want to test a hypothesis involving a population proportion. The unknown population proportion is p. The null hypothesis is p = / and the alternative hypothesis is p > /.
More informationHan-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek
J. Korean Math. Soc. 41 (2004), No. 5, pp. 883 894 CONVERGENCE OF WEIGHTED SUMS FOR DEPENDENT RANDOM VARIABLES Han-Ying Liang, Dong-Xia Zhang, and Jong-Il Baek Abstract. We discuss in this paper the strong
More informationSHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan
A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,
More informationSTAT 135 Lab 8 Hypothesis Testing Review, Mann-Whitney Test by Normal Approximation, and Wilcoxon Signed Rank Test.
STAT 135 Lab 8 Hypothesis Testing Review, Mann-Whitney Test by Normal Approximation, and Wilcoxon Signed Rank Test. Rebecca Barter March 30, 2015 Mann-Whitney Test Mann-Whitney Test Recall that the Mann-Whitney
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More informationECON Fundamentals of Probability
ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,
More informationInstitute of Actuaries of India
Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the
More informationBivariate Paired Numerical Data
Bivariate Paired Numerical Data Pearson s correlation, Spearman s ρ and Kendall s τ, tests of independence University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html
More informationWLLN for arrays of nonnegative random variables
WLLN for arrays of nonnegative random variables Stefan Ankirchner Thomas Kruse Mikhail Urusov November 8, 26 We provide a weak law of large numbers for arrays of nonnegative and pairwise negatively associated
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More information1 Probability theory. 2 Random variables and probability theory.
Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major
More informationUniversität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Hypothesis testing. Anna Wegloop Niels Landwehr/Tobias Scheffer
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Hypothesis testing Anna Wegloop iels Landwehr/Tobias Scheffer Why do a statistical test? input computer model output Outlook ull-hypothesis
More informationLecture 2: Review of Probability
Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................
More information6 The normal distribution, the central limit theorem and random samples
6 The normal distribution, the central limit theorem and random samples 6.1 The normal distribution We mentioned the normal (or Gaussian) distribution in Chapter 4. It has density f X (x) = 1 σ 1 2π e
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationConsistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis
Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,
More informationMultiple Regression Analysis
Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,
More informationNEW CONSTRUCTION OF THE EAGON-NORTHCOTT COMPLEX. Oh-Jin Kang and Joohyung Kim
Korean J Math 20 (2012) No 2 pp 161 176 NEW CONSTRUCTION OF THE EAGON-NORTHCOTT COMPLEX Oh-Jin Kang and Joohyung Kim Abstract The authors [6 introduced the concept of a complete matrix of grade g > 3 to
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationThe symmetry in the martingale inequality
Statistics & Probability Letters 56 2002 83 9 The symmetry in the martingale inequality Sungchul Lee a; ;, Zhonggen Su a;b;2 a Department of Mathematics, Yonsei University, Seoul 20-749, South Korea b
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationThe Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility
The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory
More informationClass 8 Review Problems solutions, 18.05, Spring 2014
Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots
More informationAnalysis of Experimental Designs
Analysis of Experimental Designs p. 1/? Analysis of Experimental Designs Gilles Lamothe Mathematics and Statistics University of Ottawa Analysis of Experimental Designs p. 2/? Review of Probability A short
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationPRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS
PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS REZA HASHEMI and MOLUD ABDOLAHI Department of Statistics, Faculty of Science, Razi University, 67149, Kermanshah,
More informationIntroduction to Nonparametric Statistics
Introduction to Nonparametric Statistics by James Bernhard Spring 2012 Parameters Parametric method Nonparametric method µ[x 2 X 1 ] paired t-test Wilcoxon signed rank test µ[x 1 ], µ[x 2 ] 2-sample t-test
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More information. Find E(V ) and var(v ).
Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number
More informationMultivariate Time Series: Part 4
Multivariate Time Series: Part 4 Cointegration Gerald P. Dwyer Clemson University March 2016 Outline 1 Multivariate Time Series: Part 4 Cointegration Engle-Granger Test for Cointegration Johansen Test
More informationIntermediate Econometrics
Intermediate Econometrics Heteroskedasticity Text: Wooldridge, 8 July 17, 2011 Heteroskedasticity Assumption of homoskedasticity, Var(u i x i1,..., x ik ) = E(u 2 i x i1,..., x ik ) = σ 2. That is, the
More informationShih-sen Chang, Yeol Je Cho, and Haiyun Zhou
J. Korean Math. Soc. 38 (2001), No. 6, pp. 1245 1260 DEMI-CLOSED PRINCIPLE AND WEAK CONVERGENCE PROBLEMS FOR ASYMPTOTICALLY NONEXPANSIVE MAPPINGS Shih-sen Chang, Yeol Je Cho, and Haiyun Zhou Abstract.
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationUCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7
UCSD ECE50 Handout #4 Prof Young-Han Kim Wednesday, June 6, 08 Solutions to Exercise Set #7 Polya s urn An urn initially has one red ball and one white ball Let X denote the name of the first ball drawn
More informationEconometrics Multiple Regression Analysis: Heteroskedasticity
Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties
More informationConditional moment representations for dependent random variables
Conditional moment representations for dependent random variables W lodzimierz Bryc Department of Mathematics University of Cincinnati Cincinnati, OH 45 22-0025 bryc@ucbeh.san.uc.edu November 9, 995 Abstract
More informationA REVERSE TO THE JEFFREYS LINDLEY PARADOX
PROBABILITY AND MATHEMATICAL STATISTICS Vol. 38, Fasc. 1 (2018), pp. 243 247 doi:10.19195/0208-4147.38.1.13 A REVERSE TO THE JEFFREYS LINDLEY PARADOX BY WIEBE R. P E S T M A N (LEUVEN), FRANCIS T U E R
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationSome Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2
STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square
More informationTHE EXPONENTIAL DISTRIBUTION ANALOG OF THE GRUBBS WEAVER METHOD
THE EXPONENTIAL DISTRIBUTION ANALOG OF THE GRUBBS WEAVER METHOD ANDREW V. SILLS AND CHARLES W. CHAMP Abstract. In Grubbs and Weaver (1947), the authors suggest a minimumvariance unbiased estimator for
More informationOrthogonal decompositions in growth curve models
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 4, Orthogonal decompositions in growth curve models Daniel Klein and Ivan Žežula Dedicated to Professor L. Kubáček on the occasion
More informationChapter 11 Sampling Distribution. Stat 115
Chapter 11 Sampling Distribution Stat 115 1 Definition 11.1 : Random Sample (finite population) Suppose we select n distinct elements from a population consisting of N elements, using a particular probability
More informationUSING A BIMODAL KERNEL FOR A NONPARAMETRIC REGRESSION SPECIFICATION TEST
Statistica Sinica 25 (2015), 1145-1161 doi:http://dx.doi.org/10.5705/ss.2014.008 USING A BIMODAL KERNEL FOR A NONPARAMETRIC REGRESSION SPECIFICATION TEST Cheolyong Park 1, Tae Yoon Kim 1, Jeongcheol Ha
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationStatistics (1): Estimation
Statistics (1): Estimation Marco Banterlé, Christian Robert and Judith Rousseau Practicals 2014-2015 L3, MIDO, Université Paris Dauphine 1 Table des matières 1 Random variables, probability, expectation
More informationContents. Acknowledgments. xix
Table of Preface Acknowledgments page xv xix 1 Introduction 1 The Role of the Computer in Data Analysis 1 Statistics: Descriptive and Inferential 2 Variables and Constants 3 The Measurement of Variables
More informationNegative Multinomial Model and Cancer. Incidence
Generalized Linear Model under the Extended Negative Multinomial Model and Cancer Incidence S. Lahiri & Sunil K. Dhar Department of Mathematical Sciences, CAMS New Jersey Institute of Technology, Newar,
More informationA Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms. by R. L. Eubank and Suojin Wang Texas A&M University
A Central Limit Theorem for the Sum of Generalized Linear and Quadratic Forms by R. L. Eubank and Suojin Wang Texas A&M University ABSTRACT. A central limit theorem is established for the sum of stochastically
More informationFunctions of Several Random Variables (Ch. 5.5)
(Ch. 5.5) Iowa State University Mar 7, 2013 Iowa State University Mar 7, 2013 1 / 37 Outline Iowa State University Mar 7, 2013 2 / 37 several random variables We often consider functions of random variables
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationSimple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationSTA 4322 Exam I Name: Introduction to Statistics Theory
STA 4322 Exam I Name: Introduction to Statistics Theory Fall 2013 UF-ID: Instructions: There are 100 total points. You must show your work to receive credit. Read each part of each question carefully.
More informationDeccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY. SECOND YEAR B.Sc. SEMESTER - III
Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY SECOND YEAR B.Sc. SEMESTER - III SYLLABUS FOR S. Y. B. Sc. STATISTICS Academic Year 07-8 S.Y. B.Sc. (Statistics)
More informationRank-sum Test Based on Order Restricted Randomized Design
Rank-sum Test Based on Order Restricted Randomized Design Omer Ozturk and Yiping Sun Abstract One of the main principles in a design of experiment is to use blocking factors whenever it is possible. On
More informationSummary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016
8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying
More informationRecall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n
Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible
More informationNONPARAMETRICS. Statistical Methods Based on Ranks E. L. LEHMANN HOLDEN-DAY, INC. McGRAW-HILL INTERNATIONAL BOOK COMPANY
NONPARAMETRICS Statistical Methods Based on Ranks E. L. LEHMANN University of California, Berkeley With the special assistance of H. J. M. D'ABRERA University of California, Berkeley HOLDEN-DAY, INC. San
More informationMark Scheme (Results) Summer 2007
Mark Scheme (Results) Summer 007 GCE GCE Mathematics Statistics S3 (6691) Edexcel Limited. Registered in England and Wales No. 4496750 Registered Office: One90 High Holborn, London WC1V 7BH June 007 6691
More informationChapter 18 Resampling and Nonparametric Approaches To Data
Chapter 18 Resampling and Nonparametric Approaches To Data 18.1 Inferences in children s story summaries (McConaughy, 1980): a. Analysis using Wilcoxon s rank-sum test: Younger Children Older Children
More informationNon-Parametric Statistics: When Normal Isn t Good Enough"
Non-Parametric Statistics: When Normal Isn t Good Enough" Professor Ron Fricker" Naval Postgraduate School" Monterey, California" 1/28/13 1 A Bit About Me" Academic credentials" Ph.D. and M.A. in Statistics,
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth
More information