The Distribution of F
|
|
- Paula Barber
- 5 years ago
- Views:
Transcription
1 The Distribution of F It can be shown that F = SS Treat/(t 1) SS E /(N t) F t 1,N t,λ a noncentral F-distribution with t 1 and N t degrees of freedom and noncentrality parameter λ = t i=1 n i(µ i µ) 2 /σ 2 = t i=1 n i j=1 (µ i µ) 2 /σ 2 Under H 0 : µ 1 =... = µ t this becomes the (central) F t 1,N t distribution. We reject H 0 whenever F F t 1,N t (1 α) = Fcrit = qf(1 α,t 1,N t) which denotes the (1 α)-quantile of the F t 1,N t distribution. Power function: β(λ) = P(F F t 1,N t (1 α)) = 1 pf(fcrit,t 1,N t,λ) 1
2 Discussion of Noncentrality Parameter λ The power of the ANOVA F-test is a monotone increasing function of λ = t i=1 n i(µ i µ) 2 /σ 2 = N t i=1 (n i/n)(µ i µ) 2 /σ 2 = N σ 2 µ/σ 2 = N between treatment variation/within treatment variation Thus we consider the drivers in λ. λ increases as σ decreases (provided the µ i are not all the same). The more difference there is between the treatment means µ i the higher λ Increasing the sample sizes will magnify n i (µ i µ) 2 (provided µ is stable). The sample sizes we can plan for. Later: we can reduce σ by blocking units into more homogeneous groups. 2
3 Optimal Allocation of Sample Sizes? We have N experimental units available for testing the effects of t treatments and suppose that N is a multiple of t, say N = r t (r and t integer). It would seem best to use samples of equal size r for each of the t treatments i.e., we would opt for a balanced design. That way we would not emphasize one treatment over any of the others. Is there some other optimality criterion that could be used as justification? We may plan for a balanced design upfront, but then something goes wrong with a few observations and they have to be discarded from analysis. Be careful that the deletion of observations does not bias any conclusions. 3
4 A Sample Size Allocation Rationale We may be concerned with alternatives where all means but one are the same. Since we won t know upfront which mean sticks out, we would want to maximize the minimum power against all such contingencies. Max-Min Strategy! If µ 1 = µ + and µ 2 =... = µ t = µ then µ = µ + n 1 /N and (algebra) t λ 1 = n i (µ i µ) 2 /σ 2 = N 2 n ( 1 i=1 σ 2 1 n ) 1 N N for the other cases. It is easy to see now that for fixed σ max min n 1,...,n t 1 i t [ N 2 σ 2 is achieved when n 1 =... = n t. That is because n ( i 1 n ) i N N increases for n i /N 1/2. n i N and similarly ( 1 n ) ] i N λ i = N 2 σ 2 n i N ( 1 n ) i N 4
5 Using sample.sizeanova (see web page) Suppose we have t = 3 treatments and want to determine the sample size n per treatment to achieve power β(λ) =.9 for level α =.05. It is desired to do this for a λ = λ i corresponding to the alternatives on the previous slide with /σ =.5, i.e., with N = t n λ i = N 2 σ 2 n ( 1 n ) ( = n 2 N N σ ) = n 2 t σ 2 t 1 = n λ t 0. λ 0 = ( 2 /σ 2 ) (t 1)/t can be interpreted more generally as (µ i µ) 2 /σ 2. > sample.sizeanova() > sample.sizeanova(nrange=30:100) > sample.sizeanova(nrange=70:100,power0=.9) produced the next 3 slides = n = 77. 5
6 sample.sizeanova I function (delta.per.sigma=.5,t.treat=3, nrange=2:30,alpha=.05,power0=null) { # delta.per.sigma is the ratio of delta over sigma # for which one wants to detect a delta shift in one # mean while all other means stay the same. # t.treat is the number of treatments. alpha is the # desired significance level. nrange is a range of # sample sizes over which the power will be calculated # for that delta.per.sigma. power0 is on optional value # for the target power that will be highlighted on the plot. # lambda0=((t.treat-1)/t.treat)*delta.per.sigmaˆ2 power=null 6
7 sample.sizeanova II for(n in nrange){ N=n*t.treat Fcrit=qf(1-alpha,t.treat-1,N-t.treat) power=c(power,1-pf(fcrit,t.treat-1,n-t.treat,n*lambda0))} plot(nrange,power,type="l",xlab=paste("sample size n per each of t =", t.treat," treatments"), ylab="",ylim=c(0,1)) mtext(expression(beta(lambda) "=" beta(n %*% lambda[0])),2,2.7) abline(h=seq(0,1,.02),col="grey") abline(v=nrange,col="grey") lines(nrange,power,col="red") title(substitute(delta/sigma==delta.per.sigma "," lambda[0] "=" sum((mu[i]-bar(mu))ˆ2/sigmaˆ2) "=" lambda0 ", " alpha==alpha1, list(lambda0=format(signif(lambda0,4)),alpha1=alpha, delta.per.sigma=delta.per.sigma))) 7
8 sample.sizeanova III if(!is.null(power0)){ abline(h=power0,col="blue") par(las=2) mtext(power0,4,0.2,at=power0,col="blue") par(las=0)}} 8
9 Sample Size Determination σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 9
10 Sample Size Determination (increased n) σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 10
11 Sample Size Determination (magnified) σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 11
12 Coagulation Example In order to understand the blood coagulation behavior in relation to various diets, lab animals were given 4 different diets and their subsequent blood draws were then measured for their respective coagulation times in seconds. The lab animals were assigned randomly to the various diets. The results were as follows: > ctime [1] [19] > diet [1] "A" "A" "A" "A" "B" "B" "B" "B" "B" "B" "C" "C" "C" [14] "C" "C" "C" "D" "D" "D" "D" "D" "D" "D" "D" 12
13 Plot for Coagulation Example n A = 4 n B = 6 n C = 6 n D = 8 coagulation time (sec) A A A A B B B B B B C C C C D D D D D D D diet 13
14 ANOVA for Coagulation Example Note that in the previous plot we used jitter(ctime) to plot ctime in the vertical direction and to plot its horizontal mean lines. This perturbs tied observations a small random amount to make tied observations more visible. For example, the mean lines for diet A and D would have been the same otherwise. > anova(lm(ctime as.factor(diet))) Analysis of Variance Table Response: ctime Df Sum Sq Mean Sq F value Pr(>F) as.factor(diet) e-05 *** Residuals Signif. codes: 0 *** ** 0.01 *
15 lm for Coagulation Example > out=lm(ctime as.factor(diet)) > names(out) [1] "coefficients" "residuals" "effects" [4] "rank" "fitted.values" "assign" [7] "qr" "df.residual" "contrasts" [10] "xlevels" "call" "terms" [13] "model" > out$coefficients (Intercept) as.factor(diet)b as.factor(diet)c e e e+00 as.factor(diet)d e-14 Note that these are the estimates ˆµ A, ˆµ B ˆµ A, ˆµ C ˆµ A, ˆµ D ˆµ A. 15
16 Residuals from lm for Coagulation Example > out$residuals e e e e e e e e e e e e e e e e e e e e e e e e+00 Numbers such as e-17 should be treated as 0 (computing quirks). 16
17 Fitted Values from lm for Coagulation Example > out$fitted.values
18 Comparing Treatment Means Ȳ i. When the hypothesis H 0 : µ 1 =... = µ t is not rejected at level α then there is little purpose in looking closer at differences between the sample means Ȳ i. for the various treatments. Any such perceived differences could easily have come about by simple random variation, even when the hypothesis is true. Why then read something into randomness? It is like reading tea leaves! However, when the hypothesis is rejected it is quite natural to ask in which way the hypothesis was contradicted. The best indicators for any analysis as to how the means µ i may be different would be the sample or treatment means ˆµ i = Ȳ i., i = 1,...,t. 18
19 Confidence Intervals for µ i A first step in understanding differences in the µ i is to look at their estimates ˆµ i =Ȳ i.. We should do this in the context of the sampling variability of ˆµ i. In the past we addressed this via confidence intervals for µ i based on ˆµ i. In any such confidence interval we can now use the pooled variance s 2 from all t samples and not just the variance s 2 i from the ith sample, i.e. we get ˆµ i ±t 1 α/2,n t s ni as our 100(1 α)% confidence interval for µ i. This follows as before (exercise) from the independence of ˆµ i and s, the fact that (ˆµ i µ i )/(σ/ n i ) N (0,1), and from s 2 /σ 2 χ 2 N t /(N t). The validity of this improvement (N t n i 1 when using s 2 instead of s 2 i ) depends strongly on the assumption that the population variances σ 2 behind all t samples are the same, or at least approximately so. 19
20 Plots of Confidence Intervals for Coagulation Data coagulation time (sec) using pooled s 2 = using individual s i 2 t i=1 s i2 (n i 1) (N t) diet A diet B diet C diet D 20
21 Simultaneous Confidence Intervals When constructing intervals of the type: ˆµ i ±t 1 α/2 s ni or ˆµ i ±t 1 α/2 s i ni for i = 1,...,t we should be aware that these intervals don t simultaneously cover their respective targets µ i with probability 1 α. They do so individually. For example ( ) s t ( ) P µ i ˆµ i ±t i s 1 α/2, i = 1,...,t = ni P µ i ˆµ i ±t i 1 α/2 i=1 ni = (1 α) t < 1 α. Thus we should choose α for individual intervals to get (1 α ) t = 1 α or α = 1 (1 α) 1/t α t = α t. Some problem remains when using common pooled estimate s. No independence! 21
22 α = 1 (1 α) 1/t α/t or α ~ t = α t α t * = 1 (1 α) (1 t) α * (1 t) t = 1 (1 α) α ~ t = α t t = 3 t = 5 t = 6 t = desired overall α 22
23 Dealing with Dependence from Using Pooled s When we use a common pooled estimate s for the standard deviation σ the previous confidence intervals are no longer independent. However, it can be shown that ( ) s P µ i ˆµ i ±t 1 α /2, i = 1,...,t ni t ( ) s P µ i ˆµ i ±t 1 α /2 i=1 ni = (1 α ) t = 1 α This comes from the positive dependence between confidence intervals through s, i.e., if one interval is more (less) likely to cover its target µ i due to s, so are the other intervals more (less) likely to cover their targets µ j. Using the same compensation as in the independence case would let us err on the conservative side, i.e., give us higher confidence than the targeted 1 α. 23
24 Boole s and Bonferroni s Inequality For any m events E 1,...,E m Boole s inequality states P(E 1... E m ) P(E 1 ) P(E m ) For any m events E 1,...,E m Bonferroni s inequality states P(E 1... E m ) 1 m (1 P(E i )) i=1 The statement are equivalent, since P(E 1... E m ) = 1 P(E c 1... Ec m). If E i denotes the i th coverage event { µ i ˆµ i ±t 1 α/2 s ni } with P(E i ) = 1 α, then the simultaneous coverage probability is bounded from below as follows ( ) \ t t P E i 1 (1 P(E i )) = 1 t α = 1 α if α = α t = α/t, i=1 i=1 i.e., we can achieve at least 1 α probability coverage by choosing the individual coverage appropriately, namely 1 α = 1 α/t. Almost same adjustment. 24
25 Contrasts Any linear combination C = t i=1 c iµ i with t i=1 c i = 0 is called a contrast. Note that t i=1 c iµ = 0, i.e., contrasts are zero over the hypothesis. Suppose we have 4 treatments with respective means µ 1,...,µ 4. We may be interested in contrasts of the following form C 12 = µ 1 µ 2 with c = (c 1,...,c 4 ) = (1, 1,0,0). Similarly for the other differences C i j = µ i µ ( 4) j. There are 2 = 6 such contrasts. Sometimes one of the treatments, say the first, is singled out as the control. We may then be interested in just the 3 contrasts C 12,C 13 and C 14 or we may be interested in C = µ 1 (µ 2 + µ 3 + µ 4 )/3 with c = (1, 1/3, 1/3, 1/3). Sometimes the first 2 treatment share something in common and so do the last 2. One might then try: C = (µ 1 +µ 2 )/2 (µ 3 +µ 4 )/2 with c = (1/2,1/2, 1/2, 1/2) 25
26 Estimates and Confidence Intervals for Contrasts A natural estimate of C = t i=1 c iµ i is Ĉ = t i=1 c iˆµ i = t i=1 c iȳi.. ( ) t t t We have E(Ĉ) = E c i Ȳ i. = c i E (Ȳ i.) = c i µ i = C i=1 i=1 i=1 t and var(ĉ) = var( c i Ȳ i. i=1 Under the normality assumption for the Y i j we have ) = Ĉ C t N t where s 2 = s t i=1 c2 i /n i t c 2 i var(ȳ i.) = i=1 t c 2 i σ2 /n i. i=1 t n i s 2 i /(N t) = MS E. i=1 = Ĉ±t N t,1 α/2 s t c 2 i /n i is a 100(1 α)% confidence interval for C. i=1 26
27 Testing H 0 : C = 0 Based on the duality of testing and confidence intervals we can test the hypothesis H 0 : C = 0 by rejecting it whenever the previous confidence interval does not contain C = 0. Similarly, reject H 0 : C = C 0 by rejecting it whenever the previous confidence interval does not contain C = C 0 Another notation for this interval is Ĉ ±t N t,1 α/2 SE(Ĉ) where SE(Ĉ) = s t c 2 i /n i. i=1 SE(Ĉ) is the standard error of Ĉ, the estimate of the standard deviation of Ĉ. 27
28 Paired Comparisons: Fisher s Protected LSD Method After rejecting H 0 : µ 1 =... = µ t one is often interested in looking at all ( t 2 ) pairwise contrasts C i j = µ i µ j. The following procedure is referred to as Fisher s Protected Least Significant Difference (LSD) Method. It consists of possibly two stages: 1) Perform α level F-test for testing H 0. If H 0 is not rejected, stop. 2) If H ( t ) 0 is rejected, form all 2 (1 α)-level confidence intervals for Ci j = µ i µ j : 1 Î i j = ˆµ i ˆµ j ±t N t,1 α/2 s + 1 n i n j and declare all µ i µ j 0 for which Î i j does not contain zero. 28
29 Comments on Fisher s Protected LSD Method If H 0 is true, the chance of making any statements contradicting H 0 is at most α. This is the protected aspect of this procedure. However, when H 0 is not true there are many possible contingencies, some of which can give us a higher than desired chance of pronouncing a significant difference, when in fact there is none. E.g., if all but one mean (say µ 1 ) are equal and µ 1 is far away from µ 2 =... = µ t our chance of rejecting H 0 is almost 1. However, among the intervals for µ i µ j, 2 i < j we may find a significantly higher than α proportion of cases with wrongly declared differences. This is due to the multiple comparison issue. 29
30 Pairwise Comparisons: Tukey-Kramer Method The Tukey-Kramer method is based on the distribution of { } Zi Z j i.i.d. Q t, f = max where Z 1 i< j t s 1,...,Z t N (0,1) and f s 2 χ 2 f Its cdf and quantile function are given in R as ptukey(q,nmeans,df) and qtukey(p,nmeans,df), nmeans = t is the number of means, df = f = N t denotes the degrees of freedom in s. Applying this to Z i = (ˆµ i µ i )/(σ/ n) and assuming n 1 =... = n t = n we get max i< j { } n ˆµi ˆµ j (µ i µ j ) s = max i< j ˆµ i µ i σ/ n ˆµ j µ j σ/ n s/σ = Q t, f P ( µ i µ j ˆµ i ˆµ j ± q t, f,1 α s/ n i < j ) = 1 α simultaneous (1 α)-coverage confidence intervals. Here P(Q t, f q t, f,1 α ) = 1 α or q t, f,1 α = qtukey(1 α,t,f). 30
31 Tukey-Kramer Method: Unequal Sample Sizes The simultaneous intervals for all pairwise mean differences was due to Tukey, but it was hampered by the requirement of equal sample sizes. This was addressed by Kramer in the following way. In the above confidence intervals replace n in 1/ n = 1/n by n i j, where n i j is the harmonic mean of n i and n j, i.e., 1/n i j = (1/n i +1/n j )/2. Different adjustment for each pair (i, j)! It was possible to show ( ) P µ i µ j ˆµ i ˆµ j ± Q t, f,1 α s/ n i j i < j 1 α simultaneous confidence intervals with coverage 1 α. 31
32 Tukey-Kramer Method for Coagulation Data coag.tukey = function (alpha=.05) { diets=unique(diet) mu.vec=null nvec=null mean.vec=null for(i in 1:length(diets)){ mu.vec=c(mu.vec,mean(ctime[diet==diets[i]])) nvec=c(nvec,length(ctime[diet==diets[i]])) mean.vec=c(mean.vec,rep(mu.vec[i],nvec[i])) } tr=length(nvec) N=sum(nvec) MSE=sum((ctime-mean.vec)ˆ2/(N-tr)) 32
33 Tukey-Kramer Method for Coagulation Data s=sqrt(mse) intervals=null for(i in 1:3){ for(j in (i+1):4){ nijstar=1/(.5*(1/nvec[i]+1/nvec[j])) qtk=qtukey(1-alpha,tr,n-tr) Diff=mu.vec[i]-mu.vec[j] lower=diff - qtk*s/sqrt(nijstar) upper=diff + qtk*s/sqrt(nijstar) intervals=rbind(intervals,c(lower,upper)) } } intervals } 33
34 Tukey-Kramer Results for Coagulation Data > coag.tukey() [,1] [,2] [1,] [2,] [3,] [4,] [5,] [6,] Declare significant differences in µ 1 µ 2, µ 1 µ 3, µ 2 µ 4, and µ 3 µ 4. 34
35 Scheffé s Confidence Intervals for All Contrasts Scheffé took the F-test for testing H 0 : µ 1 =... = µ t and converted it into a simultaneous coverage statement about confidence intervals for all contrasts c = (c 1,...,c t ): P t c i µ i i=1 ( t t c iˆµ i ± (t 1) F t 1,N t,1 α s i=1 i=1 1/2 c 2 i i) /n c = 1 α This is a coverage statement about an infinite number of contrasts, but can be applied conservatively to all pairwise contrasts. The resulting intervals tend to be quite conservative. But it compares well with Bonferroni type intervals if applied to many contrasts. 35
36 Pairwise Comparison Intervals for Coagulation Data (simultaneous) 95%-Intervals mean Fisher s Bonferroni Scheffé s all difference Tukey-Kramer protected LSD inequality contrasts method µ 1 µ µ 1 µ µ 1 µ µ 2 µ µ 2 µ µ 3 µ Declare significant differences in µ 1 µ 2, µ 1 µ 3, µ 2 µ 4, and µ 3 µ 4, using any of the four methods. 36
37 Simultaneous Paired Comparisons (95%) Pairwise Comparisons of Means (Coagulation Data): 1 α = Tukey Kramer pairwise comparisons Fisher's protected LSD Bonferroni intervals Scheffe's intervals for all contrasts µ 1 µ 2 µ 1 µ 3 µ 1 µ 4 µ 2 µ 3 µ 2 µ 4 µ 3 µ 4 37
38 Simultaneous Paired Comparisons (99%) Pairwise Comparisons of Means (Coagulation Data): 1 α = Tukey Kramer pairwise comparisons Fisher's protected LSD Bonferroni intervals Scheffe's intervals for all contrasts µ 1 µ 2 µ 1 µ 3 µ 1 µ 4 µ 2 µ 3 µ 2 µ 4 µ 3 µ 4 38
39 Orthogonal Contrast All ( t 2 ) pairwise comparisons for µi µ j could by very many and simultaneous intervals would become quite conservative. Since all these contrasts span a (t 1)-dimensional space one should be able to capture all differences with just t 1 orthogonal contrasts. C 1 = t c 1i µ i C 2 = i=1 t c 2i µ i i=1 C 1 C 2 = cov(ĉ 1,Ĉ 2 ) = t i=1 t c 1i c 2i /n i = 0 i=1 t c 1i c 2 j cov(ˆµ i, ˆµ j ) = j=1 ( ) wrong def. in Montgomery p.91 t c 1i c 2i σ 2 /n i = 0, i=1 i.e., Ĉ 1 and Ĉ 2 are independent and simultaneous statements for C 1,C 2,... are easier to handle, just as before when making simultaneous intervals for µ 1,...,µ t based on independent ˆµ 1,..., ˆµ t. The trick is to have meaningful or interpretable orthogonal contrast. 39
40 Simultaneous Intervals for Orthogonal Contrast If Ĉ 1,...,Ĉ k are any k orthogonal contrasts, they are independent. Just as we constructed simultaneous intervals for means based on independent mean estimates and the pooled standard deviation s we can again construct contrast confidence intervals with simultaneous coverage probability 1 α by taking 1 α with α = 1 (1 α) 1/k α k as the coverage probability for the individual intervals. 40
41 An Orthogonal Contrast Example Suppose we have t = 3 treatments of which the third is a control, i.e., we are familiar with its performance. Assume further that we have a balanced design, i.e., n 1 = n 2 = n 3. We could try the following t 1 = 2 orthogonal contrasts: c 1 = (.5,.5, 1) and c 2 = (1, 1,0). Note that C 1 = (µ 1 + µ 2 )/2 µ 3 and C 2 = µ 1 µ 2, of which the first assesses how much the average mean of the two new treatments differs from the control mean and the second assesses the difference between the two new treatments. These are seemingly orthogonal issues. 41
42 Unbalanced Case of Previous Example We have an unbalanced design, i.e., n 1, n 2, n 3 may be different. Then the following t 1 = 2 vectors: c 1 = (n 1/(n 1 + n 2 ),n 2 /(n 1 + n 2 ), 1) and c 2 = (1, 1,0) are indeed contrast vectors: n 1 /(n 1 + n 2 ) + n 2 /(n 1 + n 2 ) 1 = 0 and = 0 and they are orthogonal: n 1 /[(n 1 + n 2 )n 1 ] n 2 /[(n 1 + n 2 )n 2 ] 1 0/n 3 = 0. = C 1 = (n 1 µ 1 + n 2 µ 2 )/(n 1 + n 2 ) µ 3 = µ 12 µ 3 and C 2 = µ 1 µ 2, of which the first assesses how much the weighted average mean of the two new treatments differs from the control mean and the second assesses the difference between the two new treatments. These are seemingly orthogonal issues. 42
43 Service Center Data # of persons # of calls on call processed per hour number of calls processed per hour number of people on call 43
44 Service Center Data Here we have a new type of treatment (number of persons on call), where the different treatment levels are scalar and not just qualitative. In such situations the following orthogonal contrasts are of practical interest: c i1 c i2 c i3 c i4 c i5 C 1 = 5 j=1 c 1 j µ j C 2 = 5 j=1 c 2 j µ j C 3 = 5 j=1 c 3 j µ j C 4 = 5 j=1 c 4 j µ j For what kind of mean patterns in µ 1,...,µ 5 would C i and consequently Ĉ i be large? 44
45 Orthogonal Contrast Plots 5 C i = c i, j j using µ j = j j= C 1 C 2 C 3 C j 45
46 Interpretation of Orthogonal Contrast Plots The previous plot suggests that a pattern in the means µ j in relation to j = 1,...,5 that correlates most strongly with the corresponding pattern in the plot should yield a high value for the corresponding absolute contrast C i. Thus a large value C 1 indicates a strong linear component in the mean pattern. A large value C 2 indicates a strong quadratic component in the mean pattern. A large value C 3 indicates a strong cubic component in the mean pattern. A large value C 4 indicates a strong quartic component in the mean pattern. Typically, one hopes to rule out some (if not all) of the latter possibilities. 46
47 Simultaneous Contrast Intervals (for Service Center Data) 95% 99% C 1 [ 6.27, 11.58] [ 5.53, 12.32] C 2 [-7.02, -0.73] [ -7.89, 0.14] C 3 [-1.26, 4.06] [ -1.99, 4.79] C 4 [-9.58, 4.48] [-11.53, 6.43] From these intervals one sees that C 1 and C 2 are significantly different from zero. with 95% confidence, but C 2 not quite with 99% confidence. Hence there appears to be a sufficiently strong linear and mildly quadratic component. The original data plot suggested this and its strength is now assessed statistically. 47
48 Orthogonal Polynomial Contrast Vectors The previous orthogonal contrasts for linear, quadratic, cubic, quartic behavior were tailored to five treatments. How do we get similar contrast vectors when we have t treatments? R has a function contr.poly(t) that gives you orthogonal vectors representing the various polynomial components: linear, quadratic,... > round(contr.poly(7),3).l.q.c ˆ4 ˆ5 ˆ6 [1,] [2,] [3,] [4,] [5,] [6,] [7,]
49 Orthogonal Polynomial Contrasts from contr.poly(7) orthogonal polynomial i 49
Stat 502 Design and Analysis of Experiments One-Factor ANOVA
Stat 502 Design and Analysis of Experiments One-Factor ANOVA Fritz Scholz Department of Statistics, University of Washington January 7, 2014 2 One-Factor ANOVA ANOVA is an acronym for Analysis of Variance
More informationThese are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large?
Lecture 12 Comparing treatment effects Orthogonal Contrasts What use are contrasts? Recall the Cotton data In this case, the treatment levels have an ordering to them this is not always the case) Consider
More informationSTAT22200 Spring 2014 Chapter 5
STAT22200 Spring 2014 Chapter 5 Yibi Huang April 29, 2014 Chapter 5 Multiple Comparisons Chapter 5-1 Chapter 5 Multiple Comparisons Note the t-tests and C.I. s are constructed assuming we only do one test,
More informationLecture 5: Comparing Treatment Means Montgomery: Section 3-5
Lecture 5: Comparing Treatment Means Montgomery: Section 3-5 Page 1 Linear Combination of Means ANOVA: y ij = µ + τ i + ɛ ij = µ i + ɛ ij Linear combination: L = c 1 µ 1 + c 1 µ 2 +...+ c a µ a = a i=1
More informationMore about Single Factor Experiments
More about Single Factor Experiments 1 2 3 0 / 23 1 2 3 1 / 23 Parameter estimation Effect Model (1): Y ij = µ + A i + ɛ ij, Ji A i = 0 Estimation: µ + A i = y i. ˆµ = y..  i = y i. y.. Effect Modell
More informationLec 1: An Introduction to ANOVA
Ying Li Stockholm University October 31, 2011 Three end-aisle displays Which is the best? Design of the Experiment Identify the stores of the similar size and type. The displays are randomly assigned to
More informationSTAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)
STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons Ch. 4-5) Recall CRD means and effects models: Y ij = µ i + ϵ ij = µ + α i + ϵ ij i = 1,..., g ; j = 1,..., n ; ϵ ij s iid N0, σ 2 ) If we reject
More informationUnit 12: Analysis of Single Factor Experiments
Unit 12: Analysis of Single Factor Experiments Statistics 571: Statistical Methods Ramón V. León 7/16/2004 Unit 12 - Stat 571 - Ramón V. León 1 Introduction Chapter 8: How to compare two treatments. Chapter
More informationBIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES
BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method
More informationLinear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1
Linear Combinations Comparison of treatment means Bruce A Craig Department of Statistics Purdue University STAT 514 Topic 6 1 Linear Combinations of Means y ij = µ + τ i + ǫ ij = µ i + ǫ ij Often study
More informationANOVA Multiple Comparisons
ANOVA Multiple Comparisons Multiple comparisons When we carry out an ANOVA on k treatments, we test H 0 : µ 1 = =µ k versus H a : H 0 is false Assume we reject the null hypothesis, i.e. we have some evidence
More informationComparisons among means (or, the analysis of factor effects)
Comparisons among means (or, the analysis of factor effects) In carrying out our usual test that μ 1 = = μ r, we might be content to just reject this omnibus hypothesis but typically more is required:
More informationSTK4900/ Lecture 3. Program
STK4900/9900 - Lecture 3 Program 1. Multiple regression: Data structure and basic questions 2. The multiple linear regression model 3. Categorical predictors 4. Planned experiments and observational studies
More informationIntroduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs
Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique
More information22s:152 Applied Linear Regression. Take random samples from each of m populations.
22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each
More informationLaboratory Topics 4 & 5
PLS205 Lab 3 January 23, 2014 Orthogonal contrasts Class comparisons in SAS Trend analysis in SAS Multiple mean comparisons Laboratory Topics 4 & 5 Orthogonal contrasts Planned, single degree-of-freedom
More information1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College
1-Way ANOVA MATH 143 Department of Mathematics and Statistics Calvin College Spring 2010 The basic ANOVA situation Two variables: 1 Categorical, 1 Quantitative Main Question: Do the (means of) the quantitative
More information22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA
22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each
More informationhttp://www.statsoft.it/out.php?loc=http://www.statsoft.com/textbook/ Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences
More information2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018
Math 403 - P. & S. III - Dr. McLoughlin - 1 2018 2 Hand-out 2 Dr. M. P. M. M. M c Loughlin Revised 2018 3. Fundamentals 3.1. Preliminaries. Suppose we can produce a random sample of weights of 10 year-olds
More information1 One-way Analysis of Variance
1 One-way Analysis of Variance Suppose that a random sample of q individuals receives treatment T i, i = 1,,... p. Let Y ij be the response from the jth individual to be treated with the ith treatment
More informationOutline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013
Topic 19 - Inference - Fall 2013 Outline Inference for Means Differences in cell means Contrasts Multiplicity Topic 19 2 The Cell Means Model Expressed numerically Y ij = µ i + ε ij where µ i is the theoretical
More informationThe legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.
1 Chapter 1: Research Design Principles The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization. 2 Chapter 2: Completely Randomized Design
More informationExample: Poisondata. 22s:152 Applied Linear Regression. Chapter 8: ANOVA
s:5 Applied Linear Regression Chapter 8: ANOVA Two-way ANOVA Used to compare populations means when the populations are classified by two factors (or categorical variables) For example sex and occupation
More informationMultiple Comparison Methods for Means
SIAM REVIEW Vol. 44, No. 2, pp. 259 278 c 2002 Society for Industrial and Applied Mathematics Multiple Comparison Methods for Means John A. Rafter Martha L. Abell James P. Braselton Abstract. Multiple
More informationOrthogonal, Planned and Unplanned Comparisons
This is a chapter excerpt from Guilford Publications. Data Analysis for Experimental Design, by Richard Gonzalez Copyright 2008. 8 Orthogonal, Planned and Unplanned Comparisons 8.1 Introduction In this
More informationGroup comparison test for independent samples
Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences between means. Supposing that: samples come from normal populations
More informationW&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant
W&M CSCI 688: Design of Experiments Homework 2 Megan Rose Bryant September 25, 201 3.5 The tensile strength of Portland cement is being studied. Four different mixing techniques can be used economically.
More informationFactorial and Unbalanced Analysis of Variance
Factorial and Unbalanced Analysis of Variance Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 04-Jan-2017 Nathaniel E. Helwig (U of Minnesota)
More informationMultiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina
Multiple Testing Tim Hanson Department of Statistics University of South Carolina January, 2017 Modified from originals by Gary W. Oehlert Type I error A Type I error is to wrongly reject the null hypothesis
More informationTukey Complete Pairwise Post-Hoc Comparison
Tukey Complete Pairwise Post-Hoc Comparison Engineering Statistics II Section 10.2 Josh Engwer TTU 2018 Josh Engwer (TTU) Tukey Complete Pairwise Post-Hoc Comparison 2018 1 / 23 PART I PART I: Gosset s
More informationH0: Tested by k-grp ANOVA
Analyses of K-Group Designs : Omnibus F, Pairwise Comparisons & Trend Analyses ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures
More informationSTAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis
STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis Rebecca Barter April 6, 2015 Multiple Testing Multiple Testing Recall that when we were doing two sample t-tests, we were testing the equality
More informationStats fest Analysis of variance. Single factor ANOVA. Aims. Single factor ANOVA. Data
1 Stats fest 2007 Analysis of variance murray.logan@sci.monash.edu.au Single factor ANOVA 2 Aims Description Investigate differences between population means Explanation How much of the variation in response
More informationANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS
ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing
More informationAnalytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 2009
Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 009 September 9, 009 Today s Class Chapter 4 Analytic comparisons The need for analytic comparisons
More informationDiagnostics and Transformations Part 2
Diagnostics and Transformations Part 2 Bivariate Linear Regression James H. Steiger Department of Psychology and Human Development Vanderbilt University Multilevel Regression Modeling, 2009 Diagnostics
More informationMultiple Comparisons
Multiple Comparisons Error Rates, A Priori Tests, and Post-Hoc Tests Multiple Comparisons: A Rationale Multiple comparison tests function to tease apart differences between the groups within our IV when
More informationThe One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)
The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or
More informationMultiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model
Biostatistics 250 ANOVA Multiple Comparisons 1 ORIGIN 1 Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model When the omnibus F-Test for ANOVA rejects the null hypothesis that
More informationObjectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters
Objectives 10.1 Simple linear regression Statistical model for linear regression Estimating the regression parameters Confidence interval for regression parameters Significance test for the slope Confidence
More informationCorrelation and the Analysis of Variance Approach to Simple Linear Regression
Correlation and the Analysis of Variance Approach to Simple Linear Regression Biometry 755 Spring 2009 Correlation and the Analysis of Variance Approach to Simple Linear Regression p. 1/35 Correlation
More informationMultiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600
Multiple Comparison Procedures Cohen Chapter 13 For EDUC/PSY 6600 1 We have to go to the deductions and the inferences, said Lestrade, winking at me. I find it hard enough to tackle facts, Holmes, without
More informationChapter 1 Statistical Inference
Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations
More information22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)
22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are
More informationWeek 14 Comparing k(> 2) Populations
Week 14 Comparing k(> 2) Populations Week 14 Objectives Methods associated with testing for the equality of k(> 2) means or proportions are presented. Post-testing concepts and analysis are introduced.
More informationStat 502 Design and Analysis of Experiments General Linear Model
1 Stat 502 Design and Analysis of Experiments General Linear Model Fritz Scholz Department of Statistics, University of Washington December 6, 2013 2 General Linear Hypothesis We assume the data vector
More informationDESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective
DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective Second Edition Scott E. Maxwell Uniuersity of Notre Dame Harold D. Delaney Uniuersity of New Mexico J,t{,.?; LAWRENCE ERLBAUM ASSOCIATES,
More informationYour schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table
Your schedule of coming weeks One-way ANOVA, II 9.07 //00 Today: One-way ANOVA, part II Next week: Two-way ANOVA, parts I and II. One-way ANOVA HW due Thursday Week of May Teacher out of town all week
More informationLectures on Simple Linear Regression Stat 431, Summer 2012
Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population
More informationOHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd
Why We Use Analysis of Variance to Compare Group Means and How it Works The question of how to compare the population means of more than two groups is an important one to researchers. Let us suppose that
More informationOne-way between-subjects ANOVA. Comparing three or more independent means
One-way between-subjects ANOVA Comparing three or more independent means Data files SpiderBG.sav Attractiveness.sav Homework: sourcesofself-esteem.sav ANOVA: A Framework Understand the basic principles
More information3. Design Experiments and Variance Analysis
3. Design Experiments and Variance Analysis Isabel M. Rodrigues 1 / 46 3.1. Completely randomized experiment. Experimentation allows an investigator to find out what happens to the output variables when
More informationOrthogonal contrasts and multiple comparisons
BIOL 933 Lab 4 Fall 2017 Orthogonal contrasts Class comparisons in R Trend analysis in R Multiple mean comparisons Orthogonal contrasts and multiple comparisons Orthogonal contrasts Planned, single degree-of-freedom
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationLecture 10: Generalized likelihood ratio test
Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual
More informationChapter Seven: Multi-Sample Methods 1/52
Chapter Seven: Multi-Sample Methods 1/52 7.1 Introduction 2/52 Introduction The independent samples t test and the independent samples Z test for a difference between proportions are designed to analyze
More informationKeppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means
Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means 4.1 The Need for Analytical Comparisons...the between-groups sum of squares averages the differences
More informationChapter 10. Design of Experiments and Analysis of Variance
Chapter 10 Design of Experiments and Analysis of Variance Elements of a Designed Experiment Response variable Also called the dependent variable Factors (quantitative and qualitative) Also called the independent
More informationMath 423/533: The Main Theoretical Topics
Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)
More informationWhat If There Are More Than. Two Factor Levels?
What If There Are More Than Chapter 3 Two Factor Levels? Comparing more that two factor levels the analysis of variance ANOVA decomposition of total variability Statistical testing & analysis Checking
More informationSolutions to Final STAT 421, Fall 2008
Solutions to Final STAT 421, Fall 2008 Fritz Scholz 1. (8) Two treatments A and B were randomly assigned to 8 subjects (4 subjects to each treatment) with the following responses: 0, 1, 3, 6 and 5, 7,
More informationAnalysis of Variance (ANOVA)
Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationWhile you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1
While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 Chapter 12 Analysis of Variance McGraw-Hill, Bluman, 7th ed., Chapter 12 2
More information22s:152 Applied Linear Regression. 1-way ANOVA visual:
22s:152 Applied Linear Regression 1-way ANOVA visual: Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Y We now consider an analysis
More informationIntroduction to Analysis of Variance (ANOVA) Part 2
Introduction to Analysis of Variance (ANOVA) Part 2 Single factor Serpulid recruitment and biofilms Effect of biofilm type on number of recruiting serpulid worms in Port Phillip Bay Response variable:
More informationAnalysis of Variance II Bios 662
Analysis of Variance II Bios 662 Michael G. Hudgens, Ph.D. mhudgens@bios.unc.edu http://www.bios.unc.edu/ mhudgens 2008-10-24 17:21 BIOS 662 1 ANOVA II Outline Multiple Comparisons Scheffe Tukey Bonferroni
More informationConfidence Intervals, Testing and ANOVA Summary
Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0
More informationT-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum
T-test: means of Spock's judge versus all other judges 1 The TTEST Procedure Variable: pcwomen judge1 N Mean Std Dev Std Err Minimum Maximum OTHER 37 29.4919 7.4308 1.2216 16.5000 48.9000 SPOCKS 9 14.6222
More informationExample: Four levels of herbicide strength in an experiment on dry weight of treated plants.
The idea of ANOVA Reminders: A factor is a variable that can take one of several levels used to differentiate one group from another. An experiment has a one-way, or completely randomized, design if several
More informationANOVA: Comparing More Than Two Means
ANOVA: Comparing More Than Two Means Chapter 11 Cathy Poliak, Ph.D. cathy@math.uh.edu Office Fleming 11c Department of Mathematics University of Houston Lecture 25-3339 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationA posteriori multiple comparison tests
A posteriori multiple comparison tests 11/15/16 1 Recall the Lakes experiment Source of variation SS DF MS F P Lakes 58.000 2 29.400 8.243 0.006 Error 42.800 12 3.567 Total 101.600 14 The ANOVA tells us
More informationCOMPARISON OF MEANS OF SEVERAL RANDOM SAMPLES. ANOVA
Experimental Design and Statistical Methods Workshop COMPARISON OF MEANS OF SEVERAL RANDOM SAMPLES. ANOVA Jesús Piedrafita Arilla jesus.piedrafita@uab.cat Departament de Ciència Animal i dels Aliments
More informationχ test statistics of 2.5? χ we see that: χ indicate agreement between the two sets of frequencies.
I. T or F. (1 points each) 1. The χ -distribution is symmetric. F. The χ may be negative, zero, or positive F 3. The chi-square distribution is skewed to the right. T 4. The observed frequency of a cell
More informationANOVA (Analysis of Variance) output RLS 11/20/2016
ANOVA (Analysis of Variance) output RLS 11/20/2016 1. Analysis of Variance (ANOVA) The goal of ANOVA is to see if the variation in the data can explain enough to see if there are differences in the means.
More informationStatistical Inference: Estimation and Confidence Intervals Hypothesis Testing
Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire
More informationFormal Statement of Simple Linear Regression Model
Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor
More informationOne-Way ANOVA Calculations: In-Class Exercise Psychology 311 Spring, 2013
One-Way ANOVA Calculations: In-Class Exercise Psychology 311 Spring, 2013 1. You are planning an experiment that will involve 4 equally sized groups, including 3 experimental groups and a control. Each
More informationRegression models. Categorical covariate, Quantitative outcome. Examples of categorical covariates. Group characteristics. Faculty of Health Sciences
Faculty of Health Sciences Categorical covariate, Quantitative outcome Regression models Categorical covariate, Quantitative outcome Lene Theil Skovgaard April 29, 2013 PKA & LTS, Sect. 3.2, 3.2.1 ANOVA
More informationANOVA: Comparing More Than Two Means
1 ANOVA: Comparing More Than Two Means 10.1 ANOVA: The Completely Randomized Design Elements of a Designed Experiment Before we begin any calculations, we need to discuss some terminology. To make this
More informationIntroduction to the Analysis of Variance (ANOVA)
Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more
More informationMultiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota
Multiple Testing Gary W. Oehlert School of Statistics University of Minnesota January 28, 2016 Background Suppose that you had a 20-sided die. Nineteen of the sides are labeled 0 and one of the sides is
More informationPROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design
PROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design The purpose of this experiment was to determine differences in alkaloid concentration of tea leaves, based on herb variety (Factor A)
More informationOne-way ANOVA (Single-Factor CRD)
One-way ANOVA (Single-Factor CRD) STAT:5201 Week 3: Lecture 3 1 / 23 One-way ANOVA We have already described a completed randomized design (CRD) where treatments are randomly assigned to EUs. There is
More informationMultiple comparisons - subsequent inferences for two-way ANOVA
1 Multiple comparisons - subsequent inferences for two-way ANOVA the kinds of inferences to be made after the F tests of a two-way ANOVA depend on the results if none of the F tests lead to rejection of
More informationThe entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials.
One-Way ANOVA Summary The One-Way ANOVA procedure is designed to construct a statistical model describing the impact of a single categorical factor X on a dependent variable Y. Tests are run to determine
More information10. Issues on the determination of trial size
10. Issues on the determination of trial size 10.1. The general theory Review of hypothesis testing Null and alternative hypotheses. Simple and composite hypotheses. Test statistic and critical value.
More informationOne-Way Analysis of Variance (ANOVA) There are two key differences regarding the explanatory variable X.
One-Way Analysis of Variance (ANOVA) Also called single factor ANOVA. The response variable Y is continuous (same as in regression). There are two key differences regarding the explanatory variable X.
More informationIn ANOVA the response variable is numerical and the explanatory variables are categorical.
1 ANOVA ANOVA means ANalysis Of VAriance. The ANOVA is a tool for studying the influence of one or more qualitative variables on the mean of a numerical variable in a population. In ANOVA the response
More informationH0: Tested by k-grp ANOVA
Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures Alpha estimation reconsidered H0: Tested by k-grp ANOVA Regardless
More informationThe Random Effects Model Introduction
The Random Effects Model Introduction Sometimes, treatments included in experiment are randomly chosen from set of all possible treatments. Conclusions from such experiment can then be generalized to other
More informationInference for Regression
Inference for Regression Section 9.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 13b - 3339 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationStatistics for EES Factorial analysis of variance
Statistics for EES Factorial analysis of variance Dirk Metzler June 12, 2015 Contents 1 ANOVA and F -Test 1 2 Pairwise comparisons and multiple testing 6 3 Non-parametric: The Kruskal-Wallis Test 9 1 ANOVA
More informationSTAT22200 Spring 2014 Chapter 13B
STAT22200 Spring 2014 Chapter 13B Yibi Huang May 27, 2014 13.3.1 Crossover Designs 13.3.4 Replicated Latin Square Designs 13.4 Graeco-Latin Squares Chapter 13B - 1 13.3.1 Crossover Design (A Special Latin-Square
More informationLecture 18: Simple Linear Regression
Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength
More informationOutline. Example and Model ANOVA table F tests Pairwise treatment comparisons with LSD Sample and subsample size determination
Outline 1 The traditional approach 2 The Mean Squares approach for the Completely randomized design (CRD) CRD and one-way ANOVA Variance components and the F test Inference about the intercept Sample vs.
More information4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES
4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES FOR SINGLE FACTOR BETWEEN-S DESIGNS Planned or A Priori Comparisons We previously showed various ways to test all possible pairwise comparisons for
More informationy response variable x 1, x 2,, x k -- a set of explanatory variables
11. Multiple Regression and Correlation y response variable x 1, x 2,, x k -- a set of explanatory variables In this chapter, all variables are assumed to be quantitative. Chapters 12-14 show how to incorporate
More informationMAT3378 (Winter 2016)
MAT3378 (Winter 2016) Assignment 2 - SOLUTIONS Total number of points for Assignment 2: 12 The following questions will be marked: Q1, Q2, Q4 Q1. (4 points) Assume that Z 1,..., Z n are i.i.d. normal random
More information