The Distribution of F

Size: px
Start display at page:

Download "The Distribution of F"

Transcription

1 The Distribution of F It can be shown that F = SS Treat/(t 1) SS E /(N t) F t 1,N t,λ a noncentral F-distribution with t 1 and N t degrees of freedom and noncentrality parameter λ = t i=1 n i(µ i µ) 2 /σ 2 = t i=1 n i j=1 (µ i µ) 2 /σ 2 Under H 0 : µ 1 =... = µ t this becomes the (central) F t 1,N t distribution. We reject H 0 whenever F F t 1,N t (1 α) = Fcrit = qf(1 α,t 1,N t) which denotes the (1 α)-quantile of the F t 1,N t distribution. Power function: β(λ) = P(F F t 1,N t (1 α)) = 1 pf(fcrit,t 1,N t,λ) 1

2 Discussion of Noncentrality Parameter λ The power of the ANOVA F-test is a monotone increasing function of λ = t i=1 n i(µ i µ) 2 /σ 2 = N t i=1 (n i/n)(µ i µ) 2 /σ 2 = N σ 2 µ/σ 2 = N between treatment variation/within treatment variation Thus we consider the drivers in λ. λ increases as σ decreases (provided the µ i are not all the same). The more difference there is between the treatment means µ i the higher λ Increasing the sample sizes will magnify n i (µ i µ) 2 (provided µ is stable). The sample sizes we can plan for. Later: we can reduce σ by blocking units into more homogeneous groups. 2

3 Optimal Allocation of Sample Sizes? We have N experimental units available for testing the effects of t treatments and suppose that N is a multiple of t, say N = r t (r and t integer). It would seem best to use samples of equal size r for each of the t treatments i.e., we would opt for a balanced design. That way we would not emphasize one treatment over any of the others. Is there some other optimality criterion that could be used as justification? We may plan for a balanced design upfront, but then something goes wrong with a few observations and they have to be discarded from analysis. Be careful that the deletion of observations does not bias any conclusions. 3

4 A Sample Size Allocation Rationale We may be concerned with alternatives where all means but one are the same. Since we won t know upfront which mean sticks out, we would want to maximize the minimum power against all such contingencies. Max-Min Strategy! If µ 1 = µ + and µ 2 =... = µ t = µ then µ = µ + n 1 /N and (algebra) t λ 1 = n i (µ i µ) 2 /σ 2 = N 2 n ( 1 i=1 σ 2 1 n ) 1 N N for the other cases. It is easy to see now that for fixed σ max min n 1,...,n t 1 i t [ N 2 σ 2 is achieved when n 1 =... = n t. That is because n ( i 1 n ) i N N increases for n i /N 1/2. n i N and similarly ( 1 n ) ] i N λ i = N 2 σ 2 n i N ( 1 n ) i N 4

5 Using sample.sizeanova (see web page) Suppose we have t = 3 treatments and want to determine the sample size n per treatment to achieve power β(λ) =.9 for level α =.05. It is desired to do this for a λ = λ i corresponding to the alternatives on the previous slide with /σ =.5, i.e., with N = t n λ i = N 2 σ 2 n ( 1 n ) ( = n 2 N N σ ) = n 2 t σ 2 t 1 = n λ t 0. λ 0 = ( 2 /σ 2 ) (t 1)/t can be interpreted more generally as (µ i µ) 2 /σ 2. > sample.sizeanova() > sample.sizeanova(nrange=30:100) > sample.sizeanova(nrange=70:100,power0=.9) produced the next 3 slides = n = 77. 5

6 sample.sizeanova I function (delta.per.sigma=.5,t.treat=3, nrange=2:30,alpha=.05,power0=null) { # delta.per.sigma is the ratio of delta over sigma # for which one wants to detect a delta shift in one # mean while all other means stay the same. # t.treat is the number of treatments. alpha is the # desired significance level. nrange is a range of # sample sizes over which the power will be calculated # for that delta.per.sigma. power0 is on optional value # for the target power that will be highlighted on the plot. # lambda0=((t.treat-1)/t.treat)*delta.per.sigmaˆ2 power=null 6

7 sample.sizeanova II for(n in nrange){ N=n*t.treat Fcrit=qf(1-alpha,t.treat-1,N-t.treat) power=c(power,1-pf(fcrit,t.treat-1,n-t.treat,n*lambda0))} plot(nrange,power,type="l",xlab=paste("sample size n per each of t =", t.treat," treatments"), ylab="",ylim=c(0,1)) mtext(expression(beta(lambda) "=" beta(n %*% lambda[0])),2,2.7) abline(h=seq(0,1,.02),col="grey") abline(v=nrange,col="grey") lines(nrange,power,col="red") title(substitute(delta/sigma==delta.per.sigma "," lambda[0] "=" sum((mu[i]-bar(mu))ˆ2/sigmaˆ2) "=" lambda0 ", " alpha==alpha1, list(lambda0=format(signif(lambda0,4)),alpha1=alpha, delta.per.sigma=delta.per.sigma))) 7

8 sample.sizeanova III if(!is.null(power0)){ abline(h=power0,col="blue") par(las=2) mtext(power0,4,0.2,at=power0,col="blue") par(las=0)}} 8

9 Sample Size Determination σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 9

10 Sample Size Determination (increased n) σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 10

11 Sample Size Determination (magnified) σ = 0.5, λ 0 = (µ i µ) 2 σ 2 = , α = 0.05 β(λ) = β(n λ 0 ) sample size n per each of t = 3 treatments 11

12 Coagulation Example In order to understand the blood coagulation behavior in relation to various diets, lab animals were given 4 different diets and their subsequent blood draws were then measured for their respective coagulation times in seconds. The lab animals were assigned randomly to the various diets. The results were as follows: > ctime [1] [19] > diet [1] "A" "A" "A" "A" "B" "B" "B" "B" "B" "B" "C" "C" "C" [14] "C" "C" "C" "D" "D" "D" "D" "D" "D" "D" "D" 12

13 Plot for Coagulation Example n A = 4 n B = 6 n C = 6 n D = 8 coagulation time (sec) A A A A B B B B B B C C C C D D D D D D D diet 13

14 ANOVA for Coagulation Example Note that in the previous plot we used jitter(ctime) to plot ctime in the vertical direction and to plot its horizontal mean lines. This perturbs tied observations a small random amount to make tied observations more visible. For example, the mean lines for diet A and D would have been the same otherwise. > anova(lm(ctime as.factor(diet))) Analysis of Variance Table Response: ctime Df Sum Sq Mean Sq F value Pr(>F) as.factor(diet) e-05 *** Residuals Signif. codes: 0 *** ** 0.01 *

15 lm for Coagulation Example > out=lm(ctime as.factor(diet)) > names(out) [1] "coefficients" "residuals" "effects" [4] "rank" "fitted.values" "assign" [7] "qr" "df.residual" "contrasts" [10] "xlevels" "call" "terms" [13] "model" > out$coefficients (Intercept) as.factor(diet)b as.factor(diet)c e e e+00 as.factor(diet)d e-14 Note that these are the estimates ˆµ A, ˆµ B ˆµ A, ˆµ C ˆµ A, ˆµ D ˆµ A. 15

16 Residuals from lm for Coagulation Example > out$residuals e e e e e e e e e e e e e e e e e e e e e e e e+00 Numbers such as e-17 should be treated as 0 (computing quirks). 16

17 Fitted Values from lm for Coagulation Example > out$fitted.values

18 Comparing Treatment Means Ȳ i. When the hypothesis H 0 : µ 1 =... = µ t is not rejected at level α then there is little purpose in looking closer at differences between the sample means Ȳ i. for the various treatments. Any such perceived differences could easily have come about by simple random variation, even when the hypothesis is true. Why then read something into randomness? It is like reading tea leaves! However, when the hypothesis is rejected it is quite natural to ask in which way the hypothesis was contradicted. The best indicators for any analysis as to how the means µ i may be different would be the sample or treatment means ˆµ i = Ȳ i., i = 1,...,t. 18

19 Confidence Intervals for µ i A first step in understanding differences in the µ i is to look at their estimates ˆµ i =Ȳ i.. We should do this in the context of the sampling variability of ˆµ i. In the past we addressed this via confidence intervals for µ i based on ˆµ i. In any such confidence interval we can now use the pooled variance s 2 from all t samples and not just the variance s 2 i from the ith sample, i.e. we get ˆµ i ±t 1 α/2,n t s ni as our 100(1 α)% confidence interval for µ i. This follows as before (exercise) from the independence of ˆµ i and s, the fact that (ˆµ i µ i )/(σ/ n i ) N (0,1), and from s 2 /σ 2 χ 2 N t /(N t). The validity of this improvement (N t n i 1 when using s 2 instead of s 2 i ) depends strongly on the assumption that the population variances σ 2 behind all t samples are the same, or at least approximately so. 19

20 Plots of Confidence Intervals for Coagulation Data coagulation time (sec) using pooled s 2 = using individual s i 2 t i=1 s i2 (n i 1) (N t) diet A diet B diet C diet D 20

21 Simultaneous Confidence Intervals When constructing intervals of the type: ˆµ i ±t 1 α/2 s ni or ˆµ i ±t 1 α/2 s i ni for i = 1,...,t we should be aware that these intervals don t simultaneously cover their respective targets µ i with probability 1 α. They do so individually. For example ( ) s t ( ) P µ i ˆµ i ±t i s 1 α/2, i = 1,...,t = ni P µ i ˆµ i ±t i 1 α/2 i=1 ni = (1 α) t < 1 α. Thus we should choose α for individual intervals to get (1 α ) t = 1 α or α = 1 (1 α) 1/t α t = α t. Some problem remains when using common pooled estimate s. No independence! 21

22 α = 1 (1 α) 1/t α/t or α ~ t = α t α t * = 1 (1 α) (1 t) α * (1 t) t = 1 (1 α) α ~ t = α t t = 3 t = 5 t = 6 t = desired overall α 22

23 Dealing with Dependence from Using Pooled s When we use a common pooled estimate s for the standard deviation σ the previous confidence intervals are no longer independent. However, it can be shown that ( ) s P µ i ˆµ i ±t 1 α /2, i = 1,...,t ni t ( ) s P µ i ˆµ i ±t 1 α /2 i=1 ni = (1 α ) t = 1 α This comes from the positive dependence between confidence intervals through s, i.e., if one interval is more (less) likely to cover its target µ i due to s, so are the other intervals more (less) likely to cover their targets µ j. Using the same compensation as in the independence case would let us err on the conservative side, i.e., give us higher confidence than the targeted 1 α. 23

24 Boole s and Bonferroni s Inequality For any m events E 1,...,E m Boole s inequality states P(E 1... E m ) P(E 1 ) P(E m ) For any m events E 1,...,E m Bonferroni s inequality states P(E 1... E m ) 1 m (1 P(E i )) i=1 The statement are equivalent, since P(E 1... E m ) = 1 P(E c 1... Ec m). If E i denotes the i th coverage event { µ i ˆµ i ±t 1 α/2 s ni } with P(E i ) = 1 α, then the simultaneous coverage probability is bounded from below as follows ( ) \ t t P E i 1 (1 P(E i )) = 1 t α = 1 α if α = α t = α/t, i=1 i=1 i.e., we can achieve at least 1 α probability coverage by choosing the individual coverage appropriately, namely 1 α = 1 α/t. Almost same adjustment. 24

25 Contrasts Any linear combination C = t i=1 c iµ i with t i=1 c i = 0 is called a contrast. Note that t i=1 c iµ = 0, i.e., contrasts are zero over the hypothesis. Suppose we have 4 treatments with respective means µ 1,...,µ 4. We may be interested in contrasts of the following form C 12 = µ 1 µ 2 with c = (c 1,...,c 4 ) = (1, 1,0,0). Similarly for the other differences C i j = µ i µ ( 4) j. There are 2 = 6 such contrasts. Sometimes one of the treatments, say the first, is singled out as the control. We may then be interested in just the 3 contrasts C 12,C 13 and C 14 or we may be interested in C = µ 1 (µ 2 + µ 3 + µ 4 )/3 with c = (1, 1/3, 1/3, 1/3). Sometimes the first 2 treatment share something in common and so do the last 2. One might then try: C = (µ 1 +µ 2 )/2 (µ 3 +µ 4 )/2 with c = (1/2,1/2, 1/2, 1/2) 25

26 Estimates and Confidence Intervals for Contrasts A natural estimate of C = t i=1 c iµ i is Ĉ = t i=1 c iˆµ i = t i=1 c iȳi.. ( ) t t t We have E(Ĉ) = E c i Ȳ i. = c i E (Ȳ i.) = c i µ i = C i=1 i=1 i=1 t and var(ĉ) = var( c i Ȳ i. i=1 Under the normality assumption for the Y i j we have ) = Ĉ C t N t where s 2 = s t i=1 c2 i /n i t c 2 i var(ȳ i.) = i=1 t c 2 i σ2 /n i. i=1 t n i s 2 i /(N t) = MS E. i=1 = Ĉ±t N t,1 α/2 s t c 2 i /n i is a 100(1 α)% confidence interval for C. i=1 26

27 Testing H 0 : C = 0 Based on the duality of testing and confidence intervals we can test the hypothesis H 0 : C = 0 by rejecting it whenever the previous confidence interval does not contain C = 0. Similarly, reject H 0 : C = C 0 by rejecting it whenever the previous confidence interval does not contain C = C 0 Another notation for this interval is Ĉ ±t N t,1 α/2 SE(Ĉ) where SE(Ĉ) = s t c 2 i /n i. i=1 SE(Ĉ) is the standard error of Ĉ, the estimate of the standard deviation of Ĉ. 27

28 Paired Comparisons: Fisher s Protected LSD Method After rejecting H 0 : µ 1 =... = µ t one is often interested in looking at all ( t 2 ) pairwise contrasts C i j = µ i µ j. The following procedure is referred to as Fisher s Protected Least Significant Difference (LSD) Method. It consists of possibly two stages: 1) Perform α level F-test for testing H 0. If H 0 is not rejected, stop. 2) If H ( t ) 0 is rejected, form all 2 (1 α)-level confidence intervals for Ci j = µ i µ j : 1 Î i j = ˆµ i ˆµ j ±t N t,1 α/2 s + 1 n i n j and declare all µ i µ j 0 for which Î i j does not contain zero. 28

29 Comments on Fisher s Protected LSD Method If H 0 is true, the chance of making any statements contradicting H 0 is at most α. This is the protected aspect of this procedure. However, when H 0 is not true there are many possible contingencies, some of which can give us a higher than desired chance of pronouncing a significant difference, when in fact there is none. E.g., if all but one mean (say µ 1 ) are equal and µ 1 is far away from µ 2 =... = µ t our chance of rejecting H 0 is almost 1. However, among the intervals for µ i µ j, 2 i < j we may find a significantly higher than α proportion of cases with wrongly declared differences. This is due to the multiple comparison issue. 29

30 Pairwise Comparisons: Tukey-Kramer Method The Tukey-Kramer method is based on the distribution of { } Zi Z j i.i.d. Q t, f = max where Z 1 i< j t s 1,...,Z t N (0,1) and f s 2 χ 2 f Its cdf and quantile function are given in R as ptukey(q,nmeans,df) and qtukey(p,nmeans,df), nmeans = t is the number of means, df = f = N t denotes the degrees of freedom in s. Applying this to Z i = (ˆµ i µ i )/(σ/ n) and assuming n 1 =... = n t = n we get max i< j { } n ˆµi ˆµ j (µ i µ j ) s = max i< j ˆµ i µ i σ/ n ˆµ j µ j σ/ n s/σ = Q t, f P ( µ i µ j ˆµ i ˆµ j ± q t, f,1 α s/ n i < j ) = 1 α simultaneous (1 α)-coverage confidence intervals. Here P(Q t, f q t, f,1 α ) = 1 α or q t, f,1 α = qtukey(1 α,t,f). 30

31 Tukey-Kramer Method: Unequal Sample Sizes The simultaneous intervals for all pairwise mean differences was due to Tukey, but it was hampered by the requirement of equal sample sizes. This was addressed by Kramer in the following way. In the above confidence intervals replace n in 1/ n = 1/n by n i j, where n i j is the harmonic mean of n i and n j, i.e., 1/n i j = (1/n i +1/n j )/2. Different adjustment for each pair (i, j)! It was possible to show ( ) P µ i µ j ˆµ i ˆµ j ± Q t, f,1 α s/ n i j i < j 1 α simultaneous confidence intervals with coverage 1 α. 31

32 Tukey-Kramer Method for Coagulation Data coag.tukey = function (alpha=.05) { diets=unique(diet) mu.vec=null nvec=null mean.vec=null for(i in 1:length(diets)){ mu.vec=c(mu.vec,mean(ctime[diet==diets[i]])) nvec=c(nvec,length(ctime[diet==diets[i]])) mean.vec=c(mean.vec,rep(mu.vec[i],nvec[i])) } tr=length(nvec) N=sum(nvec) MSE=sum((ctime-mean.vec)ˆ2/(N-tr)) 32

33 Tukey-Kramer Method for Coagulation Data s=sqrt(mse) intervals=null for(i in 1:3){ for(j in (i+1):4){ nijstar=1/(.5*(1/nvec[i]+1/nvec[j])) qtk=qtukey(1-alpha,tr,n-tr) Diff=mu.vec[i]-mu.vec[j] lower=diff - qtk*s/sqrt(nijstar) upper=diff + qtk*s/sqrt(nijstar) intervals=rbind(intervals,c(lower,upper)) } } intervals } 33

34 Tukey-Kramer Results for Coagulation Data > coag.tukey() [,1] [,2] [1,] [2,] [3,] [4,] [5,] [6,] Declare significant differences in µ 1 µ 2, µ 1 µ 3, µ 2 µ 4, and µ 3 µ 4. 34

35 Scheffé s Confidence Intervals for All Contrasts Scheffé took the F-test for testing H 0 : µ 1 =... = µ t and converted it into a simultaneous coverage statement about confidence intervals for all contrasts c = (c 1,...,c t ): P t c i µ i i=1 ( t t c iˆµ i ± (t 1) F t 1,N t,1 α s i=1 i=1 1/2 c 2 i i) /n c = 1 α This is a coverage statement about an infinite number of contrasts, but can be applied conservatively to all pairwise contrasts. The resulting intervals tend to be quite conservative. But it compares well with Bonferroni type intervals if applied to many contrasts. 35

36 Pairwise Comparison Intervals for Coagulation Data (simultaneous) 95%-Intervals mean Fisher s Bonferroni Scheffé s all difference Tukey-Kramer protected LSD inequality contrasts method µ 1 µ µ 1 µ µ 1 µ µ 2 µ µ 2 µ µ 3 µ Declare significant differences in µ 1 µ 2, µ 1 µ 3, µ 2 µ 4, and µ 3 µ 4, using any of the four methods. 36

37 Simultaneous Paired Comparisons (95%) Pairwise Comparisons of Means (Coagulation Data): 1 α = Tukey Kramer pairwise comparisons Fisher's protected LSD Bonferroni intervals Scheffe's intervals for all contrasts µ 1 µ 2 µ 1 µ 3 µ 1 µ 4 µ 2 µ 3 µ 2 µ 4 µ 3 µ 4 37

38 Simultaneous Paired Comparisons (99%) Pairwise Comparisons of Means (Coagulation Data): 1 α = Tukey Kramer pairwise comparisons Fisher's protected LSD Bonferroni intervals Scheffe's intervals for all contrasts µ 1 µ 2 µ 1 µ 3 µ 1 µ 4 µ 2 µ 3 µ 2 µ 4 µ 3 µ 4 38

39 Orthogonal Contrast All ( t 2 ) pairwise comparisons for µi µ j could by very many and simultaneous intervals would become quite conservative. Since all these contrasts span a (t 1)-dimensional space one should be able to capture all differences with just t 1 orthogonal contrasts. C 1 = t c 1i µ i C 2 = i=1 t c 2i µ i i=1 C 1 C 2 = cov(ĉ 1,Ĉ 2 ) = t i=1 t c 1i c 2i /n i = 0 i=1 t c 1i c 2 j cov(ˆµ i, ˆµ j ) = j=1 ( ) wrong def. in Montgomery p.91 t c 1i c 2i σ 2 /n i = 0, i=1 i.e., Ĉ 1 and Ĉ 2 are independent and simultaneous statements for C 1,C 2,... are easier to handle, just as before when making simultaneous intervals for µ 1,...,µ t based on independent ˆµ 1,..., ˆµ t. The trick is to have meaningful or interpretable orthogonal contrast. 39

40 Simultaneous Intervals for Orthogonal Contrast If Ĉ 1,...,Ĉ k are any k orthogonal contrasts, they are independent. Just as we constructed simultaneous intervals for means based on independent mean estimates and the pooled standard deviation s we can again construct contrast confidence intervals with simultaneous coverage probability 1 α by taking 1 α with α = 1 (1 α) 1/k α k as the coverage probability for the individual intervals. 40

41 An Orthogonal Contrast Example Suppose we have t = 3 treatments of which the third is a control, i.e., we are familiar with its performance. Assume further that we have a balanced design, i.e., n 1 = n 2 = n 3. We could try the following t 1 = 2 orthogonal contrasts: c 1 = (.5,.5, 1) and c 2 = (1, 1,0). Note that C 1 = (µ 1 + µ 2 )/2 µ 3 and C 2 = µ 1 µ 2, of which the first assesses how much the average mean of the two new treatments differs from the control mean and the second assesses the difference between the two new treatments. These are seemingly orthogonal issues. 41

42 Unbalanced Case of Previous Example We have an unbalanced design, i.e., n 1, n 2, n 3 may be different. Then the following t 1 = 2 vectors: c 1 = (n 1/(n 1 + n 2 ),n 2 /(n 1 + n 2 ), 1) and c 2 = (1, 1,0) are indeed contrast vectors: n 1 /(n 1 + n 2 ) + n 2 /(n 1 + n 2 ) 1 = 0 and = 0 and they are orthogonal: n 1 /[(n 1 + n 2 )n 1 ] n 2 /[(n 1 + n 2 )n 2 ] 1 0/n 3 = 0. = C 1 = (n 1 µ 1 + n 2 µ 2 )/(n 1 + n 2 ) µ 3 = µ 12 µ 3 and C 2 = µ 1 µ 2, of which the first assesses how much the weighted average mean of the two new treatments differs from the control mean and the second assesses the difference between the two new treatments. These are seemingly orthogonal issues. 42

43 Service Center Data # of persons # of calls on call processed per hour number of calls processed per hour number of people on call 43

44 Service Center Data Here we have a new type of treatment (number of persons on call), where the different treatment levels are scalar and not just qualitative. In such situations the following orthogonal contrasts are of practical interest: c i1 c i2 c i3 c i4 c i5 C 1 = 5 j=1 c 1 j µ j C 2 = 5 j=1 c 2 j µ j C 3 = 5 j=1 c 3 j µ j C 4 = 5 j=1 c 4 j µ j For what kind of mean patterns in µ 1,...,µ 5 would C i and consequently Ĉ i be large? 44

45 Orthogonal Contrast Plots 5 C i = c i, j j using µ j = j j= C 1 C 2 C 3 C j 45

46 Interpretation of Orthogonal Contrast Plots The previous plot suggests that a pattern in the means µ j in relation to j = 1,...,5 that correlates most strongly with the corresponding pattern in the plot should yield a high value for the corresponding absolute contrast C i. Thus a large value C 1 indicates a strong linear component in the mean pattern. A large value C 2 indicates a strong quadratic component in the mean pattern. A large value C 3 indicates a strong cubic component in the mean pattern. A large value C 4 indicates a strong quartic component in the mean pattern. Typically, one hopes to rule out some (if not all) of the latter possibilities. 46

47 Simultaneous Contrast Intervals (for Service Center Data) 95% 99% C 1 [ 6.27, 11.58] [ 5.53, 12.32] C 2 [-7.02, -0.73] [ -7.89, 0.14] C 3 [-1.26, 4.06] [ -1.99, 4.79] C 4 [-9.58, 4.48] [-11.53, 6.43] From these intervals one sees that C 1 and C 2 are significantly different from zero. with 95% confidence, but C 2 not quite with 99% confidence. Hence there appears to be a sufficiently strong linear and mildly quadratic component. The original data plot suggested this and its strength is now assessed statistically. 47

48 Orthogonal Polynomial Contrast Vectors The previous orthogonal contrasts for linear, quadratic, cubic, quartic behavior were tailored to five treatments. How do we get similar contrast vectors when we have t treatments? R has a function contr.poly(t) that gives you orthogonal vectors representing the various polynomial components: linear, quadratic,... > round(contr.poly(7),3).l.q.c ˆ4 ˆ5 ˆ6 [1,] [2,] [3,] [4,] [5,] [6,] [7,]

49 Orthogonal Polynomial Contrasts from contr.poly(7) orthogonal polynomial i 49

Stat 502 Design and Analysis of Experiments One-Factor ANOVA

Stat 502 Design and Analysis of Experiments One-Factor ANOVA Stat 502 Design and Analysis of Experiments One-Factor ANOVA Fritz Scholz Department of Statistics, University of Washington January 7, 2014 2 One-Factor ANOVA ANOVA is an acronym for Analysis of Variance

More information

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large?

These are all actually contrasts (the coef sum to zero). What are these contrasts representing? What would make them large? Lecture 12 Comparing treatment effects Orthogonal Contrasts What use are contrasts? Recall the Cotton data In this case, the treatment levels have an ordering to them this is not always the case) Consider

More information

STAT22200 Spring 2014 Chapter 5

STAT22200 Spring 2014 Chapter 5 STAT22200 Spring 2014 Chapter 5 Yibi Huang April 29, 2014 Chapter 5 Multiple Comparisons Chapter 5-1 Chapter 5 Multiple Comparisons Note the t-tests and C.I. s are constructed assuming we only do one test,

More information

Lecture 5: Comparing Treatment Means Montgomery: Section 3-5

Lecture 5: Comparing Treatment Means Montgomery: Section 3-5 Lecture 5: Comparing Treatment Means Montgomery: Section 3-5 Page 1 Linear Combination of Means ANOVA: y ij = µ + τ i + ɛ ij = µ i + ɛ ij Linear combination: L = c 1 µ 1 + c 1 µ 2 +...+ c a µ a = a i=1

More information

More about Single Factor Experiments

More about Single Factor Experiments More about Single Factor Experiments 1 2 3 0 / 23 1 2 3 1 / 23 Parameter estimation Effect Model (1): Y ij = µ + A i + ɛ ij, Ji A i = 0 Estimation: µ + A i = y i. ˆµ = y..  i = y i. y.. Effect Modell

More information

Lec 1: An Introduction to ANOVA

Lec 1: An Introduction to ANOVA Ying Li Stockholm University October 31, 2011 Three end-aisle displays Which is the best? Design of the Experiment Identify the stores of the similar size and type. The displays are randomly assigned to

More information

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5) STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons Ch. 4-5) Recall CRD means and effects models: Y ij = µ i + ϵ ij = µ + α i + ϵ ij i = 1,..., g ; j = 1,..., n ; ϵ ij s iid N0, σ 2 ) If we reject

More information

Unit 12: Analysis of Single Factor Experiments

Unit 12: Analysis of Single Factor Experiments Unit 12: Analysis of Single Factor Experiments Statistics 571: Statistical Methods Ramón V. León 7/16/2004 Unit 12 - Stat 571 - Ramón V. León 1 Introduction Chapter 8: How to compare two treatments. Chapter

More information

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method

More information

Linear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1

Linear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1 Linear Combinations Comparison of treatment means Bruce A Craig Department of Statistics Purdue University STAT 514 Topic 6 1 Linear Combinations of Means y ij = µ + τ i + ǫ ij = µ i + ǫ ij Often study

More information

ANOVA Multiple Comparisons

ANOVA Multiple Comparisons ANOVA Multiple Comparisons Multiple comparisons When we carry out an ANOVA on k treatments, we test H 0 : µ 1 = =µ k versus H a : H 0 is false Assume we reject the null hypothesis, i.e. we have some evidence

More information

Comparisons among means (or, the analysis of factor effects)

Comparisons among means (or, the analysis of factor effects) Comparisons among means (or, the analysis of factor effects) In carrying out our usual test that μ 1 = = μ r, we might be content to just reject this omnibus hypothesis but typically more is required:

More information

STK4900/ Lecture 3. Program

STK4900/ Lecture 3. Program STK4900/9900 - Lecture 3 Program 1. Multiple regression: Data structure and basic questions 2. The multiple linear regression model 3. Categorical predictors 4. Planned experiments and observational studies

More information

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique

More information

22s:152 Applied Linear Regression. Take random samples from each of m populations.

22s:152 Applied Linear Regression. Take random samples from each of m populations. 22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each

More information

Laboratory Topics 4 & 5

Laboratory Topics 4 & 5 PLS205 Lab 3 January 23, 2014 Orthogonal contrasts Class comparisons in SAS Trend analysis in SAS Multiple mean comparisons Laboratory Topics 4 & 5 Orthogonal contrasts Planned, single degree-of-freedom

More information

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College 1-Way ANOVA MATH 143 Department of Mathematics and Statistics Calvin College Spring 2010 The basic ANOVA situation Two variables: 1 Categorical, 1 Quantitative Main Question: Do the (means of) the quantitative

More information

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA 22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each

More information

http://www.statsoft.it/out.php?loc=http://www.statsoft.com/textbook/ Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences

More information

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018

2 Hand-out 2. Dr. M. P. M. M. M c Loughlin Revised 2018 Math 403 - P. & S. III - Dr. McLoughlin - 1 2018 2 Hand-out 2 Dr. M. P. M. M. M c Loughlin Revised 2018 3. Fundamentals 3.1. Preliminaries. Suppose we can produce a random sample of weights of 10 year-olds

More information

1 One-way Analysis of Variance

1 One-way Analysis of Variance 1 One-way Analysis of Variance Suppose that a random sample of q individuals receives treatment T i, i = 1,,... p. Let Y ij be the response from the jth individual to be treated with the ith treatment

More information

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013

Outline. Topic 19 - Inference. The Cell Means Model. Estimates. Inference for Means Differences in cell means Contrasts. STAT Fall 2013 Topic 19 - Inference - Fall 2013 Outline Inference for Means Differences in cell means Contrasts Multiplicity Topic 19 2 The Cell Means Model Expressed numerically Y ij = µ i + ε ij where µ i is the theoretical

More information

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization.

The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization. 1 Chapter 1: Research Design Principles The legacy of Sir Ronald A. Fisher. Fisher s three fundamental principles: local control, replication, and randomization. 2 Chapter 2: Completely Randomized Design

More information

Example: Poisondata. 22s:152 Applied Linear Regression. Chapter 8: ANOVA

Example: Poisondata. 22s:152 Applied Linear Regression. Chapter 8: ANOVA s:5 Applied Linear Regression Chapter 8: ANOVA Two-way ANOVA Used to compare populations means when the populations are classified by two factors (or categorical variables) For example sex and occupation

More information

Multiple Comparison Methods for Means

Multiple Comparison Methods for Means SIAM REVIEW Vol. 44, No. 2, pp. 259 278 c 2002 Society for Industrial and Applied Mathematics Multiple Comparison Methods for Means John A. Rafter Martha L. Abell James P. Braselton Abstract. Multiple

More information

Orthogonal, Planned and Unplanned Comparisons

Orthogonal, Planned and Unplanned Comparisons This is a chapter excerpt from Guilford Publications. Data Analysis for Experimental Design, by Richard Gonzalez Copyright 2008. 8 Orthogonal, Planned and Unplanned Comparisons 8.1 Introduction In this

More information

Group comparison test for independent samples

Group comparison test for independent samples Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences between means. Supposing that: samples come from normal populations

More information

W&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant

W&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant W&M CSCI 688: Design of Experiments Homework 2 Megan Rose Bryant September 25, 201 3.5 The tensile strength of Portland cement is being studied. Four different mixing techniques can be used economically.

More information

Factorial and Unbalanced Analysis of Variance

Factorial and Unbalanced Analysis of Variance Factorial and Unbalanced Analysis of Variance Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 04-Jan-2017 Nathaniel E. Helwig (U of Minnesota)

More information

Multiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina

Multiple Testing. Tim Hanson. January, Modified from originals by Gary W. Oehlert. Department of Statistics University of South Carolina Multiple Testing Tim Hanson Department of Statistics University of South Carolina January, 2017 Modified from originals by Gary W. Oehlert Type I error A Type I error is to wrongly reject the null hypothesis

More information

Tukey Complete Pairwise Post-Hoc Comparison

Tukey Complete Pairwise Post-Hoc Comparison Tukey Complete Pairwise Post-Hoc Comparison Engineering Statistics II Section 10.2 Josh Engwer TTU 2018 Josh Engwer (TTU) Tukey Complete Pairwise Post-Hoc Comparison 2018 1 / 23 PART I PART I: Gosset s

More information

H0: Tested by k-grp ANOVA

H0: Tested by k-grp ANOVA Analyses of K-Group Designs : Omnibus F, Pairwise Comparisons & Trend Analyses ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures

More information

STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis

STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis Rebecca Barter April 6, 2015 Multiple Testing Multiple Testing Recall that when we were doing two sample t-tests, we were testing the equality

More information

Stats fest Analysis of variance. Single factor ANOVA. Aims. Single factor ANOVA. Data

Stats fest Analysis of variance. Single factor ANOVA. Aims. Single factor ANOVA. Data 1 Stats fest 2007 Analysis of variance murray.logan@sci.monash.edu.au Single factor ANOVA 2 Aims Description Investigate differences between population means Explanation How much of the variation in response

More information

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing

More information

Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 2009

Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 2009 Analytical Comparisons Among Treatment Means (Chapter 4) Analysis of Trend (Chapter 5) ERSH 8310 Fall 009 September 9, 009 Today s Class Chapter 4 Analytic comparisons The need for analytic comparisons

More information

Diagnostics and Transformations Part 2

Diagnostics and Transformations Part 2 Diagnostics and Transformations Part 2 Bivariate Linear Regression James H. Steiger Department of Psychology and Human Development Vanderbilt University Multilevel Regression Modeling, 2009 Diagnostics

More information

Multiple Comparisons

Multiple Comparisons Multiple Comparisons Error Rates, A Priori Tests, and Post-Hoc Tests Multiple Comparisons: A Rationale Multiple comparison tests function to tease apart differences between the groups within our IV when

More information

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs) The One-Way Independent-Samples ANOVA (For Between-Subjects Designs) Computations for the ANOVA In computing the terms required for the F-statistic, we won t explicitly compute any sample variances or

More information

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model

Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model Biostatistics 250 ANOVA Multiple Comparisons 1 ORIGIN 1 Multiple Pairwise Comparison Procedures in One-Way ANOVA with Fixed Effects Model When the omnibus F-Test for ANOVA rejects the null hypothesis that

More information

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters Objectives 10.1 Simple linear regression Statistical model for linear regression Estimating the regression parameters Confidence interval for regression parameters Significance test for the slope Confidence

More information

Correlation and the Analysis of Variance Approach to Simple Linear Regression

Correlation and the Analysis of Variance Approach to Simple Linear Regression Correlation and the Analysis of Variance Approach to Simple Linear Regression Biometry 755 Spring 2009 Correlation and the Analysis of Variance Approach to Simple Linear Regression p. 1/35 Correlation

More information

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600

Multiple Comparison Procedures Cohen Chapter 13. For EDUC/PSY 6600 Multiple Comparison Procedures Cohen Chapter 13 For EDUC/PSY 6600 1 We have to go to the deductions and the inferences, said Lestrade, winking at me. I find it hard enough to tackle facts, Holmes, without

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are

More information

Week 14 Comparing k(> 2) Populations

Week 14 Comparing k(> 2) Populations Week 14 Comparing k(> 2) Populations Week 14 Objectives Methods associated with testing for the equality of k(> 2) means or proportions are presented. Post-testing concepts and analysis are introduced.

More information

Stat 502 Design and Analysis of Experiments General Linear Model

Stat 502 Design and Analysis of Experiments General Linear Model 1 Stat 502 Design and Analysis of Experiments General Linear Model Fritz Scholz Department of Statistics, University of Washington December 6, 2013 2 General Linear Hypothesis We assume the data vector

More information

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective Second Edition Scott E. Maxwell Uniuersity of Notre Dame Harold D. Delaney Uniuersity of New Mexico J,t{,.?; LAWRENCE ERLBAUM ASSOCIATES,

More information

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table

Your schedule of coming weeks. One-way ANOVA, II. Review from last time. Review from last time /22/2004. Create ANOVA table Your schedule of coming weeks One-way ANOVA, II 9.07 //00 Today: One-way ANOVA, part II Next week: Two-way ANOVA, parts I and II. One-way ANOVA HW due Thursday Week of May Teacher out of town all week

More information

Lectures on Simple Linear Regression Stat 431, Summer 2012

Lectures on Simple Linear Regression Stat 431, Summer 2012 Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population

More information

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd Why We Use Analysis of Variance to Compare Group Means and How it Works The question of how to compare the population means of more than two groups is an important one to researchers. Let us suppose that

More information

One-way between-subjects ANOVA. Comparing three or more independent means

One-way between-subjects ANOVA. Comparing three or more independent means One-way between-subjects ANOVA Comparing three or more independent means Data files SpiderBG.sav Attractiveness.sav Homework: sourcesofself-esteem.sav ANOVA: A Framework Understand the basic principles

More information

3. Design Experiments and Variance Analysis

3. Design Experiments and Variance Analysis 3. Design Experiments and Variance Analysis Isabel M. Rodrigues 1 / 46 3.1. Completely randomized experiment. Experimentation allows an investigator to find out what happens to the output variables when

More information

Orthogonal contrasts and multiple comparisons

Orthogonal contrasts and multiple comparisons BIOL 933 Lab 4 Fall 2017 Orthogonal contrasts Class comparisons in R Trend analysis in R Multiple mean comparisons Orthogonal contrasts and multiple comparisons Orthogonal contrasts Planned, single degree-of-freedom

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Lecture 10: Generalized likelihood ratio test

Lecture 10: Generalized likelihood ratio test Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual

More information

Chapter Seven: Multi-Sample Methods 1/52

Chapter Seven: Multi-Sample Methods 1/52 Chapter Seven: Multi-Sample Methods 1/52 7.1 Introduction 2/52 Introduction The independent samples t test and the independent samples Z test for a difference between proportions are designed to analyze

More information

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means

Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means Keppel, G. & Wickens, T. D. Design and Analysis Chapter 4: Analytical Comparisons Among Treatment Means 4.1 The Need for Analytical Comparisons...the between-groups sum of squares averages the differences

More information

Chapter 10. Design of Experiments and Analysis of Variance

Chapter 10. Design of Experiments and Analysis of Variance Chapter 10 Design of Experiments and Analysis of Variance Elements of a Designed Experiment Response variable Also called the dependent variable Factors (quantitative and qualitative) Also called the independent

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

What If There Are More Than. Two Factor Levels?

What If There Are More Than. Two Factor Levels? What If There Are More Than Chapter 3 Two Factor Levels? Comparing more that two factor levels the analysis of variance ANOVA decomposition of total variability Statistical testing & analysis Checking

More information

Solutions to Final STAT 421, Fall 2008

Solutions to Final STAT 421, Fall 2008 Solutions to Final STAT 421, Fall 2008 Fritz Scholz 1. (8) Two treatments A and B were randomly assigned to 8 subjects (4 subjects to each treatment) with the following responses: 0, 1, 3, 6 and 5, 7,

More information

Analysis of Variance (ANOVA)

Analysis of Variance (ANOVA) Analysis of Variance (ANOVA) Two types of ANOVA tests: Independent measures and Repeated measures Comparing 2 means: X 1 = 20 t - test X 2 = 30 How can we Compare 3 means?: X 1 = 20 X 2 = 30 X 3 = 35 ANOVA

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1

While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 While you wait: Enter the following in your calculator. Find the mean and sample variation of each group. Bluman, Chapter 12 1 Chapter 12 Analysis of Variance McGraw-Hill, Bluman, 7th ed., Chapter 12 2

More information

22s:152 Applied Linear Regression. 1-way ANOVA visual:

22s:152 Applied Linear Regression. 1-way ANOVA visual: 22s:152 Applied Linear Regression 1-way ANOVA visual: Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Y We now consider an analysis

More information

Introduction to Analysis of Variance (ANOVA) Part 2

Introduction to Analysis of Variance (ANOVA) Part 2 Introduction to Analysis of Variance (ANOVA) Part 2 Single factor Serpulid recruitment and biofilms Effect of biofilm type on number of recruiting serpulid worms in Port Phillip Bay Response variable:

More information

Analysis of Variance II Bios 662

Analysis of Variance II Bios 662 Analysis of Variance II Bios 662 Michael G. Hudgens, Ph.D. mhudgens@bios.unc.edu http://www.bios.unc.edu/ mhudgens 2008-10-24 17:21 BIOS 662 1 ANOVA II Outline Multiple Comparisons Scheffe Tukey Bonferroni

More information

Confidence Intervals, Testing and ANOVA Summary

Confidence Intervals, Testing and ANOVA Summary Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0

More information

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum T-test: means of Spock's judge versus all other judges 1 The TTEST Procedure Variable: pcwomen judge1 N Mean Std Dev Std Err Minimum Maximum OTHER 37 29.4919 7.4308 1.2216 16.5000 48.9000 SPOCKS 9 14.6222

More information

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants.

Example: Four levels of herbicide strength in an experiment on dry weight of treated plants. The idea of ANOVA Reminders: A factor is a variable that can take one of several levels used to differentiate one group from another. An experiment has a one-way, or completely randomized, design if several

More information

ANOVA: Comparing More Than Two Means

ANOVA: Comparing More Than Two Means ANOVA: Comparing More Than Two Means Chapter 11 Cathy Poliak, Ph.D. cathy@math.uh.edu Office Fleming 11c Department of Mathematics University of Houston Lecture 25-3339 Cathy Poliak, Ph.D. cathy@math.uh.edu

More information

A posteriori multiple comparison tests

A posteriori multiple comparison tests A posteriori multiple comparison tests 11/15/16 1 Recall the Lakes experiment Source of variation SS DF MS F P Lakes 58.000 2 29.400 8.243 0.006 Error 42.800 12 3.567 Total 101.600 14 The ANOVA tells us

More information

COMPARISON OF MEANS OF SEVERAL RANDOM SAMPLES. ANOVA

COMPARISON OF MEANS OF SEVERAL RANDOM SAMPLES. ANOVA Experimental Design and Statistical Methods Workshop COMPARISON OF MEANS OF SEVERAL RANDOM SAMPLES. ANOVA Jesús Piedrafita Arilla jesus.piedrafita@uab.cat Departament de Ciència Animal i dels Aliments

More information

χ test statistics of 2.5? χ we see that: χ indicate agreement between the two sets of frequencies.

χ test statistics of 2.5? χ we see that: χ indicate agreement between the two sets of frequencies. I. T or F. (1 points each) 1. The χ -distribution is symmetric. F. The χ may be negative, zero, or positive F 3. The chi-square distribution is skewed to the right. T 4. The observed frequency of a cell

More information

ANOVA (Analysis of Variance) output RLS 11/20/2016

ANOVA (Analysis of Variance) output RLS 11/20/2016 ANOVA (Analysis of Variance) output RLS 11/20/2016 1. Analysis of Variance (ANOVA) The goal of ANOVA is to see if the variation in the data can explain enough to see if there are differences in the means.

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Formal Statement of Simple Linear Regression Model

Formal Statement of Simple Linear Regression Model Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor

More information

One-Way ANOVA Calculations: In-Class Exercise Psychology 311 Spring, 2013

One-Way ANOVA Calculations: In-Class Exercise Psychology 311 Spring, 2013 One-Way ANOVA Calculations: In-Class Exercise Psychology 311 Spring, 2013 1. You are planning an experiment that will involve 4 equally sized groups, including 3 experimental groups and a control. Each

More information

Regression models. Categorical covariate, Quantitative outcome. Examples of categorical covariates. Group characteristics. Faculty of Health Sciences

Regression models. Categorical covariate, Quantitative outcome. Examples of categorical covariates. Group characteristics. Faculty of Health Sciences Faculty of Health Sciences Categorical covariate, Quantitative outcome Regression models Categorical covariate, Quantitative outcome Lene Theil Skovgaard April 29, 2013 PKA & LTS, Sect. 3.2, 3.2.1 ANOVA

More information

ANOVA: Comparing More Than Two Means

ANOVA: Comparing More Than Two Means 1 ANOVA: Comparing More Than Two Means 10.1 ANOVA: The Completely Randomized Design Elements of a Designed Experiment Before we begin any calculations, we need to discuss some terminology. To make this

More information

Introduction to the Analysis of Variance (ANOVA)

Introduction to the Analysis of Variance (ANOVA) Introduction to the Analysis of Variance (ANOVA) The Analysis of Variance (ANOVA) The analysis of variance (ANOVA) is a statistical technique for testing for differences between the means of multiple (more

More information

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota

Multiple Testing. Gary W. Oehlert. January 28, School of Statistics University of Minnesota Multiple Testing Gary W. Oehlert School of Statistics University of Minnesota January 28, 2016 Background Suppose that you had a 20-sided die. Nineteen of the sides are labeled 0 and one of the sides is

More information

PROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design

PROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design PROBLEM TWO (ALKALOID CONCENTRATIONS IN TEA) 1. Statistical Design The purpose of this experiment was to determine differences in alkaloid concentration of tea leaves, based on herb variety (Factor A)

More information

One-way ANOVA (Single-Factor CRD)

One-way ANOVA (Single-Factor CRD) One-way ANOVA (Single-Factor CRD) STAT:5201 Week 3: Lecture 3 1 / 23 One-way ANOVA We have already described a completed randomized design (CRD) where treatments are randomly assigned to EUs. There is

More information

Multiple comparisons - subsequent inferences for two-way ANOVA

Multiple comparisons - subsequent inferences for two-way ANOVA 1 Multiple comparisons - subsequent inferences for two-way ANOVA the kinds of inferences to be made after the F tests of a two-way ANOVA depend on the results if none of the F tests lead to rejection of

More information

The entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials.

The entire data set consists of n = 32 widgets, 8 of which were made from each of q = 4 different materials. One-Way ANOVA Summary The One-Way ANOVA procedure is designed to construct a statistical model describing the impact of a single categorical factor X on a dependent variable Y. Tests are run to determine

More information

10. Issues on the determination of trial size

10. Issues on the determination of trial size 10. Issues on the determination of trial size 10.1. The general theory Review of hypothesis testing Null and alternative hypotheses. Simple and composite hypotheses. Test statistic and critical value.

More information

One-Way Analysis of Variance (ANOVA) There are two key differences regarding the explanatory variable X.

One-Way Analysis of Variance (ANOVA) There are two key differences regarding the explanatory variable X. One-Way Analysis of Variance (ANOVA) Also called single factor ANOVA. The response variable Y is continuous (same as in regression). There are two key differences regarding the explanatory variable X.

More information

In ANOVA the response variable is numerical and the explanatory variables are categorical.

In ANOVA the response variable is numerical and the explanatory variables are categorical. 1 ANOVA ANOVA means ANalysis Of VAriance. The ANOVA is a tool for studying the influence of one or more qualitative variables on the mean of a numerical variable in a population. In ANOVA the response

More information

H0: Tested by k-grp ANOVA

H0: Tested by k-grp ANOVA Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures Alpha estimation reconsidered H0: Tested by k-grp ANOVA Regardless

More information

The Random Effects Model Introduction

The Random Effects Model Introduction The Random Effects Model Introduction Sometimes, treatments included in experiment are randomly chosen from set of all possible treatments. Conclusions from such experiment can then be generalized to other

More information

Inference for Regression

Inference for Regression Inference for Regression Section 9.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 13b - 3339 Cathy Poliak, Ph.D. cathy@math.uh.edu

More information

Statistics for EES Factorial analysis of variance

Statistics for EES Factorial analysis of variance Statistics for EES Factorial analysis of variance Dirk Metzler June 12, 2015 Contents 1 ANOVA and F -Test 1 2 Pairwise comparisons and multiple testing 6 3 Non-parametric: The Kruskal-Wallis Test 9 1 ANOVA

More information

STAT22200 Spring 2014 Chapter 13B

STAT22200 Spring 2014 Chapter 13B STAT22200 Spring 2014 Chapter 13B Yibi Huang May 27, 2014 13.3.1 Crossover Designs 13.3.4 Replicated Latin Square Designs 13.4 Graeco-Latin Squares Chapter 13B - 1 13.3.1 Crossover Design (A Special Latin-Square

More information

Lecture 18: Simple Linear Regression

Lecture 18: Simple Linear Regression Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength

More information

Outline. Example and Model ANOVA table F tests Pairwise treatment comparisons with LSD Sample and subsample size determination

Outline. Example and Model ANOVA table F tests Pairwise treatment comparisons with LSD Sample and subsample size determination Outline 1 The traditional approach 2 The Mean Squares approach for the Completely randomized design (CRD) CRD and one-way ANOVA Variance components and the F test Inference about the intercept Sample vs.

More information

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES

4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES 4:3 LEC - PLANNED COMPARISONS AND REGRESSION ANALYSES FOR SINGLE FACTOR BETWEEN-S DESIGNS Planned or A Priori Comparisons We previously showed various ways to test all possible pairwise comparisons for

More information

y response variable x 1, x 2,, x k -- a set of explanatory variables

y response variable x 1, x 2,, x k -- a set of explanatory variables 11. Multiple Regression and Correlation y response variable x 1, x 2,, x k -- a set of explanatory variables In this chapter, all variables are assumed to be quantitative. Chapters 12-14 show how to incorporate

More information

MAT3378 (Winter 2016)

MAT3378 (Winter 2016) MAT3378 (Winter 2016) Assignment 2 - SOLUTIONS Total number of points for Assignment 2: 12 The following questions will be marked: Q1, Q2, Q4 Q1. (4 points) Assume that Z 1,..., Z n are i.i.d. normal random

More information